期刊文献+

强化学习在移动机器人自主导航中的应用 被引量:5

Application of reinforcement learning in autonomous navigation for mobile robot
下载PDF
导出
摘要 概述了移动机器人常用的自主导航算法及其优缺点,在此基础上提出了强化学习方法。描述了强化学习算法的原理,并实现了用神经网络解决泛化问题。设计了基于障碍物探测传感器信息的机器人自主导航强化学习方法,给出了学习算法中各要素的数学模型。经仿真验证,算法正确有效,具有良好的收敛性和泛化能力。 The merit and shortcomming of common algorithms of autonomous navigation for mobile robot are introduced,based on which the reinforcement learning method is proposed.The principle of the reinforcement learning is described,the generalization problem is solved by neural network.The autonomous navigation for robot based on obstacle detection sensor is designed,the mathematical model for each element of learning algorithm is proposed.The correctness ,reactiveness and the ability of generalization of the algorithm are verified by simulation tests.
出处 《计算机工程与应用》 CSCD 北大核心 2007年第18期215-217,228,共4页 Computer Engineering and Applications
关键词 强化学习 自主导航 神经网络 移动机器人 reinforcement learning autonomous navigation neural network mobile robot
  • 相关文献

参考文献10

  • 1Krogh B H.A generalized potential field approach to obstacle avoidance control[C]//Proceeding of the International Robotics Research Conference,1984:1150-1156.
  • 2Koren Y,Borenstein J.Potential field methods and their inherent limitations for mobile robot navigation[C]//Proceedings of the IEEE Conference on Robotics and Automation,Sacramento,California,1991:1394-1404.
  • 3张汝波,杨广铭,顾国昌,张国印.Q-学习及其在智能机器人局部路径规划中的应用研究[J].计算机研究与发展,1999,36(12):1430-1436. 被引量:17
  • 4Kamon I,Rivlin E,Rimon E.A new range-sensor based globally convergent navigation algorithm for mobile robots[C]//Proceedings of the IEEE International Conference on Robotics and Automation,Minneapolis,MN,1996:429-435.
  • 5Carreras M,Ridao P,Battle J,et al.Learning reactive robot behaviors with Neural-Q_learning[C]//IEEE-TTTC International Conference on Automation,Quality and Testing,Robotics,2002.
  • 6Gaskett CQ-learning for robot control[D].Australian National University,2002:5-13.
  • 7Ge S S,Cui Y J.New potential functions for mobile robot path planning[J].IEEE Transactions on Robotics and Automation,2000,16(5).
  • 8阎平凡.再励学习——原理、算法及其在智能控制中的应用[J].信息与控制,1996,25(1):28-34. 被引量:30
  • 9黄献龙,梁斌,吴宏鑫.机器人避碰规划综述[J].航天控制,2002,20(1):34-40. 被引量:41
  • 10焦鹏.水下自治机器人避碰路径规划算法研究[D].哈尔滨:哈尔滨工程大学,2006:59-66.

二级参考文献9

共引文献85

同被引文献38

  • 1张涛,吴汉生.基于神经网络的强化学习算法实现倒立摆控制[J].计算机仿真,2006,23(4):298-300. 被引量:7
  • 2王瑞霞,孙亮,阮晓钢.基于强化学习的二级倒立摆控制[J].计算机仿真,2006,23(4):305-308. 被引量:3
  • 3叶德谦,杨樱,金大兵.基于神经网络集成的强化学习算法系统设计[J].计算机工程与应用,2006,42(12):97-99. 被引量:2
  • 4黄炳强,曹广益,王占全.强化学习原理、算法及应用[J].河北工业大学学报,2006,35(6):34-38. 被引量:19
  • 5Preu P,Delepoulies S,Raqcheville J C. A generic architecture for adaptive agents based on reinforcement learning[J]. Information Sciences,2004,(161):37-55.
  • 6BO C M,WANG Z Q,LU A J. Study and application on dynamic modeling method based on SVM and sliding time window techniques[C]//Proceedings of the 6th World Congress on Intelligent Control and Automation. Piscataway: Institute of Electrical and Electronics Engineerings Inc. Press,2006:4714-4718.
  • 7Suykens J A K, Vandewale J. Least squares support vector machine classifiers[J]. Neural Processing Letters,1999,9(3):293-300.
  • 8Cicirelli G, D’Orazio T, Distante A. Neural Q-learning control architectures for wall-following behavior[C]// Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems,2003.
  • 9Carreras M, Ridao P, EI-Fakdi A. Semi-online neural Q-learning for real-time robot learning[C]//Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems,Las Vegas Nevada,2003:662-667.
  • 10Kondo T,Ito K. A reinforcement learning with evolutionary state recruitment strategy for autonomous mobile robots control[J]. Robotics ans Autonomous Systems,2004,46(2):121-124.

引证文献5

二级引证文献15

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部