期刊文献+

移动机器人的自适应式行为融合方法 被引量:5

Adaptive action fusion method for mobile robot
下载PDF
导出
摘要 介绍了一种基于先验知识的强化学习方法,它将传统的规则控制方法和强化学习方法相结合,在保留了已知的部分规则的情况下,利用强化学习方法对基本行为的融合机制进行了完善;同时,利用已知的规则知识对学习器进行指导,保证了学习向正确方向进行,有利于学习收敛速度的提高.文章给出了2种实现方法的结合方式,并给出了学习器的结构及参数和函数设定.最后以机器人围捕为研究背景,实现了移动机器人的自适应式行为融合,并利用仿真实验对其有效性进行验证.结果表明该方法具有收敛快、学习效果好的特点. A method to reinforce learning based on prior knowledge was proposed, combining the traditional rule control method with the reinforcement learning method. The action fusion mechanism preserves the partially known rules and utilizes the reinforcement learning to accomplish modification of rules. At the same time, the partially known rules give guidance to the learner, which may guarantee the correct learning direction and speed up the convergence. The combination pattern of the two methods is presented, with the architecture and the parameter setting of the learner. The method was used for adaptive action fusion of a mobile robot in a "pursuitevasion" game, and its efficiency was shown by simulation results. The results prove that this method converges in less time and has a good learning result.
出处 《哈尔滨工程大学学报》 EI CAS CSCD 北大核心 2005年第5期586-590,613,共6页 Journal of Harbin Engineering University
关键词 强化学习 多机器人 行为融合 reinforcement learning multi-agent action fusion
  • 相关文献

参考文献9

  • 1宋梅萍,顾国昌,张汝波.多移动机器人协作任务的分布式控制系统[J].机器人,2003,25(5):456-460. 被引量:19
  • 2MORENO D L. Using prior knowledge to improve reinforcement learning in mobile robotics[A]. Towards Autonomous Robotic Systems [C]. Colchester, UK, 2004.
  • 3DIXON K, MALAK R, KNOSLA P. Incorporating prior knowledge and previously learned information into reinforcement learning agents [R].Carnegie Mellon University, Institute for Complex Engineered Systems, 2000.
  • 4WATKINS C J, DAYAN P. Q-learning [J]. Machine Learning, 1992, 8(3-4): 279-292.
  • 5TAN Ming. Multi-agent reinforcement learning: Independent vs. cooperative agents [A]. Proceedings of the Tenth International Conference on Machine Learning [C]. Amherst, MA, USA, 1993.
  • 6LITTMAN M L. Friend-or-foe Q-learning in general-sum games [A]. Proceedings of the Eighteenth International Conference on Machine Learning [C]. Williamstown, MA, USA, 2001.
  • 7高阳,陈世福,陆鑫.强化学习研究综述[J].自动化学报,2004,30(1):86-100. 被引量:258
  • 8KAELBLING L P, LITTMAN M L, MOORE A W. Reinforcement learning: a survey [J]. Journal of Artificial Intelligence Research, 1996, 4(2): 237-285.
  • 9李宁,高阳,陆鑫,陈世福.一种基于强化学习的学习Agent[J].计算机研究与发展,2001,38(9):1051-1056. 被引量:26

二级参考文献13

共引文献295

同被引文献30

  • 1王鸿洁,常国岑.可拓神经网络在可拓专家系统中的研究与应用[J].系统工程与电子技术,2005,27(2):264-266. 被引量:10
  • 2阎平凡.再励学习——原理、算法及其在智能控制中的应用[J].信息与控制,1996,25(1):28-34. 被引量:30
  • 3Das T,Kar I N.Design and implementation of an adaptive fuzzy logic-based controller for wheeled mobile robots[J].IEEE Transactions on Control System Technology,2006,14 (3):501-510.
  • 4SE S, LOWED G, LITTLE J J. Vision- based global localization and mapping for mobile robots [J]. IEEE Transactions on Robotics, 2005, 21(3): 364-375.
  • 5BALTZAKIS H, ARGYROS A, TRAHANIAS P. Fusion of laser and visual data for robot motion planning and collision avoidance [ J ]. Machine Vision and Applications, 2003, 15: 92-100.
  • 6MONTEIRO SILDOMART, RIBEIRO CARLOS H C.Use of reinforcement learning algorithms under conditions of sensory ambiguity in moving robotics [ J ]. Controle and Automacao, 2004, 15(3) : 320 -338.
  • 7Maja J. Matari?. Reinforcement Learning in the Multi-Robot Domain[J] 1997,Autonomous Robots(1):73~83
  • 8Christopher J. C. H. Watkins,Peter Dayan. Q-learning[J] 1992,Machine Learning(3-4):279~292
  • 9单建华.基于行为的实时路径规划[J].控制工程,2009,16(3):367-370. 被引量:5
  • 10张世杰,段广仁.基于分布式控制的多个体系统的循环追踪控制[J].吉林大学学报(工学版),2010,40(2):506-510. 被引量:2

引证文献5

二级引证文献10

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部