期刊文献+

基于径向基函数网络的强化学习在机器人足球中的研究(英文) 被引量:2

Study on Radial Basis Function Networks Based Reinforcement Learning in Robot Soccer
下载PDF
导出
摘要 与监督学习从范例中学习的方式不同,强化学习不需要先验知识,而是具有从经验中学习的能力。将强化学习应用在大状态空间中,必须应用函数逼近的方法,如使用径向基函数网络建立输入和输出的映射关系。本文对基于径向基函数网络的强化学习在机器人足球这样的动态的多智能体环境中的应用进行了研究。实验结果证明了研究方法的可行性。 Reinforcement learning has the ability to learn from experience as opposed to supervised learning which learns from examples. Application of reinforcement learning to large state spaces necessitates the use of function approximators like Radial Basis Function Networks (RBFNs) to map between inputs and outputs. This study investigates the applicability of RBFNs based reinforcement learning methods in a dynamic multi-agent scenario of robot soccer. And the result of experiment proves that is a suitable approach. 1 Introduction1 Robot soccer is a real time, dynamic and multi-agent environment. Learning algorithms have enabled researchers to handle this kind of complex domain[1]. Reinforcement learning is increasingly attracting researchers for its ability to learn from environment by evaluating actions in the form of rewards and penalties[2]. Many reinforcement learning algorithms are based on look up table representation of the state[3]. Real life problems often involve huge state spaces, which make table based state representation impossible. Radial Basis Function Networks (RBFNs) poses as an attractive method of function approximation for the task[3][4]. This study provides an approach on RBFNs based reinforcement learning in robot soccer domain. 2 Real Robot Soccer Simulation We designed a simulation platform for real robot soccer in form of 3 vs. 3 as shown in Fig.1. The simulation game is running via the method of server/client. The platform iscomposed of three components: the server, and the two clients. The server provides virtual field, virtual vision information,kinematics model of robots and ball, collision test and treatment Fig.1 Simulator for real robot soccer kinematics model of robots and ball, collision test and treatment model[5]. Of course it also receives control command from clients and displays the game on the screen. While each client receives the information of the situation on the field sent by the server, selects one strategy, which selects the basic movement for each robot, according to the situation on the field, and it sends control command to the server, which controls the action
出处 《系统仿真学报》 CAS CSCD 2002年第8期1094-1097,共4页 Journal of System Simulation
关键词 径向基函数 强化学习 机器人足球 多智能体 multi-agent robot soccer reinforcement learning RBFNs
  • 相关文献

参考文献1

二级参考文献4

  • 1冯瑞,机器人足球研讨班论文集,1998年
  • 2高大志,机器人足球研讨班论文集,1998年
  • 3Norton P,MFC开发Windows 95/NT4应用程序,1998年
  • 4Horizon S,Microsoft 基本类库开发指南,1995年

共引文献8

同被引文献18

  • 1薛方正,方帅,徐心和.多机器人对抗系统仿真中的对手建模[J].系统仿真学报,2005,17(9):2138-2141. 被引量:7
  • 2Schneider-Fontan M,Mataric M J.Territorial multi-robot task division[J].IEEE Transactions on Robotics and Automation,1998,14(5):815-822.
  • 3Mataric M J.Learning in behavior-based multi-robot systems:policies.models,and other agents[J].Journal of Cognitive Systems Research,2001 (2):81-93.
  • 4Barto A G.Mahanevan S.Recent advances in hierarchical reinforcement leaming[J].Discrete Event Dynamic Sysems:Theory and Applications,2003,13(4):41-77.
  • 5Sutton R S,Precup D,Singh S P.Between MDPs and semi-MDPs:a framework for temporal abstraction in reinforcement learning[J].Artificial Intelligence,1999,112(1):181-211.
  • 6Chaimowicz L,Campos M F M,Kumar V.Dynamic role assignment for cooperative robots[C]//Proc of the IEEE International Conference on Robotics and Automation.IEEE Press,2002:293-298.
  • 7Perkins T J,Barto A G.Lyapunov design for safe reinforcement learning[J].Journal of Machine Learning Research,2003,3 (4/5):803-832.
  • 8Kok J R, Vlassis N. Sparse Cooperative Q-learning [C]//Greiner R, Schuurmans D, eds.: Proc. of the 21st Int. Conf. on Machine Learning, Banff, Alberta, Canada: ACM, 2004: 481-488.
  • 9Stone P, Sutton R. Scaling reinforcement learning toward RoboCup soccer[C]//Pro, of the 18th International Conf on Machine Learning, Morgan Kaufmann, Berkshires, Massachusetts, ACM, 2001.
  • 10Nakashima Tomoharu, Udo Masayo. A Fuzzy Reinforcement Learning for a Ball Interception Problem[C]//D. Polani et al. (Eds.): RoboCup 2003, Andrea Bonarini, Springer LNAI 3020, 2004: 559-567.

引证文献2

二级引证文献6

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部