期刊文献+

Multi-task Coalition Parallel Formation Strategy Based on Reinforcement Learning 被引量:6

Multi-task Coalition Parallel Formation Strategy Based on Reinforcement Learning
下载PDF
导出
摘要 代理人联盟是代理人协作和合作的一种重要方式。形成一个联盟,代理人能提高他们的能力解决问题并且获得更多的实用程序。在这份报纸,新奇多工联盟平行形成策略被介绍,并且多工联盟形成的过程是一个 Markov 决定过程的结论理论上被证明。而且,学习的加强被用来解决多工联盟平行的代理人行为策略,和这个过程形成被描述。在多工面向的领域,策略罐头有效地并且平行形式多工联盟。 Agent coalition is an important manner of agents' coordination and cooperation.Forming a coalition,agents can enhance their ability to solve problems and obtain more utilities.In this paper,a novel multi-task coalition parallel formation strategy is presented,and the conclusion that the process of multi-task coalition formation is a Markov decision process is testified theoretically.Moreover,reinforcement learning is used to solve agents' behavior strategy,and the process of multi-task coalition parallel formation is described.In multi-task oriented domains,the strategy can effectively and parallel form multi-task coalitions.
出处 《自动化学报》 EI CSCD 北大核心 2008年第3期349-352,共4页 Acta Automatica Sinica
基金 Supported by National Natural Science Foundation of China(60474035),National Research Foundation for the Doctoral Program of Higher Education of China(20050359004),Natural Science Foundation of Anhui Province(070412035)
关键词 强化学习 多任务合并 平行排列 马尔可夫决策过程 Multi-task coalition parallel formation Markov decision process reinforcement learning
  • 相关文献

参考文献4

二级参考文献77

  • 1蒋建国,夏娜,齐美彬,木春梅.一种基于蚁群算法的多任务联盟串行生成算法[J].电子学报,2005,33(12):2178-2182. 被引量:26
  • 2Justin Lenzo, Todd Sarver. Correlated Equilibrium in Evolutionary Models with Subpopulations [A]. The 15th Annual lnt Conf on Game Theory[C]. NY, 2004.
  • 3Eilon Solan, Nicolas Vieille. Correlated Equilibrium in Stochastic Games[J]. Games and Economic Behavior,2002, 38(2): 362-399.
  • 4Watkins P Dayan. Q-Learning[J]. Machine Learning,1992, 8(3); 279-292.
  • 5Michael L Littman, Szepesvari C. A Generalized Reinforcement Learning Model: Convergence and Applications[J]. Proc of the 13th Int Conf on Machine Learning[C]. Bari, 1996: 310-318.
  • 6Michael Bowling. Convergence Problems of Generalsum Multiagent Reinforcement Learning[A]. Proc of the 17th Int Conf on Machine Learning[C]. Stanford,2000 : 89-94.
  • 7Bowling M, Veloso M. Rational and Convergent Learning in Stochastic Games[A]. Proc of the 17th Int Joint Conf on Artificial Intelligence[C]. Seattle,2001 : 1021-1026.
  • 8Singh S, Kearns M, Mansour Y. Nash Convergence of Gradient Dynamics in General-sum Games [A].Proc of the 16th Conf on Uncertainty in Artificial Intelligence[C]. Stanford: Morgan Kaufman, 2000:541-548.
  • 9Fudenberg D, Levine D K. The Theory of Learning in Games[M]. Boston: MIT Press, 1999.
  • 10Michael L Littman, Peter Stone. A Polynomial-time Nash Equilibrium Algorithm for Repeated Games[A].ACM Conf on Electronic Commerce[C]. San Diego,2003: 48-54.

共引文献94

同被引文献67

引证文献6

二级引证文献97

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部