期刊文献+

Timesharing-tracking Framework for Decentralized Reinforcement Learning in Fully Cooperative Multi-agent System

下载PDF
导出
摘要 Dimension-reduced and decentralized learning i always viewed as an efficient way to solve multi-agent cooperative learning in high dimension. However, the dynamic environmen brought by the concurrent learning makes the decentralized learning hard to converge and bad in performance. To tackle thi problem, a timesharing-tracking framework(TTF), stemming from the idea that alternative learning in microscopic view results in concurrent learning in macroscopic view, is proposed in this paper, in which the joint-state best-response Q-learning(BRQ-learning) serves as the primary algorithm to adapt to the companions policies. With the properly defined switching principle, TTF makes all agents learn the best responses to others at different joint states. Thus from the view of the whole joint-state space, agents learn the optimal cooperative policy simultaneously. The simulation results illustrate that the proposed algorithm can learn the optimal joint behavior with les computation and faster speed compared with other two classica learning algorithms.
出处 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI 2014年第2期127-133,共7页 自动化学报(英文版)
  • 相关文献

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部