期刊文献+

Efficient Exploration for Multi-Agent Reinforcement Learning via Transferable Successor Features 被引量:1

下载PDF
导出
摘要 In multi-agent reinforcement learning(MARL),the behaviors of each agent can influence the learning of others,and the agents have to search in an exponentially enlarged joint-action space.Hence,it is challenging for the multi-agent teams to explore in the environment.Agents may achieve suboptimal policies and fail to solve some complex tasks.To improve the exploring efficiency as well as the performance of MARL tasks,in this paper,we propose a new approach by transferring the knowledge across tasks.Differently from the traditional MARL algorithms,we first assume that the reward functions can be computed by linear combinations of a shared feature function and a set of taskspecific weights.Then,we define a set of basic MARL tasks in the source domain and pre-train them as the basic knowledge for further use.Finally,once the weights for target tasks are available,it will be easier to get a well-performed policy to explore in the target domain.Hence,the learning process of agents for target tasks is speeded up by taking full use of the basic knowledge that was learned previously.We evaluate the proposed algorithm on two challenging MARL tasks:cooperative boxpushing and non-monotonic predator-prey.The experiment results have demonstrated the improved performance compared with state-of-the-art MARL algorithms.
出处 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2022年第9期1673-1686,共14页 自动化学报(英文版)
基金 the National Key R&D Program of China(2021ZD0112700,2018AAA0101400) the National Natural Science Foundation of China(62173251,61921004,U1713209) the Natural Science Foundation of Jiangsu Province of China(BK20202006)。
  • 相关文献

参考文献1

二级参考文献8

共引文献13

同被引文献17

引证文献1

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部