期刊文献+

Derivative-free reinforcement learning:a review 被引量:3

原文传递
导出
摘要 Reinforcement learning is about learning agent models that make the best sequential decisions in unknown environments.In an unknown environment,the agent needs to explore the environment while exploiting the collected information,which usually forms a sophisticated problem to solve.Derivative-free optimization,meanwhile,is capable of solving sophisticated problems.It commonly uses a sampling-andupdating framework to iteratively improve the solution,where exploration and exploitation are also needed to be well balanced.Therefore,derivative-free optimization deals with a similar core issue as reinforcement learning,and has been introduced in reinforcement learning approaches,under the names of learning classifier systems and neuroevolution/evolutionary reinforcement learning.Although such methods have been developed for decades,recently,derivative-free reinforcement learning exhibits attracting increasing attention.However,recent survey on this topic is still lacking.In this article,we summarize methods of derivative-free reinforcement learning to date,and organize the methods in aspects including parameter updating,model selection,exploration,and parallel/distributed methods.Moreover,we discuss some current limitations and possible future directions,hoping that this article could bring more attentions to this topic and serve as a catalyst for developing novel and efficient approaches.
出处 《Frontiers of Computer Science》 SCIE EI CSCD 2021年第6期75-93,共19页 中国计算机科学前沿(英文版)
基金 This work was supported by the Program A for Outstanding PhD Candidate of Nanjing University,National Science Foundation of China(61876077) Jiangsu Science Foundation(BK20170013) Collaborative Innovation Center of Novel Software Technology and Industrialization.
  • 相关文献

同被引文献28

引证文献3

二级引证文献2

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部