期刊文献+

Off-policy integral reinforcement learning optimal tracking control for continuous-time chaotic systems

Off-policy integral reinforcement learning optimal tracking control for continuous-time chaotic systems
下载PDF
导出
摘要 This paper estimates an off-policy integral reinforcement learning(IRL) algorithm to obtain the optimal tracking control of unknown chaotic systems. Off-policy IRL can learn the solution of the HJB equation from the system data generated by an arbitrary control. Moreover, off-policy IRL can be regarded as a direct learning method, which avoids the identification of system dynamics. In this paper, the performance index function is first given based on the system tracking error and control error. For solving the Hamilton–Jacobi–Bellman(HJB) equation, an off-policy IRL algorithm is proposed.It is proven that the iterative control makes the tracking error system asymptotically stable, and the iterative performance index function is convergent. Simulation study demonstrates the effectiveness of the developed tracking control method. This paper estimates an off-policy integral reinforcement learning(IRL) algorithm to obtain the optimal tracking control of unknown chaotic systems. Off-policy IRL can learn the solution of the HJB equation from the system data generated by an arbitrary control. Moreover, off-policy IRL can be regarded as a direct learning method, which avoids the identification of system dynamics. In this paper, the performance index function is first given based on the system tracking error and control error. For solving the Hamilton–Jacobi–Bellman(HJB) equation, an off-policy IRL algorithm is proposed.It is proven that the iterative control makes the tracking error system asymptotically stable, and the iterative performance index function is convergent. Simulation study demonstrates the effectiveness of the developed tracking control method.
出处 《Chinese Physics B》 SCIE EI CAS CSCD 2015年第9期147-152,共6页 中国物理B(英文版)
基金 Project supported by the National Natural Science Foundation of China(Grant Nos.61304079 and 61374105) the Beijing Natural Science Foundation,China(Grant Nos.4132078 and 4143065) the China Postdoctoral Science Foundation(Grant No.2013M530527) the Fundamental Research Funds for the Central Universities,China(Grant No.FRF-TP-14-119A2) the Open Research Project from State Key Laboratory of Management and Control for Complex Systems,China(Grant No.20150104)
关键词 adaptive dynamic programming approximate dynamic programming chaotic system optimal tracking control adaptive dynamic programming,approximate dynamic programming,chaotic system,optimal tracking control
  • 相关文献

参考文献32

  • 1Lü J and Lu J 2003 Chaos Soliton. Fract. 17 127.
  • 2Xu C and Wu Y 2015 Appl. Math. Model. 39 2295.
  • 3Ma T, Zhang H and Fu J 2008 Chin. Phys. B 17 4407.
  • 4Ma T and Fu J 2011 Chin. Phys. B 20 050511.
  • 5Yang D 2014 Chin. Phys. B 23 010504.
  • 6Song R, Xiao W, Sun C and Wei Q 2013 Chin. Phys. B 22 090502.
  • 7Song R, Xiao W and Wei Q 2014 Chin. Phys. B 23 050504.
  • 8Gao S, Dong H, Sun X and Ning B 2015 Chin. Phys. B 24 010501.
  • 9Wei Q and Liu D 2014 IEEE Trans. Autom. Sci. Eng. 11 1020.
  • 10Wei Q and Liu D 2015 Neurocomputing 149 106.

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部