期刊文献+

离散非线性零和博弈的事件驱动最优控制方案 被引量:2

Event-triggered optimal control scheme for discrete-time nonlinear zero-sum games
下载PDF
导出
摘要 在求解离散非线性零和博弈问题时,为了在有效降低网络通讯和控制器执行次数的同时保证良好的控制效果,本文提出了一种基于事件驱动机制的最优控制方案.首先,设计了一个采用新型事件驱动阈值的事件驱动条件,并根据贝尔曼最优性原理获得了最优控制对的表达式.为了求解该表达式中的最优值函数,提出了一种单网络值迭代算法.利用一个神经网络构建评价网.设计了新的评价网权值更新规则.通过在评价网、控制策略及扰动策略之间不断迭代,最终获得零和博弈问题的最优值函数和最优控制对.然后,利用Lyapunov稳定性理论证明了闭环系统的稳定性.最后,将该事件驱动最优控制方案应用到了两个仿真例子中,验证了所提方法的有效性. In order to reduce the network communication and controller execution frequency while guarantee a desired control performance, an event-triggered optimal control scheme is proposed for solving the optimal control pair of discretetime nonlinear zero-sum games in this paper. Firstly, an event-triggered condition with new event-triggered threshold is designed. The expression of the optimal control pair is obtained based on the Bellman optimality principle. Then, a single network value iteration algorithm is proposed to solve the optimal value function in this expression. A neural network is used to construct the critic network. Novel weight update rule of the critic network is derived. Through the iteration between the critic network, the control policy and the disturbance policy, the optimal value function and the optimal control pair can be solved. Further, the Lyapunov theory is used to prove the stability of the event-triggered closed-loop system.Finally, the event-triggered optimal control mechanism is applied to two examples to verify its effectiveness.
作者 张欣 薄迎春 崔黎黎 ZHANG Xin;BO Ying-chun;CUI Li-li(College of Information and Control Engineering,China University of Petroleum,Qingdao Shandong 266580,China;Sofeware College,Shenyang Normal University,Shenyang Liaoning 110034,China)
出处 《控制理论与应用》 EI CAS CSCD 北大核心 2018年第5期619-626,共8页 Control Theory & Applications
基金 山东省自然科学基金项目(BS2015DX009) 国家自然科学基金项目(61703289)资助~~
关键词 博弈论 事件驱动 自适应动态规划 最优控制 game theory event-triggered adaptive dynamic programming optimal control
  • 相关文献

参考文献4

二级参考文献36

  • 1Draguna VRABIE,Frank LEWIS.Adaptive dynamic programming for online solution of a zero-sum differential game[J].控制理论与应用(英文版),2011,9(3):353-360. 被引量:10
  • 2KLEINMAN D. On an iterative technique for Riccati equation com- putations [J]. IEEE Transactions on Automatic Control, 1968, 13(1): 114- 115.
  • 3FENG Y, ANDERSON B D, ROTKOWITZ M. A game theoretic al- gorithm to compute local stabilizing solutions to HJBI equations in nonlinear Hoo control [J]. Automatica 2009, 45(4): 881 - 888.
  • 4VAN DER SCHAFT AJ. L2-gain analysis of nonlinear systems and nonlinear state feedback Ho control [J]. IEEE Transactions on Au- tomatic Control, 1992, 37(6): 770 - 784.
  • 5ABU-KHALAF M, LEWIS F L. Neurodynamic programming and zero-sum games for constrained control systems [J]. IEEE Transac- tions on Neural Networks, 2008, 19(7): 1243 - 1252.
  • 6ABU-KHALAF M, LEWIS F L, HUANG J. Policy iterations on the Hamilton-Jacobi-Isaacs equation for Ha state feedback control with input saturation [J]. IEEE Transactions on Automatic Control, 2006, 51(12): 1989- 1995.
  • 7VAMVOUDAKIS, K G, LEWIS F L. Online solution of nonlinear two-player zero-sum games using synchronous policy iteration [J]. International Journal of Robust and Nonlinear Control, 2012, 22( 13): 1460- 1483.
  • 8WU H N, LUO B. Simultaneous policy update algorithms for learn- ing the solution of linear continuous-time Hoo state feedback con- trol [J]. Information Sciences, 2013, 222:472 -485.
  • 9LI H L, LIU D R, WANG D. Integral reinforcement learning for linear continuous-time zero-sum games with completely unknown dynam- ics [J]. IEEE Transactions on Automation Science and Engineering, 2014, 11(3): 706 - 714.
  • 10LEWI F L, VRABIE D L, SYRMOS V L. Optimal Control [M]. New Jersey: Wiley, 2012.

共引文献13

同被引文献32

引证文献2

二级引证文献2

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部