期刊文献+

一种改进的近似动态规划方法及其在SVC的应用 被引量:11

An improved approximate dynamic programming and its application in SVC control
下载PDF
导出
摘要 近似动态规划的基本思想是通过近似计算代价函数,从而避免动态规划中的"维数灾"问题。随机选取初值使得近似动态规划方法需要多次的学习才能最终收敛,极大地限制了在实际系统中的应用。针对上述问题,提出一种基于改进PID神经网络的直接启发式动态规划算法,将初始执行网络与PID控制器之间建立起一种等价关系,因此可以利用已经设计好的PID控制器来指导其初值选取,从而使算法收敛性大大提高。改进的神经网络与常规PID神经网络相比,结构简单且具有更好的扩展性,性能上具有更强的鲁棒性。对4机2区系统的静止无功补偿器附加阻尼控制进行仿真测试,仿真结果表明基于改进PID神经网络的直接启发式动态规划算法和初值选取方法的有效性,并且在部分状态反馈和延时两种情况下有着很好的控制效果。 The main idea of approximate dynamic programming(ADP) is approximately computing cost function to avoid the curse of dimension.However,it needs many times learning to converge due to the randomly choosing initial weights.So it is greatly limited in the application.This paper presents a direct heuristic dynamic programming(DHDP) based on an improved proportion integration differentiation PID neural network(IPIDNN).This method constructs an equivalent between the initial action network and PID controller.Therefore,well-designed PID controller can guide the initial weights choosing,so that the convergence of this algorithm will be remarkably improved.Moreover,compared with the traditional PID neural network,the configuration of IPIDNN is flexible and easy to expand,as well as a better robust performance.The simulation results show the validity of this algorithm and initial weights choosing method by the static var compensator(SVC) supplementary control in four-machine two-area system.It also has a good performance in the circumstance of partial state feedback and state delay.
出处 《电机与控制学报》 EI CSCD 北大核心 2011年第5期95-102,共8页 Electric Machines and Control
基金 国家自然科学基金海外及港澳学者合作研究基金(50828701)
关键词 近似动态规划 直接启发式动态规划 改进PID神经网络 静止无功补偿器附加阻尼控制 approximate dynamic programming direct heuristic dynamic programming improved proportion integration differentiation neural network static var compensator supplementary control
  • 相关文献

参考文献14

  • 1BELLMAN R,DREYFUS S. Applied Dynamic Programming[ M]. Princeton: Princeton University Press, 1962: 1- 15.
  • 2BORGERS T, SARIN R. Learning through reinforcement and rep- licator dynamics[ J]. Economic Theory, 1997, 77 (1) : 1 -17.
  • 3DALTON J, BALAKRISHNAN S N. A neighboring optimal adap- tive critic for missile guidance [ J]. Mathematical and Computer Modeling, 1996, 23(1-2): 175- 188.
  • 4WERBOS P. Advanced forecasting methods for global crisis warn-ing and models of intelligence [ J ]. General System Yearbook, 1977, 22:25 -38.
  • 5SI Jennie, BARTO Andrew G, POWELL Warren B, et al. Hand- book of learning and approximate dynamic programming: Scaling up to the real world[ M]. New York: IEEE Press and John Wiley & Sons, 2004:1 -17.
  • 6KIRK D E. Optimal Control Theory: An Introduction[ M]. Engle- wood Cliffs: Prentice-Hall, 1970 : 1 - 12.
  • 7BALAKRISHNAN S N, DING Jie, LEWIS Frank L. Issues on stability of ADP feedback controllers for dynamical systems [ J ]. IEEE Transactions on Systems, Man, and Cyberneties, Part B: Cybernetics, 2008, 38(4) : 913 -917.
  • 8SI Jennie, WANG Yutsung. Online learning control by association and reinforcement [ J]. IEEE Transactions on Neural Networks, 2001, 12(2) : 264 -276.
  • 9ENNS Russell, SI Jennie. Apache helicopter stabilization using neurodynamic programming[ J]. AIAA Journal of Guidance, Con- trol, and Dynamics, 2002, 25 ( 1 ) : 19 - 25.
  • 10ENNS Russell, SI Jennie. Helicopter trimming and tracking con- trol using direct neural dynamic programming[ J]. IEEE Transac- tions on Neural Networks, 2003, 14(4) : 929 -939.

同被引文献99

引证文献11

二级引证文献119

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部