In this paper,three optimal linear formation control algorithms are proposed for first-order linear multiagent systems from a linear quadratic regulator(LQR) perspective with cost functions consisting of both interact...In this paper,three optimal linear formation control algorithms are proposed for first-order linear multiagent systems from a linear quadratic regulator(LQR) perspective with cost functions consisting of both interaction energy cost and individual energy cost,because both the collective ob ject(such as formation or consensus) and the individual goal of each agent are very important for the overall system.First,we propose the optimal formation algorithm for first-order multi-agent systems without initial physical couplings.The optimal control parameter matrix of the algorithm is the solution to an algebraic Riccati equation(ARE).It is shown that the matrix is the sum of a Laplacian matrix and a positive definite diagonal matrix.Next,for physically interconnected multi-agent systems,the optimal formation algorithm is presented,and the corresponding parameter matrix is given from the solution to a group of quadratic equations with one unknown.Finally,if the communication topology between agents is fixed,the local feedback gain is obtained from the solution to a quadratic equation with one unknown.The equation is derived from the derivative of the cost function with respect to the local feedback gain.Numerical examples are provided to validate the effectiveness of the proposed approaches and to illustrate the geometrical performances of multi-agent systems.展开更多
In this paper, a novel iterative Q-learning algorithm, called "policy iteration based deterministic Qlearning algorithm", is developed to solve the optimal control problems for discrete-time deterministic no...In this paper, a novel iterative Q-learning algorithm, called "policy iteration based deterministic Qlearning algorithm", is developed to solve the optimal control problems for discrete-time deterministic nonlinear systems. The idea is to use an iterative adaptive dynamic programming(ADP) technique to construct the iterative control law which optimizes the iterative Q function. When the optimal Q function is obtained, the optimal control law can be achieved by directly minimizing the optimal Q function, where the mathematical model of the system is not necessary. Convergence property is analyzed to show that the iterative Q function is monotonically non-increasing and converges to the solution of the optimality equation. It is also proven that any of the iterative control laws is a stable control law. Neural networks are employed to implement the policy iteration based deterministic Q-learning algorithm, by approximating the iterative Q function and the iterative control law, respectively. Finally, two simulation examples are presented to illustrate the performance of the developed algorithm.展开更多
基金supported by the National Natural Science Foundation of China(No.61375072)(50%)the Natural Science Foundation of Zhejiang Province,China(No.LQ16F030005)(50%)
文摘In this paper,three optimal linear formation control algorithms are proposed for first-order linear multiagent systems from a linear quadratic regulator(LQR) perspective with cost functions consisting of both interaction energy cost and individual energy cost,because both the collective ob ject(such as formation or consensus) and the individual goal of each agent are very important for the overall system.First,we propose the optimal formation algorithm for first-order multi-agent systems without initial physical couplings.The optimal control parameter matrix of the algorithm is the solution to an algebraic Riccati equation(ARE).It is shown that the matrix is the sum of a Laplacian matrix and a positive definite diagonal matrix.Next,for physically interconnected multi-agent systems,the optimal formation algorithm is presented,and the corresponding parameter matrix is given from the solution to a group of quadratic equations with one unknown.Finally,if the communication topology between agents is fixed,the local feedback gain is obtained from the solution to a quadratic equation with one unknown.The equation is derived from the derivative of the cost function with respect to the local feedback gain.Numerical examples are provided to validate the effectiveness of the proposed approaches and to illustrate the geometrical performances of multi-agent systems.
基金supported in part by National Natural Science Foundation of China(Grant Nos.6137410561233001+1 种基金61273140)in part by Beijing Natural Science Foundation(Grant No.4132078)
文摘In this paper, a novel iterative Q-learning algorithm, called "policy iteration based deterministic Qlearning algorithm", is developed to solve the optimal control problems for discrete-time deterministic nonlinear systems. The idea is to use an iterative adaptive dynamic programming(ADP) technique to construct the iterative control law which optimizes the iterative Q function. When the optimal Q function is obtained, the optimal control law can be achieved by directly minimizing the optimal Q function, where the mathematical model of the system is not necessary. Convergence property is analyzed to show that the iterative Q function is monotonically non-increasing and converges to the solution of the optimality equation. It is also proven that any of the iterative control laws is a stable control law. Neural networks are employed to implement the policy iteration based deterministic Q-learning algorithm, by approximating the iterative Q function and the iterative control law, respectively. Finally, two simulation examples are presented to illustrate the performance of the developed algorithm.