An algorithm for numerical solution of discrete Hamilton-Jacobi-Bellman equations is proposed.The method begins with a suitable initial guess value of the solution,then finds a suitable matrix to linearize the system ...An algorithm for numerical solution of discrete Hamilton-Jacobi-Bellman equations is proposed.The method begins with a suitable initial guess value of the solution,then finds a suitable matrix to linearize the system and constructs an iteration algorithm to generate the monotone sequence.The convergence of the algorithm for nonlinear discrete Hamilton-Jacobi-Bellman equations is proved.Some numerical examples are presented to confirm the effciency of this algorithm.展开更多
In the framework of stochastic processes, the connection between the dynamic programming scheme given by the Hamilton-Jacobi-Bellman equation and a recently proposed control approach based on the Fokker-Planck equatio...In the framework of stochastic processes, the connection between the dynamic programming scheme given by the Hamilton-Jacobi-Bellman equation and a recently proposed control approach based on the Fokker-Planck equation is discussed. Under appropriate assumptions it is shown that the two strategies are equivalent in the case of expected cost functionals, while the Fokker-Planck formalism allows considering a larger classof objectives. To illustratethe connection between the two control strategies, the cases of an Itō stochastic process and of a piecewise-deterministic process are considered.展开更多
This paper presents a modified domain decomposition method for the numerical solution of discrete Hamilton-Jacobi-Bellman equations arising from a class of optimal control problems using diffusion models.A convergence...This paper presents a modified domain decomposition method for the numerical solution of discrete Hamilton-Jacobi-Bellman equations arising from a class of optimal control problems using diffusion models.A convergence theorem is established.Numerical results indicate the effectiveness and accuracy of the method.展开更多
In this paper, we consider the optimal risk sharing problem between two parties in the insurance business: the insurer and the insured. The risk is allocated between the insurer and the insured by setting a deductible...In this paper, we consider the optimal risk sharing problem between two parties in the insurance business: the insurer and the insured. The risk is allocated between the insurer and the insured by setting a deductible and coverage in the insurance contract. We obtain the optimal deductible and coverage by considering the expected product of the two parties' utilities of terminal wealth according to stochastic optimal control theory. An equilibrium policy is also derived for when there are both a deductible and coverage;this is done by modelling the problem as a stochastic game in a continuous-time framework. A numerical example is provided to illustrate the results of the paper.展开更多
This paper is concerned with a novel integrated multi-step heuristic dynamic programming(MsHDP)algorithm for solving optimal control problems.It is shown that,initialized by the zero cost function,MsHDP can converge t...This paper is concerned with a novel integrated multi-step heuristic dynamic programming(MsHDP)algorithm for solving optimal control problems.It is shown that,initialized by the zero cost function,MsHDP can converge to the optimal solution of the Hamilton-Jacobi-Bellman(HJB)equation.Then,the stability of the system is analyzed using control policies generated by MsHDP.Also,a general stability criterion is designed to determine the admissibility of the current control policy.That is,the criterion is applicable not only to traditional value iteration and policy iteration but also to MsHDP.Further,based on the convergence and the stability criterion,the integrated MsHDP algorithm using immature control policies is developed to accelerate learning efficiency greatly.Besides,actor-critic is utilized to implement the integrated MsHDP scheme,where neural networks are used to evaluate and improve the iterative policy as the parameter architecture.Finally,two simulation examples are given to demonstrate that the learning effectiveness of the integrated MsHDP scheme surpasses those of other fixed or integrated methods.展开更多
文摘An algorithm for numerical solution of discrete Hamilton-Jacobi-Bellman equations is proposed.The method begins with a suitable initial guess value of the solution,then finds a suitable matrix to linearize the system and constructs an iteration algorithm to generate the monotone sequence.The convergence of the algorithm for nonlinear discrete Hamilton-Jacobi-Bellman equations is proved.Some numerical examples are presented to confirm the effciency of this algorithm.
基金the support by the European Science Foundation Exchange OPTPDE Grantthe support of CADMOS(Center for Advances Modeling and Science)Supported in part by the European Union under Grant Agreement“Multi-ITN STRIKE-Novel Methods in Computational Finance”.Fund Project No.304617 Marie Curie Research Training Network.
文摘In the framework of stochastic processes, the connection between the dynamic programming scheme given by the Hamilton-Jacobi-Bellman equation and a recently proposed control approach based on the Fokker-Planck equation is discussed. Under appropriate assumptions it is shown that the two strategies are equivalent in the case of expected cost functionals, while the Fokker-Planck formalism allows considering a larger classof objectives. To illustratethe connection between the two control strategies, the cases of an Itō stochastic process and of a piecewise-deterministic process are considered.
文摘This paper presents a modified domain decomposition method for the numerical solution of discrete Hamilton-Jacobi-Bellman equations arising from a class of optimal control problems using diffusion models.A convergence theorem is established.Numerical results indicate the effectiveness and accuracy of the method.
基金supported by the NSF of China(11931018, 12271274)the Tianjin Natural Science Foundation (19JCYBJC30400)。
文摘In this paper, we consider the optimal risk sharing problem between two parties in the insurance business: the insurer and the insured. The risk is allocated between the insurer and the insured by setting a deductible and coverage in the insurance contract. We obtain the optimal deductible and coverage by considering the expected product of the two parties' utilities of terminal wealth according to stochastic optimal control theory. An equilibrium policy is also derived for when there are both a deductible and coverage;this is done by modelling the problem as a stochastic game in a continuous-time framework. A numerical example is provided to illustrate the results of the paper.
基金the National Key Research and Development Program of China(2021ZD0112302)the National Natural Science Foundation of China(62222301,61890930-5,62021003)the Beijing Natural Science Foundation(JQ19013).
文摘This paper is concerned with a novel integrated multi-step heuristic dynamic programming(MsHDP)algorithm for solving optimal control problems.It is shown that,initialized by the zero cost function,MsHDP can converge to the optimal solution of the Hamilton-Jacobi-Bellman(HJB)equation.Then,the stability of the system is analyzed using control policies generated by MsHDP.Also,a general stability criterion is designed to determine the admissibility of the current control policy.That is,the criterion is applicable not only to traditional value iteration and policy iteration but also to MsHDP.Further,based on the convergence and the stability criterion,the integrated MsHDP algorithm using immature control policies is developed to accelerate learning efficiency greatly.Besides,actor-critic is utilized to implement the integrated MsHDP scheme,where neural networks are used to evaluate and improve the iterative policy as the parameter architecture.Finally,two simulation examples are given to demonstrate that the learning effectiveness of the integrated MsHDP scheme surpasses those of other fixed or integrated methods.