A penalized interior point approach for constrained nonlinear programming is examined in this work. To overcome the difficulty of initialization for the interior point method, a problem equivalent to the primal proble...A penalized interior point approach for constrained nonlinear programming is examined in this work. To overcome the difficulty of initialization for the interior point method, a problem equivalent to the primal problem via incorporating an auxiliary variable is constructed. A combined approach of logarithm barrier and quadratic penalty function is proposed to solve the problem. Based on Newton's method, the global convergence of interior point and line search algorithm is proven. Only a finite number of iterations is required to reach an approximate optimal solution. Numerical tests are given to show the effectiveness of the method.展开更多
In this paper, we propose a primal-dual interior point method for solving general constrained nonlinear programming problems. To avoid the situation that the algorithm we use may converge to a saddle point or a local ...In this paper, we propose a primal-dual interior point method for solving general constrained nonlinear programming problems. To avoid the situation that the algorithm we use may converge to a saddle point or a local maximum, we utilize a merit function to guide the iterates toward a local minimum. Especially, we add the parameter ε to the Newton system when calculating the decrease directions. The global convergence is achieved by the decrease of a merit function. Furthermore, the numerical results confirm that the algorithm can solve this kind of problems in an efficient way.展开更多
A mechanism for proving global convergence in filter-SQP (sequence of quadratic programming) method with the nonlinear complementarity problem (NCP) function is described for constrained nonlinear optimization pro...A mechanism for proving global convergence in filter-SQP (sequence of quadratic programming) method with the nonlinear complementarity problem (NCP) function is described for constrained nonlinear optimization problem.We introduce an NCP function into the filter and construct a new SQP-filter algorithm.Such methods are characterized by their use of the dominance concept of multi-objective optimization,instead of a penalty parameter whose adjustment can be problematic.We prove that the algorithm has global convergence and superlinear convergence rates under some mild conditions.展开更多
Conjugate gradient optimization algorithms depend on the search directions with different choices for the parameters in the search directions. In this note, by combining the nice numerical performance of PR and HS met...Conjugate gradient optimization algorithms depend on the search directions with different choices for the parameters in the search directions. In this note, by combining the nice numerical performance of PR and HS methods with the global convergence property of the class of conjugate gradient methods presented by HU and STOREY(1991), a class of new restarting conjugate gradient methods is presented. Global convergences of the new method with two kinds of common line searches, are proved. Firstly, it is shown that, using reverse modulus of continuity function and forcing function, the new method for solving unconstrained optimization can work for a continously dif ferentiable function with Curry-Altman's step size rule and a bounded level set. Secondly, by using comparing technique, some general convergence properties of the new method with other kind of step size rule are established. Numerical experiments show that the new method is efficient by comparing with FR conjugate gradient method.展开更多
In this papert a recursive quadratic programming algorithm is proposed andstudied.The line search functions used are Han's nondifferentiable penalty functionswith a second order penalty term. In order to avoid mar...In this papert a recursive quadratic programming algorithm is proposed andstudied.The line search functions used are Han's nondifferentiable penalty functionswith a second order penalty term. In order to avoid maratos effect,Fukushima's mixeddirection is used as the direction of line search.Finallyt we prove the global convergenceand the local second order convergence of the algorithm.展开更多
In this note,by combining the nice numerical performance of PR and HS methods with the global convergence property of FR method,a class of new restarting three terms conjugate gradient methods is presented.Global conv...In this note,by combining the nice numerical performance of PR and HS methods with the global convergence property of FR method,a class of new restarting three terms conjugate gradient methods is presented.Global convergence properties of the new method with two kinds of common line searches are proved.展开更多
Conjugate gradient optimization algorithms depend on the search directions with different choices for the parameter in the search directions. In this note, conditions are given on the parameter in the conjugate gradie...Conjugate gradient optimization algorithms depend on the search directions with different choices for the parameter in the search directions. In this note, conditions are given on the parameter in the conjugate gradient directions to ensure the descent property of the search directions. Global convergence of such a class of methods is discussed. It is shown that, using reverse modulus of continuity function and forcing function, the new method for solving unconstrained optimization can work for a continuously differentiable function with a modification of the Curry-Altman's step-size rule and a bounded level set. Combining PR method with our new method, PR method is modified to have global convergence property.Numerical experiments show that the new methods are efficient by comparing with FR conjugate gradient method.展开更多
基金supported by the National Natural Science Foundation of China (Grant No.10771133)the Shanghai Leading Academic Discipline Project (Grant Nos.J50101, S30104)
文摘A penalized interior point approach for constrained nonlinear programming is examined in this work. To overcome the difficulty of initialization for the interior point method, a problem equivalent to the primal problem via incorporating an auxiliary variable is constructed. A combined approach of logarithm barrier and quadratic penalty function is proposed to solve the problem. Based on Newton's method, the global convergence of interior point and line search algorithm is proven. Only a finite number of iterations is required to reach an approximate optimal solution. Numerical tests are given to show the effectiveness of the method.
文摘In this paper, we propose a primal-dual interior point method for solving general constrained nonlinear programming problems. To avoid the situation that the algorithm we use may converge to a saddle point or a local maximum, we utilize a merit function to guide the iterates toward a local minimum. Especially, we add the parameter ε to the Newton system when calculating the decrease directions. The global convergence is achieved by the decrease of a merit function. Furthermore, the numerical results confirm that the algorithm can solve this kind of problems in an efficient way.
基金Project supported by the National Natural Science Foundation of China (Grant Nos.10571137,10771162)
文摘A mechanism for proving global convergence in filter-SQP (sequence of quadratic programming) method with the nonlinear complementarity problem (NCP) function is described for constrained nonlinear optimization problem.We introduce an NCP function into the filter and construct a new SQP-filter algorithm.Such methods are characterized by their use of the dominance concept of multi-objective optimization,instead of a penalty parameter whose adjustment can be problematic.We prove that the algorithm has global convergence and superlinear convergence rates under some mild conditions.
文摘Conjugate gradient optimization algorithms depend on the search directions with different choices for the parameters in the search directions. In this note, by combining the nice numerical performance of PR and HS methods with the global convergence property of the class of conjugate gradient methods presented by HU and STOREY(1991), a class of new restarting conjugate gradient methods is presented. Global convergences of the new method with two kinds of common line searches, are proved. Firstly, it is shown that, using reverse modulus of continuity function and forcing function, the new method for solving unconstrained optimization can work for a continously dif ferentiable function with Curry-Altman's step size rule and a bounded level set. Secondly, by using comparing technique, some general convergence properties of the new method with other kind of step size rule are established. Numerical experiments show that the new method is efficient by comparing with FR conjugate gradient method.
文摘In this papert a recursive quadratic programming algorithm is proposed andstudied.The line search functions used are Han's nondifferentiable penalty functionswith a second order penalty term. In order to avoid maratos effect,Fukushima's mixeddirection is used as the direction of line search.Finallyt we prove the global convergenceand the local second order convergence of the algorithm.
基金Supported by the National Natural Science Foundation of China(10571106) Supported by the Fundamental Research Funds for the Central Universities(10CX04044A)
文摘In this note,by combining the nice numerical performance of PR and HS methods with the global convergence property of FR method,a class of new restarting three terms conjugate gradient methods is presented.Global convergence properties of the new method with two kinds of common line searches are proved.
文摘Conjugate gradient optimization algorithms depend on the search directions with different choices for the parameter in the search directions. In this note, conditions are given on the parameter in the conjugate gradient directions to ensure the descent property of the search directions. Global convergence of such a class of methods is discussed. It is shown that, using reverse modulus of continuity function and forcing function, the new method for solving unconstrained optimization can work for a continuously differentiable function with a modification of the Curry-Altman's step-size rule and a bounded level set. Combining PR method with our new method, PR method is modified to have global convergence property.Numerical experiments show that the new methods are efficient by comparing with FR conjugate gradient method.