In this paper, a new superlinearly convergent algorithm for nonlinearly constrained optimization problems is presented. The search directions are directly computed by a few formulas, and neither quadratic programming ...In this paper, a new superlinearly convergent algorithm for nonlinearly constrained optimization problems is presented. The search directions are directly computed by a few formulas, and neither quadratic programming nor linear equation need to be sovled. Under mild assumptions, the new algorithm is shown to possess global and superlinear convergence.展开更多
An effective continuous algorithm is proposed to find approximate solutions of NP-hard max-cut problems. The algorithm relaxes the max-cut problem into a continuous nonlinear programming problem by replacing n discret...An effective continuous algorithm is proposed to find approximate solutions of NP-hard max-cut problems. The algorithm relaxes the max-cut problem into a continuous nonlinear programming problem by replacing n discrete constraints in the original problem with one single continuous constraint. A feasible direction method is designed to solve the resulting nonlinear programming problem. The method employs only the gradient evaluations of the objective function, and no any matrix calculations and no line searches are required. This greatly reduces the calculation cost of the method, and is suitable for the solution of large size max-cut problems. The convergence properties of the proposed method to KKT points of the nonlinear programming are analyzed. If the solution obtained by the proposed method is a global solution of the nonlinear programming problem, the solution will provide an upper bound on the max-cut value. Then an approximate solution to the max-cut problem is generated from the solution of the nonlinear programming and provides a lower bound on the max-cut value. Numerical experiments and comparisons on some max-cut test problems (small and large size) show that the proposed algorithm is efficient to get the exact solutions for all small test problems andwell satisfied solutions for most of the large size test problems with less calculation costs.展开更多
文摘In this paper, a new superlinearly convergent algorithm for nonlinearly constrained optimization problems is presented. The search directions are directly computed by a few formulas, and neither quadratic programming nor linear equation need to be sovled. Under mild assumptions, the new algorithm is shown to possess global and superlinear convergence.
基金This work is supported by National Natural Science Foundation of China at 10231060.
文摘An effective continuous algorithm is proposed to find approximate solutions of NP-hard max-cut problems. The algorithm relaxes the max-cut problem into a continuous nonlinear programming problem by replacing n discrete constraints in the original problem with one single continuous constraint. A feasible direction method is designed to solve the resulting nonlinear programming problem. The method employs only the gradient evaluations of the objective function, and no any matrix calculations and no line searches are required. This greatly reduces the calculation cost of the method, and is suitable for the solution of large size max-cut problems. The convergence properties of the proposed method to KKT points of the nonlinear programming are analyzed. If the solution obtained by the proposed method is a global solution of the nonlinear programming problem, the solution will provide an upper bound on the max-cut value. Then an approximate solution to the max-cut problem is generated from the solution of the nonlinear programming and provides a lower bound on the max-cut value. Numerical experiments and comparisons on some max-cut test problems (small and large size) show that the proposed algorithm is efficient to get the exact solutions for all small test problems andwell satisfied solutions for most of the large size test problems with less calculation costs.