This work explores a family of two-block nonconvex optimization problems subject to linear constraints.We first introduce a simple but universal Bregman-style improved alternating direction method of multipliers(ADMM)...This work explores a family of two-block nonconvex optimization problems subject to linear constraints.We first introduce a simple but universal Bregman-style improved alternating direction method of multipliers(ADMM)based on the iteration framework of ADMM and the Bregman distance.Then,we utilize the smooth performance of one of the components to develop a linearized version of it.Compared to the traditional ADMM,both proposed methods integrate a convex combination strategy into the multiplier update step.For each proposed method,we demonstrate the convergence of the entire iteration sequence to a unique critical point of the augmented Lagrangian function utilizing the powerful Kurdyka–Łojasiewicz property,and we also derive convergence rates for both the sequence of merit function values and the iteration sequence.Finally,some numerical results show that the proposed methods are effective and encouraging for the Lasso model.展开更多
This work is about a splitting method for solving a nonconvex nonseparable optimization problem with linear constraints,where the objective function consists of two separable functions and a coupled term.First,based o...This work is about a splitting method for solving a nonconvex nonseparable optimization problem with linear constraints,where the objective function consists of two separable functions and a coupled term.First,based on the ideas from Bregman distance and Peaceman–Rachford splitting method,the Bregman Peaceman–Rachford splitting method with different relaxation factors for the multiplier is proposed.Second,the global and strong convergence of the proposed algorithm are proved under general conditions including the region of the two relaxation factors as well as the crucial Kurdyka–Łojasiewicz property.Third,when the associated Kurdyka–Łojasiewicz property function has a special structure,the sublinear and linear convergence rates of the proposed algorithm are guaranteed.Furthermore,some preliminary numerical results are shown to indicate the effectiveness of the proposed algorithm.展开更多
The alternating direction method of multipliers(ADMM)is one of the most successful and powerful methods for separable minimization optimization.Based on the idea of symmetric ADMM in two-block optimization,we add an u...The alternating direction method of multipliers(ADMM)is one of the most successful and powerful methods for separable minimization optimization.Based on the idea of symmetric ADMM in two-block optimization,we add an updating formula for the Lagrange multiplier without restricting its position for multiblock one.Then,combining with the Bregman distance,in this work,a Bregman-style partially symmetric ADMM is presented for nonconvex multi-block optimization with linear constraints,and the Lagrange multiplier is updated twice with different relaxation factors in the iteration scheme.Under the suitable conditions,the global convergence,strong convergence and convergence rate of the presented method are analyzed and obtained.Finally,some preliminary numerical results are reported to support the correctness of the theoretical assertions,and these show that the presented method is numerically effective.展开更多
In this paper, the nonlinear minimax problems are discussed. By means of the Sequential Quadratic Programming (SQP), a new descent algorithm for solving the problems is presented. At each iteration of the proposed a...In this paper, the nonlinear minimax problems are discussed. By means of the Sequential Quadratic Programming (SQP), a new descent algorithm for solving the problems is presented. At each iteration of the proposed algorithm, a main search direction is obtained by solving a Quadratic Programming (QP) which always has a solution. In order to avoid the Maratos effect, a correction direction is obtained by updating the main direction with a simple explicit formula. Under mild conditions without the strict complementarity, the global and superlinear convergence of the algorithm can be obtained. Finally, some numerical experiments are reported.展开更多
In this paper, a bundle modification strategy is proposed for nonsmooth convex constrained min- imization problems. As a result, a new feasible point bundle method is presented by applying this strategy. Whenever the ...In this paper, a bundle modification strategy is proposed for nonsmooth convex constrained min- imization problems. As a result, a new feasible point bundle method is presented by applying this strategy. Whenever the stability center is updated, some points in the bundle will be substituted by new ones which have lower objective values and/or constraint values, aiming at getting a better bundle. The method generates feasible serious iterates on which the objective function is monotonically decreasing. Global convergence of the algorithm is established, and some preliminary numerical results show that our method performs better than the standard feasible point bundle method.展开更多
In this paper,we present a QP-free algorithm without a penalty function or a filter for nonlinear semidefinite programming.At each iteration,two systems of linear equations with the same coefficient matrix are solved ...In this paper,we present a QP-free algorithm without a penalty function or a filter for nonlinear semidefinite programming.At each iteration,two systems of linear equations with the same coefficient matrix are solved to determine search direction;the nonmonotone line search ensures that the objective function or constraint violation function is sufficiently reduced.There is no feasibility restoration phase in our algorithm,which is necessary for traditional filter methods.The proposed algorithm is globally convergent under some mild conditions.Preliminary numerical results indicate that the proposed algorithm is comparable.展开更多
基金the National Natural Science Foundation of China(Nos.12171106 and 72071202)the Natural Science Foundation of Guangxi Province(No.2020GXNSFDA238017)Key Laboratory of Mathematics and Engineering Applications,Ministry of Education.
文摘This work explores a family of two-block nonconvex optimization problems subject to linear constraints.We first introduce a simple but universal Bregman-style improved alternating direction method of multipliers(ADMM)based on the iteration framework of ADMM and the Bregman distance.Then,we utilize the smooth performance of one of the components to develop a linearized version of it.Compared to the traditional ADMM,both proposed methods integrate a convex combination strategy into the multiplier update step.For each proposed method,we demonstrate the convergence of the entire iteration sequence to a unique critical point of the augmented Lagrangian function utilizing the powerful Kurdyka–Łojasiewicz property,and we also derive convergence rates for both the sequence of merit function values and the iteration sequence.Finally,some numerical results show that the proposed methods are effective and encouraging for the Lasso model.
基金supported by the National Natural Science Foundation of China(No.12171106)the Natural Science Foundation of Guangxi Province(Nos.2020GXNSFDA238017 and 2018GXNSFFA281007).
文摘This work is about a splitting method for solving a nonconvex nonseparable optimization problem with linear constraints,where the objective function consists of two separable functions and a coupled term.First,based on the ideas from Bregman distance and Peaceman–Rachford splitting method,the Bregman Peaceman–Rachford splitting method with different relaxation factors for the multiplier is proposed.Second,the global and strong convergence of the proposed algorithm are proved under general conditions including the region of the two relaxation factors as well as the crucial Kurdyka–Łojasiewicz property.Third,when the associated Kurdyka–Łojasiewicz property function has a special structure,the sublinear and linear convergence rates of the proposed algorithm are guaranteed.Furthermore,some preliminary numerical results are shown to indicate the effectiveness of the proposed algorithm.
基金supported by the National Natural Science Foundation of China (No.12171106)the Natural Science Foundation of Guangxi Province (No.2020GXNSFDA238017)。
文摘The alternating direction method of multipliers(ADMM)is one of the most successful and powerful methods for separable minimization optimization.Based on the idea of symmetric ADMM in two-block optimization,we add an updating formula for the Lagrange multiplier without restricting its position for multiblock one.Then,combining with the Bregman distance,in this work,a Bregman-style partially symmetric ADMM is presented for nonconvex multi-block optimization with linear constraints,and the Lagrange multiplier is updated twice with different relaxation factors in the iteration scheme.Under the suitable conditions,the global convergence,strong convergence and convergence rate of the presented method are analyzed and obtained.Finally,some preliminary numerical results are reported to support the correctness of the theoretical assertions,and these show that the presented method is numerically effective.
基金the National Natural Science Foundation of China(No.10261001)Guangxi Science Foundation(Nos.0236001,0640001)China as well as Guangxi University Key Program for Science and Technology Research(No.2005ZD02).
文摘In this paper, the nonlinear minimax problems are discussed. By means of the Sequential Quadratic Programming (SQP), a new descent algorithm for solving the problems is presented. At each iteration of the proposed algorithm, a main search direction is obtained by solving a Quadratic Programming (QP) which always has a solution. In order to avoid the Maratos effect, a correction direction is obtained by updating the main direction with a simple explicit formula. Under mild conditions without the strict complementarity, the global and superlinear convergence of the algorithm can be obtained. Finally, some numerical experiments are reported.
基金Project supported by the National Natural Science Foundation of China(11761013,11771383)Guangxi Natural Science Foundation(2013GXNSFAA019013,2014GXNSFFA118001,2016GXNSFDA380019)the Open Project of Guangxi Colleges and Universities Key Laboratory of Complex System Optimization and Big Data Processing(2016CSOBDP0203)
文摘In this paper, a bundle modification strategy is proposed for nonsmooth convex constrained min- imization problems. As a result, a new feasible point bundle method is presented by applying this strategy. Whenever the stability center is updated, some points in the bundle will be substituted by new ones which have lower objective values and/or constraint values, aiming at getting a better bundle. The method generates feasible serious iterates on which the objective function is monotonically decreasing. Global convergence of the algorithm is established, and some preliminary numerical results show that our method performs better than the standard feasible point bundle method.
基金supported by the National Natural Science Foundation(No.11561005)the National Science Foundation of Guangxi(No.2016GXNSFAA380248)。
文摘In this paper,we present a QP-free algorithm without a penalty function or a filter for nonlinear semidefinite programming.At each iteration,two systems of linear equations with the same coefficient matrix are solved to determine search direction;the nonmonotone line search ensures that the objective function or constraint violation function is sufficiently reduced.There is no feasibility restoration phase in our algorithm,which is necessary for traditional filter methods.The proposed algorithm is globally convergent under some mild conditions.Preliminary numerical results indicate that the proposed algorithm is comparable.