Improving numerical forecasting skill in the atmospheric and oceanic sciences by solving optimization problems is an important issue. One such method is to compute the conditional nonlinear optimal perturbation(CNOP),...Improving numerical forecasting skill in the atmospheric and oceanic sciences by solving optimization problems is an important issue. One such method is to compute the conditional nonlinear optimal perturbation(CNOP), which has been applied widely in predictability studies. In this study, the Differential Evolution(DE) algorithm, which is a derivative-free algorithm and has been applied to obtain CNOPs for exploring the uncertainty of terrestrial ecosystem processes, was employed to obtain the CNOPs for finite-dimensional optimization problems with ball constraint conditions using Burgers' equation. The aim was first to test if the CNOP calculated by the DE algorithm is similar to that computed by traditional optimization algorithms, such as the Spectral Projected Gradient(SPG2) algorithm. The second motive was to supply a possible route through which the CNOP approach can be applied in predictability studies in the atmospheric and oceanic sciences without obtaining a model adjoint system, or for optimization problems with non-differentiable cost functions. A projection skill was first explanted to the DE algorithm to calculate the CNOPs. To validate the algorithm, the SPG2 algorithm was also applied to obtain the CNOPs for the same optimization problems. The results showed that the CNOPs obtained by the DE algorithm were nearly the same as those obtained by the SPG2 algorithm in terms of their spatial distributions and nonlinear evolutions. The implication is that the DE algorithm could be employed to calculate the optimal values of optimization problems, especially for non-differentiable and nonlinear optimization problems associated with the atmospheric and oceanic sciences.展开更多
The linear conjugate gradient method is an optimal method for convex quadratic minimization due to the Krylov subspace minimization property. The proposition of limited-memory BFGS method and Barzilai-Borwein gradient...The linear conjugate gradient method is an optimal method for convex quadratic minimization due to the Krylov subspace minimization property. The proposition of limited-memory BFGS method and Barzilai-Borwein gradient method, however, heavily restricted the use of conjugate gradient method for largescale nonlinear optimization. This is, to the great extent, due to the requirement of a relatively exact line search at each iteration and the loss of conjugacy property of the search directions in various occasions. On the contrary, the limited-memory BFGS method and the Barzilai-Bowein gradient method share the so-called asymptotical one stepsize per line-search property, namely, the trial stepsize in the method will asymptotically be accepted by the line search when the iteration is close to the solution. This paper will focus on the analysis of the subspace minimization conjugate gradient method by Yuan and Stoer(1995). Specifically, if choosing the parameter in the method by combining the Barzilai-Borwein idea, we will be able to provide some efficient Barzilai-Borwein conjugate gradient(BBCG) methods. The initial numerical experiments show that one of the variants, BBCG3, is specially efficient among many others without line searches. This variant of the BBCG method might enjoy the asymptotical one stepsize per line-search property and become a strong candidate for large-scale nonlinear optimization.展开更多
基金provided by grants from the LASG State Key Laboratory Special Fundthe National Natural Science Foundation of China (Grant Nos. 40905050, 40830955, and 41375111)
文摘Improving numerical forecasting skill in the atmospheric and oceanic sciences by solving optimization problems is an important issue. One such method is to compute the conditional nonlinear optimal perturbation(CNOP), which has been applied widely in predictability studies. In this study, the Differential Evolution(DE) algorithm, which is a derivative-free algorithm and has been applied to obtain CNOPs for exploring the uncertainty of terrestrial ecosystem processes, was employed to obtain the CNOPs for finite-dimensional optimization problems with ball constraint conditions using Burgers' equation. The aim was first to test if the CNOP calculated by the DE algorithm is similar to that computed by traditional optimization algorithms, such as the Spectral Projected Gradient(SPG2) algorithm. The second motive was to supply a possible route through which the CNOP approach can be applied in predictability studies in the atmospheric and oceanic sciences without obtaining a model adjoint system, or for optimization problems with non-differentiable cost functions. A projection skill was first explanted to the DE algorithm to calculate the CNOPs. To validate the algorithm, the SPG2 algorithm was also applied to obtain the CNOPs for the same optimization problems. The results showed that the CNOPs obtained by the DE algorithm were nearly the same as those obtained by the SPG2 algorithm in terms of their spatial distributions and nonlinear evolutions. The implication is that the DE algorithm could be employed to calculate the optimal values of optimization problems, especially for non-differentiable and nonlinear optimization problems associated with the atmospheric and oceanic sciences.
基金supported by National Natural Science Foundation of China (Grant Nos. 81173633, 11401038 and 11331012)the Chinese Academy of Sciences Grant (Grant No. kjcx-yw-s7-03)+2 种基金National Natural Science Foundation of China for Distinguished Young Scientists (Grant No. 11125107)the Key Project of Chinese National Programs for Fundamental Research and Development (Grant No. 2015CB856000)the Fundamental Research Funds for the Central Universities (Grant No. 2014RC0904)
文摘The linear conjugate gradient method is an optimal method for convex quadratic minimization due to the Krylov subspace minimization property. The proposition of limited-memory BFGS method and Barzilai-Borwein gradient method, however, heavily restricted the use of conjugate gradient method for largescale nonlinear optimization. This is, to the great extent, due to the requirement of a relatively exact line search at each iteration and the loss of conjugacy property of the search directions in various occasions. On the contrary, the limited-memory BFGS method and the Barzilai-Bowein gradient method share the so-called asymptotical one stepsize per line-search property, namely, the trial stepsize in the method will asymptotically be accepted by the line search when the iteration is close to the solution. This paper will focus on the analysis of the subspace minimization conjugate gradient method by Yuan and Stoer(1995). Specifically, if choosing the parameter in the method by combining the Barzilai-Borwein idea, we will be able to provide some efficient Barzilai-Borwein conjugate gradient(BBCG) methods. The initial numerical experiments show that one of the variants, BBCG3, is specially efficient among many others without line searches. This variant of the BBCG method might enjoy the asymptotical one stepsize per line-search property and become a strong candidate for large-scale nonlinear optimization.