期刊文献+
共找到482篇文章
< 1 2 25 >
每页显示 20 50 100
ON THE GLOBAL CONVERGENCE OF CONJUGATE GRADIENT METHODS WITH INEXACT LINESEARCH
1
作者 刘光辉 韩继业 《Numerical Mathematics A Journal of Chinese Universities(English Series)》 SCIE 1995年第2期147-153,共7页
In this paper we consider the global convergence of any conjugate gradient method of the form d1=-g1,dk+1=-gk+1+βkdk(k≥1)with any βk satisfying sume conditions,and with the strong wolfe line search conditions.Under... In this paper we consider the global convergence of any conjugate gradient method of the form d1=-g1,dk+1=-gk+1+βkdk(k≥1)with any βk satisfying sume conditions,and with the strong wolfe line search conditions.Under the convex assumption on the objective function,we preve the descenf property and the global convergence of this method. 展开更多
关键词 conjugate gradient method STRONG Wolfe line search global convergence.
下载PDF
GLOBAL CONVERGENCE OF THE GENERAL THREE TERM CONJUGATE GRADIENT METHODS WITH THE RELAXED STRONG WOLFE LINE SEARCH
2
作者 Xu Zeshui Yue ZhenjunInstitute of Sciences,PLA University of Science and Technology,Nanjing,210016. 《Applied Mathematics(A Journal of Chinese Universities)》 SCIE CSCD 2001年第1期58-62,共5页
The global convergence of the general three term conjugate gradient methods with the relaxed strong Wolfe line search is proved.
关键词 conjugate gradient method inexact line search global convergence.
下载PDF
A New Class of Nonlinear Conjugate Gradient Methods with Global Convergence Properties 被引量:1
3
作者 陈忠 《长江大学学报(自科版)(上旬)》 CAS 2014年第3期I0001-I0003,共3页
非线性共轭梯度法由于其迭代简单和储存量小,且搜索方向不需要满足正割条件,在求解大规模无约束优化问题时占据及其重要的地位.提出了一类新的共轭梯度法,其搜索方向是目标函数的下降方向.若假设目标函数连续可微且梯度满足Lipschitz条... 非线性共轭梯度法由于其迭代简单和储存量小,且搜索方向不需要满足正割条件,在求解大规模无约束优化问题时占据及其重要的地位.提出了一类新的共轭梯度法,其搜索方向是目标函数的下降方向.若假设目标函数连续可微且梯度满足Lipschitz条件,线性搜索满足Wolfe原则,讨论了所设计算法的全局收敛性. 展开更多
关键词 摘要 编辑部 编辑工作 读者
下载PDF
Convergence Analysis on a Class of Nonmonotone Conjugate Gradient Methods without Sufficient Decrease Condition 被引量:1
4
作者 DUShou-qiang CHENYuan-yuan 《Chinese Quarterly Journal of Mathematics》 CSCD 2004年第2期142-145,共4页
In [3] Liu et al. investigated global convergence of conjugate gradient methods. In that paper they allowed βκ to be selected in a wider range and the global convergence of the corresponding algorithm without suffic... In [3] Liu et al. investigated global convergence of conjugate gradient methods. In that paper they allowed βκ to be selected in a wider range and the global convergence of the corresponding algorithm without sufficient decrease condition was proved. This paper investigates global convergence of nonmonotone conjugate gradient method under the same conditions. 展开更多
关键词 nonmonotone conjugate gradient global convergence nonmonotone line search
下载PDF
A CLASSOF NONMONOTONE CONJUGATE GRADIENT METHODSFOR NONCONVEX FUNCTIONS
5
作者 LiuYun WeiZengxin 《Applied Mathematics(A Journal of Chinese Universities)》 SCIE CSCD 2002年第2期208-214,共7页
This paper discusses the global convergence of a class of nonmonotone conjugate gra- dient methods(NM methods) for nonconvex object functions.This class of methods includes the nonmonotone counterpart of modified Po... This paper discusses the global convergence of a class of nonmonotone conjugate gra- dient methods(NM methods) for nonconvex object functions.This class of methods includes the nonmonotone counterpart of modified Polak- Ribière method and modified Hestenes- Stiefel method as special cases 展开更多
关键词 nonmonotone conjugate gradient method nonmonotone line search global convergence unconstrained optimization.
下载PDF
CONVERGENCE ANALYSIS ON A CLASS OF CONJUGATE GRADIENT METHODS WITHOUTSUFFICIENT DECREASE CONDITION 被引量:1
6
作者 刘光辉 韩继业 +1 位作者 戚厚铎 徐中玲 《Acta Mathematica Scientia》 SCIE CSCD 1998年第1期11-16,共6页
Recently, Gilbert and Nocedal([3]) investigated global convergence of conjugate gradient methods related to Polak-Ribiere formular, they restricted beta(k) to non-negative value. [5] discussed the same problem as that... Recently, Gilbert and Nocedal([3]) investigated global convergence of conjugate gradient methods related to Polak-Ribiere formular, they restricted beta(k) to non-negative value. [5] discussed the same problem as that in [3] and relaxed beta(k) to be negative with the objective function being convex. This paper allows beta(k) to be selected in a wider range than [5]. Especially, the global convergence of the corresponding algorithm without sufficient decrease condition is proved. 展开更多
关键词 Polak-Ribiere conjugate gradient method strong Wolfe line search global convergence
全文增补中
An Adaptive Spectral Conjugate Gradient Method with Restart Strategy
7
作者 Zhou Jincheng Jiang Meixuan +2 位作者 Zhong Zining Wu Yanqiang Shao Hu 《数学理论与应用》 2024年第3期106-118,共13页
As a generalization of the two-term conjugate gradient method(CGM),the spectral CGM is one of the effective methods for solving unconstrained optimization.In this paper,we enhance the JJSL conjugate parameter,initiall... As a generalization of the two-term conjugate gradient method(CGM),the spectral CGM is one of the effective methods for solving unconstrained optimization.In this paper,we enhance the JJSL conjugate parameter,initially proposed by Jiang et al.(Computational and Applied Mathematics,2021,40:174),through the utilization of a convex combination technique.And this improvement allows for an adaptive search direction by integrating a newly constructed spectral gradient-type restart strategy.Then,we develop a new spectral CGM by employing an inexact line search to determine the step size.With the application of the weak Wolfe line search,we establish the sufficient descent property of the proposed search direction.Moreover,under general assumptions,including the employment of the strong Wolfe line search for step size calculation,we demonstrate the global convergence of our new algorithm.Finally,the given unconstrained optimization test results show that the new algorithm is effective. 展开更多
关键词 Unconstrained optimization Spectral conjugate gradient method Restart strategy inexact line search Global convergence
下载PDF
CONVERGENCE OF NONLINEAR CONJUGATE GRADIENT METHODS 被引量:1
8
作者 Yu-hong Dai (LSEC, ICMSEC, Academy of Mathematics and Systems Sciences, Chinese Academy of Sciences,Beijing 100080, China) 《Journal of Computational Mathematics》 SCIE EI CSCD 2001年第5期539-548,共10页
Presents a study on methods for unconstrained optimization. Assumptions of the study; Main results; Convergence properties of the methods under simplified Armijo-type line search.
关键词 unconstrained optimization conjugate gradient (generalized) line search global convergence
原文传递
A Scaled Conjugate Gradient Method Based on New BFGS Secant Equation with Modified Nonmonotone Line Search
9
作者 Tsegay Giday Woldu Haibin Zhang Yemane Hailu Fissuh 《American Journal of Computational Mathematics》 2020年第1期1-22,共22页
In this paper, we provide and analyze a new scaled conjugate gradient method and its performance, based on the modified secant equation of the Broyden-Fletcher-Goldfarb-Shanno (BFGS) method and on a new modified nonmo... In this paper, we provide and analyze a new scaled conjugate gradient method and its performance, based on the modified secant equation of the Broyden-Fletcher-Goldfarb-Shanno (BFGS) method and on a new modified nonmonotone line search technique. The method incorporates the modified BFGS secant equation in an effort to include the second order information of the objective function. The new secant equation has both gradient and function value information, and its update formula inherits the positive definiteness of Hessian approximation for general convex function. In order to improve the likelihood of finding a global optimal solution, we introduce a new modified nonmonotone line search technique. It is shown that, for nonsmooth convex problems, the proposed algorithm is globally convergent. Numerical results show that this new scaled conjugate gradient algorithm is promising and efficient for solving not only convex but also some large scale nonsmooth nonconvex problems in the sense of the Dolan-Moré performance profiles. 展开更多
关键词 conjugate gradient METHOD BFGS METHOD MODIFIED SECANT EQUATION NONMONOTONE line search Nonsmooth Optimization
下载PDF
A class of globally convergent conjugate gradient methods 被引量:6
10
作者 戴彧虹 袁亚湘 《Science China Mathematics》 SCIE 2003年第2期251-261,共11页
Conjugate gradient methods are very important ones for solving nonlinear optimization problems,especially for large scale problems. However, unlike quasi-Newton methods, conjugate gradient methods wereusually analyzed... Conjugate gradient methods are very important ones for solving nonlinear optimization problems,especially for large scale problems. However, unlike quasi-Newton methods, conjugate gradient methods wereusually analyzed individually. In this paper, we propose a class of conjugate gradient methods, which can beregarded as some kind of convex combination of the Fletcher-Reeves method and the method proposed byDai et al. To analyze this class of methods, we introduce some unified tools that concern a general methodwith the scalarβk having the form of φk/φk-1. Consequently, the class of conjugate gradient methods canuniformly be analyzed. 展开更多
关键词 UNCONSTRAINED optimization conjugate gradient line search global convergence.
原文传递
GLOBAL CONVERGENCE RESULTS OF A THREE TERM MEMORY GRADIENT METHOD WITH A NON-MONOTONE LINE SEARCH TECHNIQUE 被引量:12
11
作者 孙清滢 《Acta Mathematica Scientia》 SCIE CSCD 2005年第1期170-178,共9页
In this paper, a new class of three term memory gradient method with non-monotone line search technique for unconstrained optimization is presented. Global convergence properties of the new methods are discussed. Comb... In this paper, a new class of three term memory gradient method with non-monotone line search technique for unconstrained optimization is presented. Global convergence properties of the new methods are discussed. Combining the quasi-Newton method with the new method, the former is modified to have global convergence property. Numerical results show that the new algorithm is efficient. 展开更多
关键词 Non-linear programming three term memory gradient method convergence non-monotone line search technique numerical experiment
下载PDF
CONVERGENCE PROPERTIES OF THE DEPENDENT PRP CONJUGATE GRADIENT METHODS 被引量:1
12
作者 Shujun LIAN Changyu WANG Lixia CAO 《Journal of Systems Science & Complexity》 SCIE EI CSCD 2006年第2期288-296,共9页
In this paper, a new region of βk with respect to ;βk^PRP is given. With two Armijo-type line searches, the authors investigate the global convergence properties of the dependent PRP conjugate gradient methods, whic... In this paper, a new region of βk with respect to ;βk^PRP is given. With two Armijo-type line searches, the authors investigate the global convergence properties of the dependent PRP conjugate gradient methods, which extend the global convergence results of PRP conjugate gradient method proved by Grippo and Lucidi (1997) and Dai and Yuan (2002). 展开更多
关键词 conjugate gradient convergence property line search.
原文传递
A New Nonlinear Conjugate Gradient Method for Unconstrained Optimization Problems 被引量:1
13
作者 LIU Jin-kui WANG Kai-rong +1 位作者 SONG Xiao-qian DU Xiang-lin 《Chinese Quarterly Journal of Mathematics》 CSCD 2010年第3期444-450,共7页
In this paper,an efficient conjugate gradient method is given to solve the general unconstrained optimization problems,which can guarantee the sufficient descent property and the global convergence with the strong Wol... In this paper,an efficient conjugate gradient method is given to solve the general unconstrained optimization problems,which can guarantee the sufficient descent property and the global convergence with the strong Wolfe line search conditions.Numerical results show that the new method is efficient and stationary by comparing with PRP+ method,so it can be widely used in scientific computation. 展开更多
关键词 unconstrained optimization conjugate gradient method strong Wolfe line search sufficient descent property global convergence
下载PDF
A SUBSPACE PROJECTED CONJUGATE GRADIENT ALGORITHM FOR LARGE BOUND CONSTRAINED QUADRATIC PROGRAMMING 被引量:3
14
作者 倪勤 《Numerical Mathematics A Journal of Chinese Universities(English Series)》 SCIE 1998年第1期51-60,共10页
A subspace projected conjugate gradient method is proposed for solving large bound constrained quadratic programming. The conjugate gradient method is used to update the variables with indices outside of the active se... A subspace projected conjugate gradient method is proposed for solving large bound constrained quadratic programming. The conjugate gradient method is used to update the variables with indices outside of the active set, while the projected gradient method is used to update the active variables. At every iterative level, the search direction consists of two parts, one of which is a subspace trumcated Newton direction, another is a modified gradient direction. With the projected search the algorithm is suitable to large problems. The convergence of the method is proved and same numerical tests with dimensions ranging from 5000 to 20000 are given. 展开更多
关键词 Projected search conjugate gradient method LARGE problem BOUND constrained quadraic programming.
下载PDF
Global Convergence of a Hybrid Conjugate Gradient Method
15
作者 吴雪莎 《Chinese Quarterly Journal of Mathematics》 2015年第3期408-415,共8页
Conjugate gradient method is one of successful methods for solving the unconstrained optimization problems. In this paper, absorbing the advantages of FR and CD methods, a hybrid conjugate gradient method is proposed.... Conjugate gradient method is one of successful methods for solving the unconstrained optimization problems. In this paper, absorbing the advantages of FR and CD methods, a hybrid conjugate gradient method is proposed. Under the general Wolfe linear searches, the proposed method can generate the sufficient descent direction at each iterate,and its global convergence property also can be established. Some preliminary numerical results show that the proposed method is effective and stable for the given test problems. 展开更多
关键词 conjugate gradient method general Wolfe linear search SUFFICIENT DESCENT condition global CONVERGENCE
下载PDF
PRP-Type Direct Search Methods for Unconstrained Optimization
16
作者 Qunfeng Liu Wanyou Cheng 《Applied Mathematics》 2011年第6期725-731,共7页
Three PRP-type direct search methods for unconstrained optimization are presented. The methods adopt three kinds of recently developed descent conjugate gradient methods and the idea of frame-based direct search metho... Three PRP-type direct search methods for unconstrained optimization are presented. The methods adopt three kinds of recently developed descent conjugate gradient methods and the idea of frame-based direct search method. Global convergence is shown for continuously differentiable functions. Data profile and performance profile are adopted to analyze the numerical experiments and the results show that the proposed methods are effective. 展开更多
关键词 Direct search methods DESCENT conjugate gradient methods Frame-Based methods Global Convergence Data PROFILE Performance PROFILE
下载PDF
A hybrid conjugate gradient method for optimization problems
17
作者 Xiangrong Li Xupei Zhao 《Natural Science》 2011年第1期85-90,共6页
A hybrid method of the Polak-Ribière-Polyak (PRP) method and the Wei-Yao-Liu (WYL) method is proposed for unconstrained optimization pro- blems, which possesses the following properties: i) This method inherits a... A hybrid method of the Polak-Ribière-Polyak (PRP) method and the Wei-Yao-Liu (WYL) method is proposed for unconstrained optimization pro- blems, which possesses the following properties: i) This method inherits an important property of the well known PRP method: the tendency to turn towards the steepest descent direction if a small step is generated away from the solution, preventing a sequence of tiny steps from happening;ii) The scalar holds automatically;iii) The global convergence with some line search rule is established for nonconvex functions. Numerical results show that the method is effective for the test problems. 展开更多
关键词 line search UNCONSTRAINED Optimization conjugate gradient Method Global CONVERGENCE
下载PDF
一种近似BFGS的自适应双参数共轭梯度法
18
作者 李向利 莫元健 梅建平 《应用数学》 北大核心 2024年第1期89-99,共11页
为了更加有效的求解大规模无约束优化问题,本文基于自调比无记忆BFGS拟牛顿法,提出一个自适应双参数共轭梯度法,设计的搜索方向满足充分下降性,在一般假设和标准Wolfe线搜索准则下,证明该方法具有全局收敛性,数值实验结果证明提出的新... 为了更加有效的求解大规模无约束优化问题,本文基于自调比无记忆BFGS拟牛顿法,提出一个自适应双参数共轭梯度法,设计的搜索方向满足充分下降性,在一般假设和标准Wolfe线搜索准则下,证明该方法具有全局收敛性,数值实验结果证明提出的新算法是有效的. 展开更多
关键词 大规模无约束优化 共轭梯度法 WOLFE线搜索 全局收敛性
下载PDF
一种WYL型谱共轭梯度法的全局收敛性 被引量:1
19
作者 蔡宇 周光辉 《数学物理学报(A辑)》 CSCD 北大核心 2024年第1期173-184,共12页
为解决大规模无约束优化问题,该文结合WYL共轭梯度法和谱共轭梯度法,给出了一种WYL型谱共轭梯度法.在不依赖于任何线搜索的条件下,该方法产生的搜索方向均满足充分下降性,且在强Wolfe线搜索下证明了该方法的全局收敛性.与WYL共轭梯度法... 为解决大规模无约束优化问题,该文结合WYL共轭梯度法和谱共轭梯度法,给出了一种WYL型谱共轭梯度法.在不依赖于任何线搜索的条件下,该方法产生的搜索方向均满足充分下降性,且在强Wolfe线搜索下证明了该方法的全局收敛性.与WYL共轭梯度法的收敛性相比,WYL型谱共轭梯度法推广了线搜索中参数σ的取值范围.最后,相应的数值结果表明了该方法是有效的. 展开更多
关键词 无约束优化 谱共轭梯度法 强Wolfe线搜索 全局收敛性
下载PDF
Conjugate Gradient Methods with Armijo-type Line Searches 被引量:12
20
作者 Yu-Hong DAIState Key Laboratory of Scientific and Engineering Computing, Institute of Computational Mathematics, Academy of Mathematics and System Sciences, Chinese Academy of Sciences, Beijing 100080, China 《Acta Mathematicae Applicatae Sinica》 SCIE CSCD 2002年第1期123-130,共8页
Two Armijo-type line searches are proposed in this paper for nonlinear conjugate gradient methods. Under these line searches, global convergence results are established for several famous conjugate gradient methods, i... Two Armijo-type line searches are proposed in this paper for nonlinear conjugate gradient methods. Under these line searches, global convergence results are established for several famous conjugate gradient methods, including the Fletcher-Reeves method, the Polak-Ribiere-Polyak method, and the conjugate descent method. 展开更多
关键词 Unconstrained optimization conjugate gradient method line search global convergence
全文增补中
上一页 1 2 25 下一页 到第
使用帮助 返回顶部