期刊文献+
共找到693篇文章
< 1 2 35 >
每页显示 20 50 100
A Globally Convergent Polak-Ribiere-Polyak Conjugate Gradient Method with Armijo-Type Line Search 被引量:11
1
作者 Gaohang Yu Lutai Guan Zengxin Wei 《Numerical Mathematics A Journal of Chinese Universities(English Series)》 SCIE 2006年第4期357-366,共10页
In this paper, we propose a globally convergent Polak-Ribiere-Polyak (PRP) conjugate gradient method for nonconvex minimization of differentiable functions by employing an Armijo-type line search which is simpler and ... In this paper, we propose a globally convergent Polak-Ribiere-Polyak (PRP) conjugate gradient method for nonconvex minimization of differentiable functions by employing an Armijo-type line search which is simpler and less demanding than those defined in [4,10]. A favorite property of this method is that we can choose the initial stepsize as the one-dimensional minimizer of a quadratic modelΦ(t):= f(xk)+tgkTdk+(1/2) t2dkTQkdk, where Qk is a positive definite matrix that carries some second order information of the objective function f. So, this line search may make the stepsize tk more easily accepted. Preliminary numerical results show that this method is efficient. 展开更多
关键词 非约束最优化 共轭梯度法 整体收敛 可微函数
下载PDF
A Scaled Conjugate Gradient Method Based on New BFGS Secant Equation with Modified Nonmonotone Line Search
2
作者 Tsegay Giday Woldu Haibin Zhang Yemane Hailu Fissuh 《American Journal of Computational Mathematics》 2020年第1期1-22,共22页
In this paper, we provide and analyze a new scaled conjugate gradient method and its performance, based on the modified secant equation of the Broyden-Fletcher-Goldfarb-Shanno (BFGS) method and on a new modified nonmo... In this paper, we provide and analyze a new scaled conjugate gradient method and its performance, based on the modified secant equation of the Broyden-Fletcher-Goldfarb-Shanno (BFGS) method and on a new modified nonmonotone line search technique. The method incorporates the modified BFGS secant equation in an effort to include the second order information of the objective function. The new secant equation has both gradient and function value information, and its update formula inherits the positive definiteness of Hessian approximation for general convex function. In order to improve the likelihood of finding a global optimal solution, we introduce a new modified nonmonotone line search technique. It is shown that, for nonsmooth convex problems, the proposed algorithm is globally convergent. Numerical results show that this new scaled conjugate gradient algorithm is promising and efficient for solving not only convex but also some large scale nonsmooth nonconvex problems in the sense of the Dolan-Moré performance profiles. 展开更多
关键词 conjugate gradient method BFGS method MODIFIED SECANT EQUATION NONMONOTONE line search Nonsmooth Optimization
下载PDF
The global convergence of the non-quasi-Newton methods with non-monotone line search
3
作者 焦宝聪 刘洪伟 《Journal of Harbin Institute of Technology(New Series)》 EI CAS 2006年第6期758-762,共5页
The non-quasi-Newton methods for unconstrained optimization was investigated. Non-monotone line search procedure is introduced, which is combined with the non-quasi-Newton family. Under the uniform convexity assumptio... The non-quasi-Newton methods for unconstrained optimization was investigated. Non-monotone line search procedure is introduced, which is combined with the non-quasi-Newton family. Under the uniform convexity assumption on objective function, the global convergence of the non-quasi-Newton family was proved. Numerical experiments showed that the non-monotone line search was more effective. 展开更多
关键词 凸函数 函数方程 牛顿 函数优化
下载PDF
GLOBAL CONVERGENCE RESULTS OF A THREE TERM MEMORY GRADIENT METHOD WITH A NON-MONOTONE LINE SEARCH TECHNIQUE 被引量:12
4
作者 孙清滢 《Acta Mathematica Scientia》 SCIE CSCD 2005年第1期170-178,共9页
In this paper, a new class of three term memory gradient method with non-monotone line search technique for unconstrained optimization is presented. Global convergence properties of the new methods are discussed. Comb... In this paper, a new class of three term memory gradient method with non-monotone line search technique for unconstrained optimization is presented. Global convergence properties of the new methods are discussed. Combining the quasi-Newton method with the new method, the former is modified to have global convergence property. Numerical results show that the new algorithm is efficient. 展开更多
关键词 Non-linear programming three term memory gradient method convergence non-monotone line search technique numerical experiment
下载PDF
ON THE GLOBAL CONVERGENCE OF CONJUGATE GRADIENT METHODS WITH INEXACT LINESEARCH
5
作者 刘光辉 韩继业 《Numerical Mathematics A Journal of Chinese Universities(English Series)》 SCIE 1995年第2期147-153,共7页
In this paper we consider the global convergence of any conjugate gradient method of the form d1=-g1,dk+1=-gk+1+βkdk(k≥1)with any βk satisfying sume conditions,and with the strong wolfe line search conditions.Under... In this paper we consider the global convergence of any conjugate gradient method of the form d1=-g1,dk+1=-gk+1+βkdk(k≥1)with any βk satisfying sume conditions,and with the strong wolfe line search conditions.Under the convex assumption on the objective function,we preve the descenf property and the global convergence of this method. 展开更多
关键词 conjugate gradient method STRONG Wolfe line search global convergence.
下载PDF
A New Class of Nonlinear Conjugate Gradient Methods with Global Convergence Properties 被引量:1
6
作者 陈忠 《长江大学学报(自科版)(上旬)》 CAS 2014年第3期I0001-I0003,共3页
非线性共轭梯度法由于其迭代简单和储存量小,且搜索方向不需要满足正割条件,在求解大规模无约束优化问题时占据及其重要的地位.提出了一类新的共轭梯度法,其搜索方向是目标函数的下降方向.若假设目标函数连续可微且梯度满足Lipschitz条... 非线性共轭梯度法由于其迭代简单和储存量小,且搜索方向不需要满足正割条件,在求解大规模无约束优化问题时占据及其重要的地位.提出了一类新的共轭梯度法,其搜索方向是目标函数的下降方向.若假设目标函数连续可微且梯度满足Lipschitz条件,线性搜索满足Wolfe原则,讨论了所设计算法的全局收敛性. 展开更多
关键词 摘要 编辑部 编辑工作 读者
下载PDF
GLOBAL CONVERGENCE PROPERTIES OF THREE-TERM CONJUGATE GRADIENT METHOD WITH NEW-TYPE LINE SEARCH 被引量:13
7
作者 WANGChangyu DUShouqiang CHENYuanyuan 《Journal of Systems Science & Complexity》 SCIE EI CSCD 2004年第3期412-420,共9页
Abstract. In this paper, a new Wolfe-type line search and a new Armijo-type line search are proposed, and some global convergence properties of a three-term conjugate gradient method with the two line searches are pro... Abstract. In this paper, a new Wolfe-type line search and a new Armijo-type line search are proposed, and some global convergence properties of a three-term conjugate gradient method with the two line searches are proved. 展开更多
关键词 全局收敛 线性搜索 无约束优化 三期共轭梯度方法 索引子集
原文传递
A Descent Gradient Method and Its Global Convergence
8
作者 LIU Jin-kui 《Chinese Quarterly Journal of Mathematics》 CSCD 2014年第1期142-150,共9页
Y Liu and C Storey(1992)proposed the famous LS conjugate gradient method which has good numerical results.However,the LS method has very weak convergence under the Wolfe-type line search.In this paper,we give a new de... Y Liu and C Storey(1992)proposed the famous LS conjugate gradient method which has good numerical results.However,the LS method has very weak convergence under the Wolfe-type line search.In this paper,we give a new descent gradient method based on the LS method.It can guarantee the sufficient descent property at each iteration and the global convergence under the strong Wolfe line search.Finally,we also present extensive preliminary numerical experiments to show the efficiency of the proposed method by comparing with the famous PRP^+method. 展开更多
关键词 unconstrained optimization conjugate gradient method strong Wolfe line search sufficient descent property global convergence
下载PDF
A New Nonlinear Conjugate Gradient Method for Unconstrained Optimization Problems 被引量:1
9
作者 LIU Jin-kui WANG Kai-rong +1 位作者 SONG Xiao-qian DU Xiang-lin 《Chinese Quarterly Journal of Mathematics》 CSCD 2010年第3期444-450,共7页
在这份报纸,一有效结合坡度方法被给解决一般非强迫的优化问题,它能保证有强壮的沃尔夫线的足够的降下性质和全球集中寻找新方法由与 PRP+method 作比较有效、静止的 conditions.Numerical 结果表演,它能广泛地因此在科学计算被使用。
关键词 非强迫的优化 结合坡度方法 强壮的沃尔夫线搜索 足够的降下性质 全球集中
下载PDF
A New Descent Nonlinear Conjugate Gradient Method for Unconstrained Optimization
10
作者 Hao Fan Zhibin Zhu Anwa Zhou 《Applied Mathematics》 2011年第9期1119-1123,共5页
In this paper, a new nonlinear conjugate gradient method is proposed for large-scale unconstrained optimization. The sufficient descent property holds without any line searches. We use some steplength technique which ... In this paper, a new nonlinear conjugate gradient method is proposed for large-scale unconstrained optimization. The sufficient descent property holds without any line searches. We use some steplength technique which ensures the Zoutendijk condition to be held, this method is proved to be globally convergent. Finally, we improve it, and do further analysis. 展开更多
关键词 Large Scale UNCONSTRAINED Optimization conjugate gradient method SUFFICIENT DESCENT Property globally CONVERGENT
下载PDF
Global Convergence of an Extended Descent Algorithm without Line Search for Unconstrained Optimization
11
作者 Cuiling Chen Liling Luo +1 位作者 Caihong Han Yu Chen 《Journal of Applied Mathematics and Physics》 2018年第1期130-137,共8页
In this paper, we extend a descent algorithm without line search for solving unconstrained optimization problems. Under mild conditions, its global convergence is established. Further, we generalize the search directi... In this paper, we extend a descent algorithm without line search for solving unconstrained optimization problems. Under mild conditions, its global convergence is established. Further, we generalize the search direction to more general form, and also obtain the global convergence of corresponding algorithm. The numerical results illustrate that the new algorithm is effective. 展开更多
关键词 UNCONSTRAINED Optimization DESCENT method line search global convergence
下载PDF
Several New Line Search Methods and Their Convergence
12
作者 Zhenjun Shi Kimberly Kendricks +1 位作者 Zhiwei Xu Yongning Tang 《American Journal of Operations Research》 2013年第5期421-430,共10页
In this paper, we propose several new line search rules for solving unconstrained minimization problems. These new line search rules can extend the accepted scope of step sizes to a wider extent than the corresponding... In this paper, we propose several new line search rules for solving unconstrained minimization problems. These new line search rules can extend the accepted scope of step sizes to a wider extent than the corresponding original ones and give an adequate initial step size at each iteration. It is proved that the resulting line search algorithms have global convergence under some mild conditions. It is also proved that the search direction plays an important role in line search methods and that the step size approaches mainly guarantee global convergence in general cases. The convergence rate of these methods is also investigated. Some numerical results show that these new line search algorithms are effective in practical computation. 展开更多
关键词 UNCONSTRAINED MINIMIZATION line search method global convergence convergence RATE
下载PDF
A NEW CONJUGATE GRADIENT METHOD AND ITS GLOBAL CONVERGENCE PROPERTIES 被引量:2
13
作者 LI Zhengfeng CHEN Jing DENG Naiyang(Division of Basic Sciences, China Agricultural University East Campus, Beijing 100083, China) 《Systems Science and Mathematical Sciences》 SCIE EI CSCD 1998年第1期53-60,共8页
This paper presents a new conjugate gradient method for unconstrained opti-mization. This method reduces to the Polak-Ribiere-Polyak method when line searches areexact. But their performances are differellt in the cas... This paper presents a new conjugate gradient method for unconstrained opti-mization. This method reduces to the Polak-Ribiere-Polyak method when line searches areexact. But their performances are differellt in the case of inexact line search. By a simpleexample, we show that the Wolf e conditions do not ensure that the present method and thePolak- Ribiere- Polyak method will pro duce descent direct i0ns even u nder t h e ass umpt ionthat the objective function is Strictly convex. This result contradicts the F0lk axiom thatthe Polak-Ribiere-Polyak with the Wolf e line search should find the minimizer of a strictlyconvex objective function. Finally, we show that there are two ways to improve the newmethod such that it is globally convergent. 展开更多
关键词 conjugate gradient method global convergence UNCONSTRAINED optimization line searches.
原文传递
A modified three–term conjugate gradient method with sufficient descent property 被引量:1
14
作者 Saman Babaie–Kafaki 《Applied Mathematics(A Journal of Chinese Universities)》 SCIE CSCD 2015年第3期263-272,共10页
A hybridization of the three–term conjugate gradient method proposed by Zhang et al. and the nonlinear conjugate gradient method proposed by Polak and Ribi`ere, and Polyak is suggested. Based on an eigenvalue analysi... A hybridization of the three–term conjugate gradient method proposed by Zhang et al. and the nonlinear conjugate gradient method proposed by Polak and Ribi`ere, and Polyak is suggested. Based on an eigenvalue analysis, it is shown that search directions of the proposed method satisfy the sufficient descent condition, independent of the line search and the objective function convexity. Global convergence of the method is established under an Armijo–type line search condition. Numerical experiments show practical efficiency of the proposed method. 展开更多
关键词 unconstrained optimization conjugate gradient method EIGENVALUE sufficient descent condition global convergence
下载PDF
A hybrid conjugate gradient method for optimization problems
15
作者 Xiangrong Li Xupei Zhao 《Natural Science》 2011年第1期85-90,共6页
A hybrid method of the Polak-Ribière-Polyak (PRP) method and the Wei-Yao-Liu (WYL) method is proposed for unconstrained optimization pro- blems, which possesses the following properties: i) This method inherits a... A hybrid method of the Polak-Ribière-Polyak (PRP) method and the Wei-Yao-Liu (WYL) method is proposed for unconstrained optimization pro- blems, which possesses the following properties: i) This method inherits an important property of the well known PRP method: the tendency to turn towards the steepest descent direction if a small step is generated away from the solution, preventing a sequence of tiny steps from happening;ii) The scalar holds automatically;iii) The global convergence with some line search rule is established for nonconvex functions. Numerical results show that the method is effective for the test problems. 展开更多
关键词 line search UNCONSTRAINED Optimization conjugate gradient method global convergence
下载PDF
A Non-Monotone Trust Region Method with Non-Monotone Wolfe-Type Line Search Strategy for Unconstrained Optimization
16
作者 Changyuan Li Qinghua Zhou Xiao Wu 《Journal of Applied Mathematics and Physics》 2015年第6期707-712,共6页
In this paper, we propose and analyze a non-monotone trust region method with non-monotone line search strategy for unconstrained optimization problems. Unlike the traditional non-monotone trust region method, our alg... In this paper, we propose and analyze a non-monotone trust region method with non-monotone line search strategy for unconstrained optimization problems. Unlike the traditional non-monotone trust region method, our algorithm utilizes non-monotone Wolfe line search to get the next point if a trial step is not adopted. Thus, it can reduce the number of solving sub-problems. Theoretical analysis shows that the new proposed method has a global convergence under some mild conditions. 展开更多
关键词 UNCONSTRAINED Optimization Non-Monotone TRUST Region method Non-Monotone line search global convergence
下载PDF
GLOBAL CONVERGENCE OF THE GENERAL THREE TERM CONJUGATE GRADIENT METHODS WITH THE RELAXED STRONG WOLFE LINE SEARCH
17
作者 Xu Zeshui Yue ZhenjunInstitute of Sciences,PLA University of Science and Technology,Nanjing,210016. 《Applied Mathematics(A Journal of Chinese Universities)》 SCIE CSCD 2001年第1期58-62,共5页
The global convergence of the general three term conjugate gradient methods with the relaxed strong Wolfe line search is proved.
关键词 conjugate gradient method inexact line search global convergence.
下载PDF
The Global Convergence of Self-Scaling BFGS Algorithm with Nonmonotone Line Search for Unconstrained Nonconvex Optimization Problems
18
作者 Hong Xia YIN Dong Lei DU 《Acta Mathematica Sinica,English Series》 SCIE CSCD 2007年第7期1233-1240,共8页
The self-scaling quasi-Newton method solves an unconstrained optimization problem by scaling the Hessian approximation matrix before it is updated at each iteration to avoid the possible large eigenvalues in the Hessi... The self-scaling quasi-Newton method solves an unconstrained optimization problem by scaling the Hessian approximation matrix before it is updated at each iteration to avoid the possible large eigenvalues in the Hessian approximation matrices of the objective function. It has been proved in the literature that this method has the global and superlinear convergence when the objective function is convex (or even uniformly convex). We propose to solve unconstrained nonconvex optimization problems by a self-scaling BFGS algorithm with nonmonotone linear search. Nonmonotone line search has been recognized in numerical practices as a competitive approach for solving large-scale nonlinear problems. We consider two different nonmonotone line search forms and study the global convergence of these nonmonotone self-scale BFGS algorithms. We prove that, under some weaker condition than that in the literature, both forms of the self-scaling BFGS algorithm are globally convergent for unconstrained nonconvex optimization problems. 展开更多
关键词 nonmonotone line search self-scaling BFGS method global convergence
原文传递
A Modified Conjugate Gradient Method with Global Convergence Property 被引量:2
19
作者 HONG Ling MO Li Liu WEI Zeng Xin 《Journal of Mathematical Research and Exposition》 CSCD 2009年第1期65-75,共11页
A new conjugate gradient method is proposed in this paper. For any (inexact) line search, our scheme satifies the sufficient descent property. The method is proved to be globally convergent if the restricted Wolfe-P... A new conjugate gradient method is proposed in this paper. For any (inexact) line search, our scheme satifies the sufficient descent property. The method is proved to be globally convergent if the restricted Wolfe-Powell line search is used. Preliminary numerical result shows that it is efficient. 展开更多
关键词 unconstrained optimization conjugate gradient method Wolfe-Powell line search global convergence.
下载PDF
Conjugate Gradient Methods with Armijo-type Line Searches 被引量:11
20
作者 Yu-Hong DAIState Key Laboratory of Scientific and Engineering Computing, Institute of Computational Mathematics, Academy of Mathematics and System Sciences, Chinese Academy of Sciences, Beijing 100080, China 《Acta Mathematicae Applicatae Sinica》 SCIE CSCD 2002年第1期123-130,共8页
Two Armijo-type line searches are proposed in this paper for nonlinear conjugate gradient methods. Under these line searches, global convergence results are established for several famous conjugate gradient methods, i... Two Armijo-type line searches are proposed in this paper for nonlinear conjugate gradient methods. Under these line searches, global convergence results are established for several famous conjugate gradient methods, including the Fletcher-Reeves method, the Polak-Ribiere-Polyak method, and the conjugate descent method. 展开更多
关键词 Unconstrained optimization conjugate gradient method line search global convergence
全文增补中
上一页 1 2 35 下一页 到第
使用帮助 返回顶部