Y Liu and C Storey(1992)proposed the famous LS conjugate gradient method which has good numerical results.However,the LS method has very weak convergence under the Wolfe-type line search.In this paper,we give a new de...Y Liu and C Storey(1992)proposed the famous LS conjugate gradient method which has good numerical results.However,the LS method has very weak convergence under the Wolfe-type line search.In this paper,we give a new descent gradient method based on the LS method.It can guarantee the sufficient descent property at each iteration and the global convergence under the strong Wolfe line search.Finally,we also present extensive preliminary numerical experiments to show the efficiency of the proposed method by comparing with the famous PRP^+method.展开更多
A hybridization of the three–term conjugate gradient method proposed by Zhang et al. and the nonlinear conjugate gradient method proposed by Polak and Ribi`ere, and Polyak is suggested. Based on an eigenvalue analysi...A hybridization of the three–term conjugate gradient method proposed by Zhang et al. and the nonlinear conjugate gradient method proposed by Polak and Ribi`ere, and Polyak is suggested. Based on an eigenvalue analysis, it is shown that search directions of the proposed method satisfy the sufficient descent condition, independent of the line search and the objective function convexity. Global convergence of the method is established under an Armijo–type line search condition. Numerical experiments show practical efficiency of the proposed method.展开更多
The distribution of sampling data influences completeness of rule base so that extrapolating missing rules is very difficult. Based on data mining, a self-learning method is developed for identifying fuzzy model and e...The distribution of sampling data influences completeness of rule base so that extrapolating missing rules is very difficult. Based on data mining, a self-learning method is developed for identifying fuzzy model and extrapolating missing rules, by means of confidence measure and the improved gradient descent method. The proposed approach can not only identify fuzzy model, update its parameters and determine optimal output fuzzy sets simultaneously, but also resolve the uncontrollable problem led by the regions that data do not cover. The simulation results show the effectiveness and accuracy of the proposed approach with the classical truck backer-upper control problem verifying.展开更多
A new algorithm to exploit the learning rates of gradient descent method is presented, based on the second-order Taylor expansion of the error energy function with respect to learning rate, at some values decided by &...A new algorithm to exploit the learning rates of gradient descent method is presented, based on the second-order Taylor expansion of the error energy function with respect to learning rate, at some values decided by "award-punish" strategy. Detailed deduction of the algorithm applied to RBF networks is given. Simulation studies show that this algorithm can increase the rate of convergence and improve the performance of the gradient descent method.展开更多
In this paper, a new nonlinear conjugate gradient method is proposed for large-scale unconstrained optimization. The sufficient descent property holds without any line searches. We use some steplength technique which ...In this paper, a new nonlinear conjugate gradient method is proposed for large-scale unconstrained optimization. The sufficient descent property holds without any line searches. We use some steplength technique which ensures the Zoutendijk condition to be held, this method is proved to be globally convergent. Finally, we improve it, and do further analysis.展开更多
In this paper, three new hybrid nonlinear conjugate gradient methods are presented, which produce suf?cient descent search direction at every iteration. This property is independent of any line search or the convexity...In this paper, three new hybrid nonlinear conjugate gradient methods are presented, which produce suf?cient descent search direction at every iteration. This property is independent of any line search or the convexity of the objective function used. Under suitable conditions, we prove that the proposed methods converge globally for general nonconvex functions. The numerical results show that all these three new hybrid methods are efficient for the given test problems.展开更多
Designing a fuzzy inference system(FIS)from data can be divided into two main phases:structure identification and parameter optimization.First,starting from a simple initial topology,the membership functions and syste...Designing a fuzzy inference system(FIS)from data can be divided into two main phases:structure identification and parameter optimization.First,starting from a simple initial topology,the membership functions and system rules are defined as specific structures.Second,to speed up the convergence of the learning algorithm and lighten the oscillation,an improved descent method for FIS generation is developed.Furthermore, the convergence and the oscillation of the algorithm are system- atically analyzed.Third,using the information obtained from the previous phase,it can be decided in which region of the in- put space the density of fuzzy rules should be enhanced and for which variable the number of fuzzy sets that used to partition the domain must be increased.Consequently,this produces a new and more appropriate structure.Finally,the proposed method is applied to the problem of nonlinear function approximation.展开更多
Three PRP-type direct search methods for unconstrained optimization are presented. The methods adopt three kinds of recently developed descent conjugate gradient methods and the idea of frame-based direct search metho...Three PRP-type direct search methods for unconstrained optimization are presented. The methods adopt three kinds of recently developed descent conjugate gradient methods and the idea of frame-based direct search method. Global convergence is shown for continuously differentiable functions. Data profile and performance profile are adopted to analyze the numerical experiments and the results show that the proposed methods are effective.展开更多
The application of Sobolev gradient methods for finding critical points of the Huxley and Fisher models is demonstrated. A comparison is given between the Euclidean, weighted and unweighted Sobolev gradients. Results ...The application of Sobolev gradient methods for finding critical points of the Huxley and Fisher models is demonstrated. A comparison is given between the Euclidean, weighted and unweighted Sobolev gradients. Results are given for the one dimensional Huxley and Fisher models.展开更多
基金Supported by The Youth Project Foundation of Chongqing Three Gorges University(13QN17)Supported by the Fund of Scientific Research in Southeast University(the Support Project of Fundamental Research)
文摘Y Liu and C Storey(1992)proposed the famous LS conjugate gradient method which has good numerical results.However,the LS method has very weak convergence under the Wolfe-type line search.In this paper,we give a new descent gradient method based on the LS method.It can guarantee the sufficient descent property at each iteration and the global convergence under the strong Wolfe line search.Finally,we also present extensive preliminary numerical experiments to show the efficiency of the proposed method by comparing with the famous PRP^+method.
基金Supported by Research Council of Semnan University
文摘A hybridization of the three–term conjugate gradient method proposed by Zhang et al. and the nonlinear conjugate gradient method proposed by Polak and Ribi`ere, and Polyak is suggested. Based on an eigenvalue analysis, it is shown that search directions of the proposed method satisfy the sufficient descent condition, independent of the line search and the objective function convexity. Global convergence of the method is established under an Armijo–type line search condition. Numerical experiments show practical efficiency of the proposed method.
基金This project was supported by State Science &Technology Pursuing Project (2001BA204B01) of China and Foundation forUniversity Key Teacher by the Ministry of Education of China.
文摘The distribution of sampling data influences completeness of rule base so that extrapolating missing rules is very difficult. Based on data mining, a self-learning method is developed for identifying fuzzy model and extrapolating missing rules, by means of confidence measure and the improved gradient descent method. The proposed approach can not only identify fuzzy model, update its parameters and determine optimal output fuzzy sets simultaneously, but also resolve the uncontrollable problem led by the regions that data do not cover. The simulation results show the effectiveness and accuracy of the proposed approach with the classical truck backer-upper control problem verifying.
基金Open Foundation of State Key Lab of Transmission of Wide-Band FiberTechnologies of Communication Systems
文摘A new algorithm to exploit the learning rates of gradient descent method is presented, based on the second-order Taylor expansion of the error energy function with respect to learning rate, at some values decided by "award-punish" strategy. Detailed deduction of the algorithm applied to RBF networks is given. Simulation studies show that this algorithm can increase the rate of convergence and improve the performance of the gradient descent method.
文摘In this paper, a new nonlinear conjugate gradient method is proposed for large-scale unconstrained optimization. The sufficient descent property holds without any line searches. We use some steplength technique which ensures the Zoutendijk condition to be held, this method is proved to be globally convergent. Finally, we improve it, and do further analysis.
文摘In this paper, three new hybrid nonlinear conjugate gradient methods are presented, which produce suf?cient descent search direction at every iteration. This property is independent of any line search or the convexity of the objective function used. Under suitable conditions, we prove that the proposed methods converge globally for general nonconvex functions. The numerical results show that all these three new hybrid methods are efficient for the given test problems.
基金Supported by National Basic Research Program of China(973 Program)(2007CB714006)
文摘Designing a fuzzy inference system(FIS)from data can be divided into two main phases:structure identification and parameter optimization.First,starting from a simple initial topology,the membership functions and system rules are defined as specific structures.Second,to speed up the convergence of the learning algorithm and lighten the oscillation,an improved descent method for FIS generation is developed.Furthermore, the convergence and the oscillation of the algorithm are system- atically analyzed.Third,using the information obtained from the previous phase,it can be decided in which region of the in- put space the density of fuzzy rules should be enhanced and for which variable the number of fuzzy sets that used to partition the domain must be increased.Consequently,this produces a new and more appropriate structure.Finally,the proposed method is applied to the problem of nonlinear function approximation.
文摘Three PRP-type direct search methods for unconstrained optimization are presented. The methods adopt three kinds of recently developed descent conjugate gradient methods and the idea of frame-based direct search method. Global convergence is shown for continuously differentiable functions. Data profile and performance profile are adopted to analyze the numerical experiments and the results show that the proposed methods are effective.
文摘The application of Sobolev gradient methods for finding critical points of the Huxley and Fisher models is demonstrated. A comparison is given between the Euclidean, weighted and unweighted Sobolev gradients. Results are given for the one dimensional Huxley and Fisher models.