期刊文献+
共找到2篇文章
< 1 >
每页显示 20 50 100
GLOBAL CONVERGENCE RESULTS OF A THREE TERM MEMORY GRADIENT METHOD WITH A NON-MONOTONE LINE SEARCH TECHNIQUE 被引量:12
1
作者 孙清滢 《Acta Mathematica Scientia》 SCIE CSCD 2005年第1期170-178,共9页
In this paper, a new class of three term memory gradient method with non-monotone line search technique for unconstrained optimization is presented. Global convergence properties of the new methods are discussed. Comb... In this paper, a new class of three term memory gradient method with non-monotone line search technique for unconstrained optimization is presented. Global convergence properties of the new methods are discussed. Combining the quasi-Newton method with the new method, the former is modified to have global convergence property. Numerical results show that the new algorithm is efficient. 展开更多
关键词 Non-linear programming three term memory gradient method convergence non-monotone line search technique numerical experiment
在线阅读 下载PDF
A NEW DESCENT MEMORY GRADIENT METHOD AND ITS GLOBAL CONVERGENCE 被引量:3
2
作者 Min SUN Qingguo BAI 《Journal of Systems Science & Complexity》 SCIE EI CSCD 2011年第4期784-794,共11页
In this article, a new descent memory gradient method without restarts is proposed for solving large scale unconstrained optimization problems. The method has the following attractive properties: 1) The search direc... In this article, a new descent memory gradient method without restarts is proposed for solving large scale unconstrained optimization problems. The method has the following attractive properties: 1) The search direction is always a sufficiently descent direction at every iteration without the line search used; 2) The search direction always satisfies the angle property, which is independent of the convexity of the objective function. Under mild conditions, the authors prove that the proposed method has global convergence, and its convergence rate is also investigated. The numerical results show that the new descent memory method is efficient for the given test problems. 展开更多
关键词 Global convergence memory gradient method sufficiently descent.
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部