期刊文献+
共找到621篇文章
< 1 2 32 >
每页显示 20 50 100
A SQP METHOD FOR MINIMIZING A CLASS OF NONSMOOTH FUNCTIONS
1
作者 孙小玲 张连生 白延琴 《Numerical Mathematics A Journal of Chinese Universities(English Series)》 SCIE 1996年第2期139-146,共8页
In this paper,we present a successive quadratic programming(SQP)method for minimizing a class of nonsmooth functions,which are the sum of a convex function and a nonsmooth composite function.The method generates new i... In this paper,we present a successive quadratic programming(SQP)method for minimizing a class of nonsmooth functions,which are the sum of a convex function and a nonsmooth composite function.The method generates new iterations by using the Armijo-type line search technique after having found the search directions.Global convergence property is established under mild assumptions.Numerical results are also offered. 展开更多
关键词 nonsmooth optimization sqp method global convergence.
在线阅读 下载PDF
A Modified Limited SQP Method For Constrained Optimization
2
作者 Gonglin Yuan Sha Lu Zhengxin Wei 《Applied Mathematics》 2010年第1期8-17,共10页
In this paper, a modified variation of the Limited SQP method is presented for constrained optimization. This method possesses not only the information of gradient but also the information of function value. Moreover,... In this paper, a modified variation of the Limited SQP method is presented for constrained optimization. This method possesses not only the information of gradient but also the information of function value. Moreover, the proposed method requires no more function or derivative evaluations and hardly more storage or arithmetic operations. Under suitable conditions, the global convergence is established. 展开更多
关键词 CONSTRAINED optimization LIMITED method sqp method global convergence
在线阅读 下载PDF
A FILTER-TRUST-REGION METHOD FOR LC^1 UNCONSTRAINED OPTIMIZATION AND ITS GLOBAL CONVERGENCE 被引量:1
3
作者 ZhenghaoYang Wenyu Sun Chuangyin Dang 《Analysis in Theory and Applications》 2008年第1期55-66,共12页
In this paper we present a filter-trust-region algorithm for solving LC1 unconstrained optimization problems which uses the second Dini upper directional derivative. We establish the global convergence of the algorith... In this paper we present a filter-trust-region algorithm for solving LC1 unconstrained optimization problems which uses the second Dini upper directional derivative. We establish the global convergence of the algorithm under reasonable assumptions. 展开更多
关键词 nonsmooth optimization filter method trust region algorithm global conver- gence LC1 optimization
在线阅读 下载PDF
Feasible SQP Descent Method for Inequality Constrained Optimization Problems and Its Convergence 被引量:1
4
作者 张和平 叶留青 《Chinese Quarterly Journal of Mathematics》 CSCD 2009年第3期469-474,共6页
In this paper,the new SQP feasible descent algorithm for nonlinear constrained optimization problems presented,and under weaker conditions of relative,we proofed the new method still possesses global convergence and i... In this paper,the new SQP feasible descent algorithm for nonlinear constrained optimization problems presented,and under weaker conditions of relative,we proofed the new method still possesses global convergence and its strong convergence.The numerical results illustrate that the new methods are valid. 展开更多
关键词 nonlinearly constrained optimization sqp the generalized projection line search global convergence strong convergence.
在线阅读 下载PDF
A CLASS OF TRUST REGION METHODS FOR LINEAR INEQUALITY CONSTRAINED OPTIMIZATION AND ITS THEORY ANALYSIS:I.ALGORITHM AND GLOBAL CONVERGENCEXIU NAIHUA
5
作者 XIU NAIHUA 《Applied Mathematics(A Journal of Chinese Universities)》 SCIE CSCD 1995年第3期287-296,共10页
A class of trust region methods for solving linear inequality constrained problems is proposed in this paper. It is shown that the algorithm is of global convergence.The algorithm uses a version of the two-sided proje... A class of trust region methods for solving linear inequality constrained problems is proposed in this paper. It is shown that the algorithm is of global convergence.The algorithm uses a version of the two-sided projection and the strategy of the unconstrained trust region methods. It keeps the good convergence properties of the unconstrained case and has the merits of the projection method. In some sense, our algorithm can be regarded as an extension and improvement of the projected type algorithm. 展开更多
关键词 Linear inequality constrained optimization trust region method global convergence
在线阅读 下载PDF
Global Convergence of Curve Search Methods for Unconstrained Optimization
6
作者 Zhiwei Xu Yongning Tang Zhen-Jun Shi 《Applied Mathematics》 2016年第7期721-735,共15页
In this paper we propose a new family of curve search methods for unconstrained optimization problems, which are based on searching a new iterate along a curve through the current iterate at each iteration, while line... In this paper we propose a new family of curve search methods for unconstrained optimization problems, which are based on searching a new iterate along a curve through the current iterate at each iteration, while line search methods are based on finding a new iterate on a line starting from the current iterate at each iteration. The global convergence and linear convergence rate of these curve search methods are investigated under some mild conditions. Numerical results show that some curve search methods are stable and effective in solving some large scale minimization problems. 展开更多
关键词 Unconstrained optimization Curve Search method global convergence convergence Rate
在线阅读 下载PDF
A ROBUST SQP METHOD FOR OPTIMIZATION WITH INEQUALITY CONSTRAINTS 被引量:3
7
作者 JuliangZhang XiangsunZhang 《Journal of Computational Mathematics》 SCIE CSCD 2003年第2期247-256,共10页
A new algorithm for inequality constrained optimization is presented, which solves a linear programming subproblem and a quadratic subproblem at each iteration. The algorithm can circumvent the difficulties associated... A new algorithm for inequality constrained optimization is presented, which solves a linear programming subproblem and a quadratic subproblem at each iteration. The algorithm can circumvent the difficulties associated with the possible inconsistency of QP subproblem of the original SQP method. Moreover, the algorithm can converge to a point which satisfies a certain first-order necessary condition even if the original problem is itself infeasible. Under certain condition, some global convergence results are proved and local superlinear convergence results are also obtained. Preliminary numerical results are reported. 展开更多
关键词 nonlinear optimization sqp method global convergence superlinear conver- gence.
原文传递
A ROBUST SQP METHOD BASED ON A SMOOTHING APPROXIMATE PENALTY FUNCTION FOR INEQUALITY CONSTRAINED OPTIMIZATION 被引量:1
8
作者 ZHANGJuliang ZHANGXiangsun 《Journal of Systems Science & Complexity》 SCIE EI CSCD 2002年第1期102-112,共11页
A robust SQP method, which is analogous to Facchinei’s algorithm, is introduced. The algorithm is globally convergent. It uses automatic rules for choosing penalty parameter, and can efficiently cope with the possibl... A robust SQP method, which is analogous to Facchinei’s algorithm, is introduced. The algorithm is globally convergent. It uses automatic rules for choosing penalty parameter, and can efficiently cope with the possible inconsistency of the quadratic search subproblem. In addition, the algorithm employs a differentiable approximate exact penalty function as a merit function. Unlike the merit function in Facchinei’s algorithm, which is quite complicated and is not easy to be implemented in practice, this new merit function is very simple. As a result, we can use the Facchinei’s idea to construct an algorithm which is easy to be implemented in practice. 展开更多
关键词 sqp method global convergence inequality constrained optimization approximate differentiable exact penalty regularity condition.
原文传递
GLOBAL CONVERGENCE OF A CLASS OF OPTIMALLY CONDITIONED SSVM METHODS
9
作者 杨正方 夏爱生 +1 位作者 韩立兴 刘光辉 《Transactions of Tianjin University》 EI CAS 1997年第1期73-76,共4页
This paper explores the convergence of a class of optimally conditioned self scaling variable metric (OCSSVM) methods for unconstrained optimization. We show that this class of methods with Wolfe line search are glob... This paper explores the convergence of a class of optimally conditioned self scaling variable metric (OCSSVM) methods for unconstrained optimization. We show that this class of methods with Wolfe line search are globally convergent for general convex functions. 展开更多
关键词 optimally conditioned self scaling variable metric methods global convergence unconstrained optimization
在线阅读 下载PDF
GLOBAL CONVERGENCE OF TRUST REGION ALGORITHM FOR EQUALITY AND BOUND CONSTRAINED NONLINEAR OPTIMIZATION
10
作者 TongXiaojiao ZhouShuzi 《Applied Mathematics(A Journal of Chinese Universities)》 SCIE CSCD 2003年第1期83-94,共12页
This paper presents a trust region two phase model algorithm for solving the equality and bound constrained nonlinear optimization problem. A concept of substationary point is given. Under suitable assumptions,the gl... This paper presents a trust region two phase model algorithm for solving the equality and bound constrained nonlinear optimization problem. A concept of substationary point is given. Under suitable assumptions,the global convergence of this algorithm is proved without assuming the linear independence of the gradient of active constraints. A numerical example is also presented. 展开更多
关键词 nonlinear optimization equality and bound constrained problem trust-region method global convergence.
在线阅读 下载PDF
A SQP Method for Inequality Constrained Optimization 被引量:5
11
作者 Ju-liang ZHANG, Xiang-sun ZHANGInstitute of Applied Mathematics, Academy of Mathematics and System Sciences, Chinese Academy of Sciences, Beijing 100080, China 《Acta Mathematicae Applicatae Sinica》 SCIE CSCD 2002年第1期77-84,共8页
In this paper, a new SQP method for inequality constrained optimization is proposed and the global convergence is obtained under very mild conditions.
关键词 sqp method global convergence inequality constrained optimization nondifferentiable exact penalty function
全文增补中
AN SQP METHOD BASED ON SMOOTHING PENALTY FUNCTION FOR NONLINEAR OPTIMIZATION WITH INEQUALITY CONSTRAINT 被引量:4
12
作者 ZHANG Juliang ZHANG Xiangsun (Institute of Applied Mathematics, Academy of Mathematics and Systems Science, Chinese Academy of Sciences, Beijing 100080, China) 《Journal of Systems Science & Complexity》 SCIE EI CSCD 2001年第2期212-217,共6页
In this paper, we use the smoothing penalty function proposed in [1] as the merit function of SQP method for nonlinear optimization with inequality constraints. The global convergence of the method is obtained.
关键词 sqp method global convergence INEQUALITY CONSTRAINED optimization SMOOTHING PENALTY function.
原文传递
GLOBAL COVERGENCE OF THE NON-QUASI-NEWTON METHOD FOR UNCONSTRAINED OPTIMIZATION PROBLEMS 被引量:6
13
作者 Liu Hongwei Wang Mingjie +1 位作者 Li Jinshan Zhang Xiangsun 《Applied Mathematics(A Journal of Chinese Universities)》 SCIE CSCD 2006年第3期276-288,共13页
In this paper, the non-quasi-Newton's family with inexact line search applied to unconstrained optimization problems is studied. A new update formula for non-quasi-Newton's family is proposed. It is proved that the ... In this paper, the non-quasi-Newton's family with inexact line search applied to unconstrained optimization problems is studied. A new update formula for non-quasi-Newton's family is proposed. It is proved that the constituted algorithm with either Wolfe-type or Armijotype line search converges globally and Q-superlinearly if the function to be minimized has Lipschitz continuous gradient. 展开更多
关键词 non-quasi-Newton method inexact line search global convergence unconstrained optimization superlinear convergence.
在线阅读 下载PDF
A New Two-Parameter Family of Nonlinear Conjugate Gradient Method Without Line Search for Unconstrained Optimization Problem
14
作者 ZHU Tiefeng 《Wuhan University Journal of Natural Sciences》 CAS CSCD 2024年第5期403-411,共9页
This paper puts forward a two-parameter family of nonlinear conjugate gradient(CG)method without line search for solving unconstrained optimization problem.The main feature of this method is that it does not rely on a... This paper puts forward a two-parameter family of nonlinear conjugate gradient(CG)method without line search for solving unconstrained optimization problem.The main feature of this method is that it does not rely on any line search and only requires a simple step size formula to always generate a sufficient descent direction.Under certain assumptions,the proposed method is proved to possess global convergence.Finally,our method is compared with other potential methods.A large number of numerical experiments show that our method is more competitive and effective. 展开更多
关键词 unconstrained optimization conjugate gradient method without line search global convergence
原文传递
A GLOBALLY AND SUPERLINEARLY CONVERGENT TRUST REGION METHOD FOR LC^1 OPTIMIZATION PROBLEMS 被引量:1
15
作者 Zhang Liping Lai Yanlian Institute of Applied Mathematics,Academia Sinica,Beijing 100080. 《Applied Mathematics(A Journal of Chinese Universities)》 SCIE CSCD 2001年第1期72-80,共9页
A new trust region algorithm for solving convex LC 1 optimization problem is presented.It is proved that the algorithm is globally convergent and the rate of convergence is superlinear under some reasonable assum... A new trust region algorithm for solving convex LC 1 optimization problem is presented.It is proved that the algorithm is globally convergent and the rate of convergence is superlinear under some reasonable assumptions. 展开更多
关键词 LC 1 optimization problem global and superlinear convergence trust region method.
在线阅读 下载PDF
Global Convergence of an Extended Descent Algorithm without Line Search for Unconstrained Optimization 被引量:1
16
作者 Cuiling Chen Liling Luo +1 位作者 Caihong Han Yu Chen 《Journal of Applied Mathematics and Physics》 2018年第1期130-137,共8页
In this paper, we extend a descent algorithm without line search for solving unconstrained optimization problems. Under mild conditions, its global convergence is established. Further, we generalize the search directi... In this paper, we extend a descent algorithm without line search for solving unconstrained optimization problems. Under mild conditions, its global convergence is established. Further, we generalize the search direction to more general form, and also obtain the global convergence of corresponding algorithm. The numerical results illustrate that the new algorithm is effective. 展开更多
关键词 UNCONSTRAINED optimization DESCENT method Line SEARCH global convergence
在线阅读 下载PDF
A New Class of Nonlinear Conjugate Gradient Methods with Global Convergence Properties 被引量:1
17
作者 CHEN Zhong 《长江大学学报(自科版)(上旬)》 CAS 2014年第3期I0001-I0003,共3页
Nonlinear conjugate gradient methods have played an important role in solving large scale unconstrained optimi-zation problems,it is characterized by the simplicity of their iteration and their low memory requirements... Nonlinear conjugate gradient methods have played an important role in solving large scale unconstrained optimi-zation problems,it is characterized by the simplicity of their iteration and their low memory requirements.It is well-known that the direction generated by a conjugate gradient method may be not a descent direction.In this paper,a new class of nonlinear conjugate gradient method is presented,its search direction is a descent direction for the objective function.If the objective function is differentiable and its gradient is Lipschitz continuous,the line sbarch satisfies strong Wolfe condition,the global convergence result is established. 展开更多
关键词 conjugate gradient method line search global convergence unconstrained optimization
在线阅读 下载PDF
A Modified PRP-HS Hybrid Conjugate Gradient Algorithm for Solving Unconstrained Optimization Problems
18
作者 LI Xiangli WANG Zhiling LI Binglan 《应用数学》 北大核心 2025年第2期553-564,共12页
In this paper,we propose a three-term conjugate gradient method for solving unconstrained optimization problems based on the Hestenes-Stiefel(HS)conjugate gradient method and Polak-Ribiere-Polyak(PRP)conjugate gradien... In this paper,we propose a three-term conjugate gradient method for solving unconstrained optimization problems based on the Hestenes-Stiefel(HS)conjugate gradient method and Polak-Ribiere-Polyak(PRP)conjugate gradient method.Under the condition of standard Wolfe line search,the proposed search direction is the descent direction.For general nonlinear functions,the method is globally convergent.Finally,numerical results show that the proposed method is efficient. 展开更多
关键词 Conjugate gradient method Unconstrained optimization Sufficient descent condition global convergence
在线阅读 下载PDF
A Descent Gradient Method and Its Global Convergence
19
作者 LIU Jin-kui 《Chinese Quarterly Journal of Mathematics》 CSCD 2014年第1期142-150,共9页
Y Liu and C Storey(1992)proposed the famous LS conjugate gradient method which has good numerical results.However,the LS method has very weak convergence under the Wolfe-type line search.In this paper,we give a new de... Y Liu and C Storey(1992)proposed the famous LS conjugate gradient method which has good numerical results.However,the LS method has very weak convergence under the Wolfe-type line search.In this paper,we give a new descent gradient method based on the LS method.It can guarantee the sufficient descent property at each iteration and the global convergence under the strong Wolfe line search.Finally,we also present extensive preliminary numerical experiments to show the efficiency of the proposed method by comparing with the famous PRP^+method. 展开更多
关键词 unconstrained optimization conjugate gradient method strong Wolfe line search sufficient descent property global convergence
在线阅读 下载PDF
The global convergence of the non-quasi-Newton methods with non-monotone line search
20
作者 焦宝聪 刘洪伟 《Journal of Harbin Institute of Technology(New Series)》 EI CAS 2006年第6期758-762,共5页
The non-quasi-Newton methods for unconstrained optimization was investigated. Non-monotone line search procedure is introduced, which is combined with the non-quasi-Newton family. Under the uniform convexity assumptio... The non-quasi-Newton methods for unconstrained optimization was investigated. Non-monotone line search procedure is introduced, which is combined with the non-quasi-Newton family. Under the uniform convexity assumption on objective function, the global convergence of the non-quasi-Newton family was proved. Numerical experiments showed that the non-monotone line search was more effective. 展开更多
关键词 non-quasi-Newton method non-monotone line search global convergence unconstrained optimization
在线阅读 下载PDF
上一页 1 2 32 下一页 到第
使用帮助 返回顶部