期刊文献+
共找到69篇文章
< 1 2 4 >
每页显示 20 50 100
GLOBAL COVERGENCE OF THE NON-QUASI-NEWTON METHOD FOR UNCONSTRAINED OPTIMIZATION PROBLEMS 被引量:6
1
作者 Liu Hongwei Wang Mingjie +1 位作者 Li Jinshan Zhang Xiangsun 《Applied Mathematics(A Journal of Chinese Universities)》 SCIE CSCD 2006年第3期276-288,共13页
In this paper, the non-quasi-Newton's family with inexact line search applied to unconstrained optimization problems is studied. A new update formula for non-quasi-Newton's family is proposed. It is proved that the ... In this paper, the non-quasi-Newton's family with inexact line search applied to unconstrained optimization problems is studied. A new update formula for non-quasi-Newton's family is proposed. It is proved that the constituted algorithm with either Wolfe-type or Armijotype line search converges globally and Q-superlinearly if the function to be minimized has Lipschitz continuous gradient. 展开更多
关键词 non-quasi-Newton method inexact line search global convergence unconstrained optimization superlinear convergence.
在线阅读 下载PDF
Bayesian network learning algorithm based on unconstrained optimization and ant colony optimization 被引量:3
2
作者 Chunfeng Wang Sanyang Liu Mingmin Zhu 《Journal of Systems Engineering and Electronics》 SCIE EI CSCD 2012年第5期784-790,共7页
Structure learning of Bayesian networks is a wellresearched but computationally hard task.For learning Bayesian networks,this paper proposes an improved algorithm based on unconstrained optimization and ant colony opt... Structure learning of Bayesian networks is a wellresearched but computationally hard task.For learning Bayesian networks,this paper proposes an improved algorithm based on unconstrained optimization and ant colony optimization(U-ACO-B) to solve the drawbacks of the ant colony optimization(ACO-B).In this algorithm,firstly,an unconstrained optimization problem is solved to obtain an undirected skeleton,and then the ACO algorithm is used to orientate the edges,thus returning the final structure.In the experimental part of the paper,we compare the performance of the proposed algorithm with ACO-B algorithm.The experimental results show that our method is effective and greatly enhance convergence speed than ACO-B algorithm. 展开更多
关键词 Bayesian network structure learning ant colony optimization unconstrained optimization
在线阅读 下载PDF
A DERIVATIVE-FREE ALGORITHM FOR UNCONSTRAINED OPTIMIZATION 被引量:1
3
作者 Peng Yehui Liu Zhenhai 《Applied Mathematics(A Journal of Chinese Universities)》 SCIE CSCD 2005年第4期491-498,共8页
In this paper a hybrid algorithm which combines the pattern search method and the genetic algorithm for unconstrained optimization is presented. The algorithm is a deterministic pattern search algorithm,but in the sea... In this paper a hybrid algorithm which combines the pattern search method and the genetic algorithm for unconstrained optimization is presented. The algorithm is a deterministic pattern search algorithm,but in the search step of pattern search algorithm,the trial points are produced by a way like the genetic algorithm. At each iterate, by reduplication,crossover and mutation, a finite set of points can be used. In theory,the algorithm is globally convergent. The most stir is the numerical results showing that it can find the global minimizer for some problems ,which other pattern search algorithms don't bear. 展开更多
关键词 unconstrained optimization pattern search method genetic algorithm global minimizer.
在线阅读 下载PDF
A New Nonlinear Conjugate Gradient Method for Unconstrained Optimization Problems 被引量:1
4
作者 LIU Jin-kui WANG Kai-rong +1 位作者 SONG Xiao-qian DU Xiang-lin 《Chinese Quarterly Journal of Mathematics》 CSCD 2010年第3期444-450,共7页
In this paper,an efficient conjugate gradient method is given to solve the general unconstrained optimization problems,which can guarantee the sufficient descent property and the global convergence with the strong Wol... In this paper,an efficient conjugate gradient method is given to solve the general unconstrained optimization problems,which can guarantee the sufficient descent property and the global convergence with the strong Wolfe line search conditions.Numerical results show that the new method is efficient and stationary by comparing with PRP+ method,so it can be widely used in scientific computation. 展开更多
关键词 unconstrained optimization conjugate gradient method strong Wolfe line search sufficient descent property global convergence
在线阅读 下载PDF
Subspace Minimization Conjugate Gradient Method Based on Cubic Regularization Model for Unconstrained Optimization 被引量:1
5
作者 Ting Zhao Hongwei Liu 《Journal of Harbin Institute of Technology(New Series)》 CAS 2021年第5期61-69,共9页
Many methods have been put forward to solve unconstrained optimization problems,among which conjugate gradient method(CG)is very important.With the increasing emergence of large⁃scale problems,the subspace technology ... Many methods have been put forward to solve unconstrained optimization problems,among which conjugate gradient method(CG)is very important.With the increasing emergence of large⁃scale problems,the subspace technology has become particularly important and widely used in the field of optimization.In this study,a new CG method was put forward,which combined subspace technology and a cubic regularization model.Besides,a special scaled norm in a cubic regularization model was analyzed.Under certain conditions,some significant characteristics of the search direction were given and the convergence of the algorithm was built.Numerical comparisons show that for the 145 test functions under the CUTEr library,the proposed method is better than two classical CG methods and two new subspaces conjugate gradient methods. 展开更多
关键词 cubic regularization model conjugate gradient method subspace technique unconstrained optimization
在线阅读 下载PDF
New type of conjugate gradient algorithms for unconstrained optimization problems
6
作者 Caiying Wu Guoqing Chen 《Journal of Systems Engineering and Electronics》 SCIE EI CSCD 2010年第6期1000-1007,共8页
Two new formulaes of the main parameter βk of the conjugate gradient method are presented, which respectively can be seen as the modifications of method HS and PRP. In comparison with classic conjugate gradient metho... Two new formulaes of the main parameter βk of the conjugate gradient method are presented, which respectively can be seen as the modifications of method HS and PRP. In comparison with classic conjugate gradient methods, the new methods take both available gradient and function value information. Furthermore, their modifications are proposed. These methods are shown to be global convergent under some assumptions. Numerical results are also reported. 展开更多
关键词 conjugate gradient unconstrained optimization global convergence conjugacy condition.
在线阅读 下载PDF
An Improved Quasi-Newton Method for Unconstrained Optimization
7
作者 Fei Pusheng Chen Zhong (Department of Mathematics, Wuhan University, Wuhan 430072, China) 《Wuhan University Journal of Natural Sciences》 CAS 1996年第1期35-37,共3页
We present an improved method. If we assume that the objective function is twice continuously differentiable and uniformly convex, we discuss global and superlinear convergence of the improved quasi-Newton method.
关键词 quasi-Newton method superlinear convergence unconstrained optimization
在线阅读 下载PDF
Global Convergence of Curve Search Methods for Unconstrained Optimization
8
作者 Zhiwei Xu Yongning Tang Zhen-Jun Shi 《Applied Mathematics》 2016年第7期721-735,共15页
In this paper we propose a new family of curve search methods for unconstrained optimization problems, which are based on searching a new iterate along a curve through the current iterate at each iteration, while line... In this paper we propose a new family of curve search methods for unconstrained optimization problems, which are based on searching a new iterate along a curve through the current iterate at each iteration, while line search methods are based on finding a new iterate on a line starting from the current iterate at each iteration. The global convergence and linear convergence rate of these curve search methods are investigated under some mild conditions. Numerical results show that some curve search methods are stable and effective in solving some large scale minimization problems. 展开更多
关键词 unconstrained optimization Curve Search Method Global Convergence Convergence Rate
在线阅读 下载PDF
Cyclic Gradient Methods for Unconstrained Optimization
9
作者 Ya Zhang Cong Sun 《Journal of the Operations Research Society of China》 EI CSCD 2024年第3期809-828,共20页
Gradient method is popular for solving large-scale problems.In this work,the cyclic gradient methods for quadratic function minimization are extended to general smooth unconstrained optimization problems.Combining wit... Gradient method is popular for solving large-scale problems.In this work,the cyclic gradient methods for quadratic function minimization are extended to general smooth unconstrained optimization problems.Combining with nonmonotonic line search,we prove its global convergence.Furthermore,the proposed algorithms have sublinear convergence rate for general convex functions,and R-linear convergence rate for strongly convex problems.Numerical experiments show that the proposed methods are effective compared to the state of the arts. 展开更多
关键词 Gradient method unconstrained optimization Nonmonotonic line search Global convergence
原文传递
A New Two-Parameter Family of Nonlinear Conjugate Gradient Method Without Line Search for Unconstrained Optimization Problem
10
作者 ZHU Tiefeng 《Wuhan University Journal of Natural Sciences》 CAS CSCD 2024年第5期403-411,共9页
This paper puts forward a two-parameter family of nonlinear conjugate gradient(CG)method without line search for solving unconstrained optimization problem.The main feature of this method is that it does not rely on a... This paper puts forward a two-parameter family of nonlinear conjugate gradient(CG)method without line search for solving unconstrained optimization problem.The main feature of this method is that it does not rely on any line search and only requires a simple step size formula to always generate a sufficient descent direction.Under certain assumptions,the proposed method is proved to possess global convergence.Finally,our method is compared with other potential methods.A large number of numerical experiments show that our method is more competitive and effective. 展开更多
关键词 unconstrained optimization conjugate gradient method without line search global convergence
原文传递
A New Huang Class and Its Properties for Unconstrained Optimization Problems 被引量:1
11
作者 韦增欣 李桥兴 《Journal of Mathematical Research and Exposition》 CSCD 北大核心 2005年第1期64-71,共8页
This paper presents a new class of quasi-Newton methods for solving unconstrained minimization problems. The methods can be regarded as a generalization of Huang class of quasi-Newton methods. We prove that the direct... This paper presents a new class of quasi-Newton methods for solving unconstrained minimization problems. The methods can be regarded as a generalization of Huang class of quasi-Newton methods. We prove that the directions and the iterations generated by the methods of the new class depend only on the parameter p if the exact line searches are made in each steps. 展开更多
关键词 unconstrained optimization quasi-Newton equation quasi-Newton method
在线阅读 下载PDF
A NONMONOTONE CONJUGATE GRADIENT ALGORITHM FOR UNCONSTRAINED OPTIMIZATION 被引量:30
12
《Journal of Systems Science & Complexity》 SCIE EI CSCD 2002年第2期139-145,共7页
Abstract. Conjugate gradient methods are very important methods for unconstrainedoptimization, especially for large scale problems. In this paper, we propose a new conjugategradient method, in which the technique of n... Abstract. Conjugate gradient methods are very important methods for unconstrainedoptimization, especially for large scale problems. In this paper, we propose a new conjugategradient method, in which the technique of nonmonotone line search is used. Under mildassumptions, we prove the global convergence of the method. Some numerical results arealso presented. 展开更多
关键词 unconstrained optimization conjugate gradient nonmonotone line search global convergence.
原文传递
CURVILINEAR PATHS AND TRUST REGION METHODS WITH NONMONOTONIC BACK TRACKING TECHNIQUE FOR UNCONSTRAINED OPTIMIZATION 被引量:26
13
作者 De-tong Zhu (Department of Mathematics, Shanghai Normal University, Shanghai 200234, China) 《Journal of Computational Mathematics》 SCIE EI CSCD 2001年第3期241-258,共18页
Focuses on a study which examined the modification of type approximate trust region methods via two curvilinear paths for unconstrained optimization. Properties of the curvilinear paths; Description of a method which ... Focuses on a study which examined the modification of type approximate trust region methods via two curvilinear paths for unconstrained optimization. Properties of the curvilinear paths; Description of a method which combines line search technique with an approximate trust region algorithm; Information on the convergence analysis; Details on the numerical experiments. 展开更多
关键词 curvilinear paths trust region methods nonmonotonic technique unconstrained optimization
原文传递
TESTING DIFFERENT CONJUGATE GRADIENT METHODS FOR LARGE-SCALE UNCONSTRAINED OPTIMIZATION 被引量:10
14
作者 Yu-hongDai QinNi 《Journal of Computational Mathematics》 SCIE CSCD 2003年第3期311-320,共10页
In this paper we test different conjugate gradient (CG) methods for solving large-scale unconstrained optimization problems. The methods are divided in two groups: the first group includes five basic CG methods and th... In this paper we test different conjugate gradient (CG) methods for solving large-scale unconstrained optimization problems. The methods are divided in two groups: the first group includes five basic CG methods and the second five hybrid CG methods. A collection of medium-scale and large-scale test problems are drawn from a standard code of test problems, CUTE. The conjugate gradient methods are ranked according to the numerical results. Some remarks are given. 展开更多
关键词 Conjugate gradient methods LARGE-SCALE unconstrained optimization Numerical tests.
原文传递
NON-QUASI-NEWTON UPDATES FOR UNCONSTRAINED OPTIMIZATION 被引量:25
15
作者 YUAN, YX BYRD, RH 《Journal of Computational Mathematics》 SCIE CSCD 1995年第2期95-107,共13页
In this report we present some new numerical methods for unconstrained optimization. These methods apply update formulae that do not satisfy the quasi-Newton equation. We derive these new formulae by considering diffe... In this report we present some new numerical methods for unconstrained optimization. These methods apply update formulae that do not satisfy the quasi-Newton equation. We derive these new formulae by considering different techniques of approximating the objective function. Theoretical analyses are given to show the advantages of using non-quasi-Newton updates. Under mild conditions we prove that our new update formulae preserve global convergence properties. Numerical results are also presented. 展开更多
关键词 NON-QUASI-NEWTON UPDATES FOR unconstrained optimization TH
原文传递
An improved trust region method for unconstrained optimization 被引量:5
16
作者 ZHOU QingHua ZHANG YaRui +2 位作者 XU FengXia GENG Yan SUN XiaoDian 《Science China Mathematics》 SCIE 2013年第2期425-434,共10页
In this paper,we propose an improved trust region method for solving unconstrained optimization problems.Different with traditional trust region methods,our algorithm does not resolve the subproblem within the trust r... In this paper,we propose an improved trust region method for solving unconstrained optimization problems.Different with traditional trust region methods,our algorithm does not resolve the subproblem within the trust region centered at the current iteration point,but within an improved one centered at some point located in the direction of the negative gradient,while the current iteration point is on the boundary set.We prove the global convergence properties of the new improved trust region algorithm and give the computational results which demonstrate the effectiveness of our algorithm. 展开更多
关键词 unconstrained optimization trust region methods global convergence negative gradient direction ITERATIVE
原文传递
A NEW FAMILY OF TRUST REGION ALGORITHMS FOR UNCONSTRAINED OPTIMIZATION 被引量:5
17
作者 Yuhong Dai Dachuan Xu(State Key Laboratory of Scientific/Engineering Computing, Institute of Computational Mathematicsand Scientific/Engineering Computing, Academy of Mathematics and System Sciences, ChineseAcademy of Sciences, P.O. Box 2719, Beijing 100080, China) 《Journal of Computational Mathematics》 SCIE CSCD 2003年第2期221-228,共8页
Trust region (TR) algorithms are a class of recently developed algorithms for nonlinear optimization. A new family of TR algorithms for unconstrained optimization, which is the extension of the usual TR method, is pre... Trust region (TR) algorithms are a class of recently developed algorithms for nonlinear optimization. A new family of TR algorithms for unconstrained optimization, which is the extension of the usual TR method, is presented in this paper. When the objective function is bounded below and continuously, differentiable, and the norm of the Hesse approximations increases at most linearly with the iteration number, we prove the global convergence of the algorithms. Limited numerical results are reported, which indicate that our new TR algorithm is competitive. 展开更多
关键词 trust region method global convergence quasi-Newton method unconstrained optimization nonlinear programming.
原文传递
A New Restarting Adaptive Trust-Region Method for Unconstrained Optimization 被引量:1
18
作者 Morteza Kimiaei Susan Ghaderi 《Journal of the Operations Research Society of China》 EI CSCD 2017年第4期487-507,共21页
In this paper,we present a new adaptive trust-region method for solving nonlinear unconstrained optimization problems.More precisely,a trust-region radius based on a nonmonotone technique uses an approximation of Hes... In this paper,we present a new adaptive trust-region method for solving nonlinear unconstrained optimization problems.More precisely,a trust-region radius based on a nonmonotone technique uses an approximation of Hessian which is adaptively chosen.We produce a suitable trust-region radius;preserve the global convergence under classical assumptions to the first-order critical points;improve the practical performance of the new algorithm compared to other exiting variants.Moreover,the quadratic convergence rate is established under suitable conditions.Computational results on the CUTEst test collection of unconstrained problems are presented to show the effectiveness of the proposed algorithm compared with some exiting methods. 展开更多
关键词 unconstrained optimization Trust-region methods Nonmonotone technique Adaptive radius Theoretical convergence
原文传递
A QUASI-NEWTON ALGORITHM WITHOUT CALCULATING DERIVATIVES FOR UNCONSTRAINED OPTIMIZATION 被引量:1
19
作者 Sun Lin-ping(Depatheent of Mathematics, Nanjing Universitg, Jiangsu, China) 《Journal of Computational Mathematics》 SCIE CSCD 1994年第4期380-386,共7页
A new algorithm for unconstrained optimization is developed, by using the product form of the OCSSR1 update. The implementation is especially useful when gradient information is estimated by difference formulae. Preli... A new algorithm for unconstrained optimization is developed, by using the product form of the OCSSR1 update. The implementation is especially useful when gradient information is estimated by difference formulae. Preliminary tests show that new algorithm can perform well. 展开更多
关键词 TF LINE A QUASI-NEWTON ALGORITHM WITHOUT CALCULATING DERIVATIVES FOR unconstrained optimization MATH
原文传递
The Substitution Secant/Finite Difference Method for Large Scale Sparse Unconstrained Optimization
20
作者 Hong-wei Zhang Jun-xiang Li 《Acta Mathematicae Applicatae Sinica》 SCIE CSCD 2005年第4期581-596,共16页
This paper studies a substitution secant/finite difference (SSFD) method for solving large scale sparse unconstrained optimization problems. This method is a combination of a secant method and a finite difference me... This paper studies a substitution secant/finite difference (SSFD) method for solving large scale sparse unconstrained optimization problems. This method is a combination of a secant method and a finite difference method, which depends on a consistent partition of the columns of the lower triangular part of the Hessian matrix. A q-superlinear convergence result and an r-convergence rate estimate show that this method has good local convergence properties. The numerical results show that this method may be competitive with some currently used algorithms. 展开更多
关键词 unconstrained optimization SUBSTITUTION HESSIAN SPARSITY secant method
原文传递
上一页 1 2 4 下一页 到第
使用帮助 返回顶部