16,641 research outputs found
A Three-Term Conjugate Gradient Method with Sufficient Descent Property for Unconstrained Optimization
Conjugate gradient methods are widely used for solving large-scale unconstrained optimization problems, because they do not need the storage of matrices. In this paper, we propose a general form of three-term conjugate gradient methods which always generate a sufficient descent direction. We give a sufficient condition for the global convergence of the proposed general method. Moreover, we present a specific three-term conjugate gradient method based on the multi-step quasi-Newton method. Finally, some numerical results of the proposed method are given
A new, globally convergent Riemannian conjugate gradient method
This article deals with the conjugate gradient method on a Riemannian
manifold with interest in global convergence analysis. The existing conjugate
gradient algorithms on a manifold endowed with a vector transport need the
assumption that the vector transport does not increase the norm of tangent
vectors, in order to confirm that generated sequences have a global convergence
property. In this article, the notion of a scaled vector transport is
introduced to improve the algorithm so that the generated sequences may have a
global convergence property under a relaxed assumption. In the proposed
algorithm, the transported vector is rescaled in case its norm has increased
during the transport. The global convergence is theoretically proved and
numerically observed with examples. In fact, numerical experiments show that
there exist minimization problems for which the existing algorithm generates
divergent sequences, but the proposed algorithm generates convergent sequences.Comment: 22 pages, 8 figure
- …