85 research outputs found

    Divergence of the ADAM algorithm with fixed-stepsize: a (very) simple example

    Full text link
    A very simple unidimensional function with Lipschitz continuous gradient is constructed such that the ADAM algorithm with constant stepsize, started from the origin, diverges when applied to minimize this function in the absence of noise on the gradient. Divergence occurs irrespective of the choice of the method parameters

    Adaptive Regularization for Nonconvex Optimization Using Inexact Function Values and Randomly Perturbed Derivatives

    Get PDF
    A regularization algorithm allowing random noise in derivatives and inexact function values is proposed for computing approximate local critical points of any order for smooth unconstrained optimization problems. For an objective function with Lipschitz continuous pp-th derivative and given an arbitrary optimality order qpq \leq p, it is shown that this algorithm will, in expectation, compute such a point in at most O((minj{1,,q}ϵj)p+1pq+1)O\left(\left(\min_{j\in\{1,\ldots,q\}}\epsilon_j\right)^{-\frac{p+1}{p-q+1}}\right) inexact evaluations of ff and its derivatives whenever q{1,2}q\in\{1,2\}, where ϵj\epsilon_j is the tolerance for jjth order accuracy. This bound becomes at most O((minj{1,,q}ϵj)q(p+1)p)O\left(\left(\min_{j\in\{1,\ldots,q\}}\epsilon_j\right)^{-\frac{q(p+1)}{p}}\right) inexact evaluations if q>2q>2 and all derivatives are Lipschitz continuous. Moreover these bounds are sharp in the order of the accuracy tolerances. An extension to convexly constrained problems is also outlined.Comment: 22 page

    Adaptive Regularization Algorithms with Inexact Evaluations for Nonconvex Optimization

    Get PDF
    A regularization algorithm using inexact function values and inexact derivatives is proposed and its evaluation complexity analyzed. This algorithm is applicable to unconstrained problems and to problems with inexpensive constraints (that is constraints whose evaluation and enforcement has negligible cost) under the assumption that the derivative of highest degree is β\beta-H\"{o}lder continuous. It features a very flexible adaptive mechanism for determining the inexactness which is allowed, at each iteration, when computing objective function values and derivatives. The complexity analysis covers arbitrary optimality order and arbitrary degree of available approximate derivatives. It extends results of Cartis, Gould and Toint (2018) on the evaluation complexity to the inexact case: if a qqth order minimizer is sought using approximations to the first pp derivatives, it is proved that a suitable approximate minimizer within ϵ\epsilon is computed by the proposed algorithm in at most O(ϵp+βpq+β)O(\epsilon^{-\frac{p+\beta}{p-q+\beta}}) iterations and at most O(log(ϵ)ϵp+βpq+β)O(|\log(\epsilon)|\epsilon^{-\frac{p+\beta}{p-q+\beta}}) approximate evaluations. An algorithmic variant, although more rigid in practice, can be proved to find such an approximate minimizer in O(log(ϵ)+ϵp+βpq+β)O(|\log(\epsilon)|+\epsilon^{-\frac{p+\beta}{p-q+\beta}}) evaluations.While the proposed framework remains so far conceptual for high degrees and orders, it is shown to yield simple and computationally realistic inexact methods when specialized to the unconstrained and bound-constrained first- and second-order cases. The deterministic complexity results are finally extended to the stochastic context, yielding adaptive sample-size rules for subsampling methods typical of machine learning.Comment: 32 page

    OFFO minimization algorithms for second-order optimality and their complexity

    Full text link
    An Adagrad-inspired class of algorithms for smooth unconstrained optimization is presented in which the objective function is never evaluated and yet the gradient norms decrease at least as fast as \calO(1/\sqrt{k+1}) while second-order optimality measures converge to zero at least as fast as \calO(1/(k+1)^{1/3}). This latter rate of convergence is shown to be essentially sharp and is identical to that known for more standard algorithms (like trust-region or adaptive-regularization methods) using both function and derivatives' evaluations. A related "divergent stepsize" method is also described, whose essentially sharp rate of convergence is slighly inferior. It is finally discussed how to obtain weaker second-order optimality guarantees at a (much) reduced computional cost

    Multilevel Objective-Function-Free Optimization with an Application to Neural Networks Training

    Get PDF
    A class of multi-level algorithms for unconstrained nonlinear optimization is presented which does not require the evaluation of the objective function. The class contains the momentum-less AdaGrad method as a particular (single-level) instance. The choice of avoiding the evaluation of the objective function is intended to make the algorithms of the class less sensitive to noise, while the multi-level feature aims at reducing their computational cost. The evaluation complexity of these algorithms is analyzed and their behaviour in the presence of noise is then illustrated in the context of training deep neural networks for supervised learning applications.Comment: 29 pages, 4 figure

    An adaptive regularization method in Banach spaces

    Get PDF
    This paper considers optimization of nonconvex functionals in smooth infinite dimensional spaces. It is first proved that functionals in a class containing multivariate polynomials augmented with a sufficiently smooth regularization can be minimized by a simple linesearch-based algorithm. Sufficient smoothness depends on gradients satisfying a novel two-terms generalized Lipschitz condition. A first-order adaptive regularization method applicable to functionals with β-Hölder continuous derivatives is then proposed, that uses the linesearch approach to compute a suitable trial step. It is shown to find an ϵ-approximate first-order point in at most (Formula presented.) evaluations of the functional and its first p derivatives.</p
    corecore