7,748 research outputs found
A Quasi-Newton Method for Large Scale Support Vector Machines
This paper adapts a recently developed regularized stochastic version of the
Broyden, Fletcher, Goldfarb, and Shanno (BFGS) quasi-Newton method for the
solution of support vector machine classification problems. The proposed method
is shown to converge almost surely to the optimal classifier at a rate that is
linear in expectation. Numerical results show that the proposed method exhibits
a convergence rate that degrades smoothly with the dimensionality of the
feature vectors.Comment: 5 pages, To appear in International Conference on Acoustics, Speech,
and Signal Processing (ICASSP) 201
Practical Inexact Proximal Quasi-Newton Method with Global Complexity Analysis
Recently several methods were proposed for sparse optimization which make
careful use of second-order information [10, 28, 16, 3] to improve local
convergence rates. These methods construct a composite quadratic approximation
using Hessian information, optimize this approximation using a first-order
method, such as coordinate descent and employ a line search to ensure
sufficient descent. Here we propose a general framework, which includes
slightly modified versions of existing algorithms and also a new algorithm,
which uses limited memory BFGS Hessian approximations, and provide a novel
global convergence rate analysis, which covers methods that solve subproblems
via coordinate descent
Sharpened Lazy Incremental Quasi-Newton Method
We consider the finite sum minimization of strongly convex and smooth
functions with Lipschitz continuous Hessians in dimensions. In many
applications where such problems arise, including maximum likelihood
estimation, empirical risk minimization, and unsupervised learning, the number
of observations is large, and it becomes necessary to use incremental or
stochastic algorithms whose per-iteration complexity is independent of . Of
these, the incremental/stochastic variants of the Newton method exhibit
superlinear convergence, but incur a per-iteration complexity of ,
which may be prohibitive in large-scale settings. On the other hand, the
incremental Quasi-Newton method incurs a per-iteration complexity of
but its superlinear convergence rate has only been characterized
asymptotically. This work puts forth the Sharpened Lazy Incremental
Quasi-Newton (SLIQN) method that achieves the best of both worlds: an explicit
superlinear convergence rate with a per-iteration complexity of .
Building upon the recently proposed Sharpened Quasi-Newton method, the proposed
incremental variant incorporates a hybrid update strategy incorporating both
classic and greedy BFGS updates. The proposed lazy update rule distributes the
computational complexity between the iterations, so as to enable a
per-iteration complexity of . Numerical tests demonstrate the
superiority of SLIQN over all other incremental and stochastic Quasi-Newton
variants.Comment: 39 pages, 3 figure
The algorithms of Broyden-CG for unconstrained optimization problems
The conjugate gradient method plays an important role in solving large-scaled problems and the quasi-Newton method is known as the most efficient method in solving unconstrained optimization problems. Therefore, in this paper, the new hybrid method between the conjugate gradient method and the quasi-newton method for solving optimization problem is suggested. The Broyden family formula is used as an approximation of Hessian in the hybrid method and the quasi-Newton method. Our numerical analysis provides strong evidence that our Broyden-CG method is more efficient than the ordinary Broyden method. Furthermore, we also prove that new algorithm is globally convergent and gratify the sufficient descent condition
Online Equivalence Learning Through A Quasi-Newton Method
International audienceRecently, the community has shown a growing interest in building online learning models. In this paper, we are interested in the framework of fuzzy equivalences obtained by residual implications. Models are generally based on the relevance degree between pairs of objects of the learning set, and the update is obtained by using a standard stochastic (online) gradient descent. This paper proposes another method for learning fuzzy equivalences using a Quasi-Newton optimization. The two methods are extensively compared on real data sets for the task of nearest sample(s) classification
- …