84 research outputs found
Accelerated algorithms for linearly constrained convex minimization
νμλ
Όλ¬Έ (λ°μ¬)-- μμΈλνκ΅ λνμ : μ리과νλΆ, 2014. 2. κ°λͺ
μ£Ό.μ ν μ ν 쑰건μ μνμ μ΅μ νλ λ€μν μμ μ²λ¦¬ λ¬Έμ μ λͺ¨λΈλ‘μ μ¬
μ©λκ³ μλ€. μ΄ λ
Όλ¬Έμμλ μ΄ μ ν μ ν 쑰건μ μνμ μ΅μ ν λ¬Έμ λ₯Ό
νκΈ°μν λΉ λ₯Έ μκ³ λ¦¬λ¬λ€μ μκ°νκ³ μ νλ€. μ°λ¦¬κ° μ μνλ λ°©λ²λ€
μ 곡ν΅μ μΌλ‘ Nesterovμ μν΄μ κ°λ°λμλ κ°μνν νλ‘μλ§ κ·Έλ λ
μΈνΈ λ°©λ²μμ μ¬μ©λμλ 보μΈλ²μ κΈ°μ΄λ‘ νκ³ μλ€. μ¬κΈ°μμ μ°λ¦¬λ
ν¬κ²λ³΄μμ λκ°μ§ μκ³ λ¦¬λ¬μ μ μνκ³ μ νλ€. 첫λ²μ§Έ λ°©λ²μ κ°μνν
Bregman λ°©λ²μ΄λ©°, μμΆμΌμ±λ¬Έμ μ μ μ©νμ¬μ μλμ Bregman λ°©λ²λ³΄λ€
κ°μνν λ°©λ²μ΄ λ λΉ λ¦μ νμΈνλ€. λλ²μ§Έ λ°©λ²μ κ°μνν μ΄κ·Έλ¨Όν°λ
λΌκ·Έλμ§μ λ°©λ²μ νμ₯ν κ²μΈλ°, μ΄κ·Έλ¨Όν°λ λΌκ·Έλμ§μ λ°©λ²μ λ΄λΆ
λ¬Έμ λ₯Ό κ°μ§κ³ μκ³ , μ΄λ° λ΄λΆλ¬Έμ λ μΌλ°μ μΌλ‘ μ νν λ΅μ κ³μ°ν μ
μλ€. κ·Έλ κΈ° λλ¬Έμ μ΄λ° λ΄λΆλ¬Έμ λ₯Ό μ λΉν 쑰건μ λ§μ‘±νλλ‘ λΆμ νν
κ² νλλΌλ κ°μνν μ΄κ·Έλ¨Όν°λ λΌκ·Έλμ§ λ°©λ²μ΄ μ ννκ² λ΄λΆλ¬Έμ λ₯Ό
νλμ κ°μ μλ ΄μ±μ κ°λ 쑰건μ μ μνλ€. μ°λ¦¬λ λν κ°μνν μΌν°
λ€μ΄ν
λλ μ
λ°©λ²λ° λν΄μλ λΉμ·ν λ΄μ©μ μ κ°νλ€.Abstract i
1 Introduction 1
2 Previous Methods 5
2.1 Mathematical Preliminary . . . . . . . . . . . . . . . . . . . . 5
2.2 The algorithms for solving the linearly constrained convex
minimization . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
2.2.1 Augmented Lagrangian Method . . . . . . . . . . . . . 8
2.2.2 Bregman Methods . . . . . . . . . . . . . . . . . . . . 9
2.2.3 Alternating direction method of multipliers . . . . . . . 13
2.3 The accelerating algorithms for unconstrained convex minimization problem . . . . . . . . . . . . . . . . . . . . . . . . . 15
2.3.1 Fast inexact iterative shrinkage thresholding algorithm 16
2.3.2 Inexact accelerated proximal point method . . . . . . . 19
3 Proposed Algorithms 23
3.1 Proposed Algorithm 1 : Accelerated Bregman method . . . . . 23
3.1.1 Equivalence to the accelerated augmented Lagrangian
method . . . . . . . . . . . . . . . . . . . . . . . . . . 24
3.1.2 Complexity of the accelerated Bregman method . . . . 27
3.2 Proposed Algorithm 2 : I-AALM . . . . . . . . . . . . . . . . 35
3.3 Proposed Algorithm 3 : I-AADMM . . . . . . . . . . . . . . . 43
3.4 Numerical Results . . . . . . . . . . . . . . . . . . . . . . . . . 54
3.4.1 Comparison to Bregman method with accelerated Bregman method . . . . . . . . . . . . . . . . . . . . . . . . 54
3.4.2 Numerical results of inexact accelerated augmented Lagrangian method using various subproblem solvers . . . 60
3.4.3 Comparison to the inexact accelerated augmented Lagrangian method with other methods . . . . . . . . . . 63
3.4.4 Inexact accelerated alternating direction method of multipliers for Multiplicative Noise Removal . . . . . . . . 69
4 Conclusion 86
Abstract (in Korean) 94Docto
Augmented L1 and Nuclear-Norm Models with a Globally Linearly Convergent Algorithm
This paper studies the long-existing idea of adding a nice smooth function to
"smooth" a non-differentiable objective function in the context of sparse
optimization, in particular, the minimization of
, where is a vector, as well as the
minimization of , where is a matrix and
and are the nuclear and Frobenius norms of ,
respectively. We show that they can efficiently recover sparse vectors and
low-rank matrices. In particular, they enjoy exact and stable recovery
guarantees similar to those known for minimizing and under
the conditions on the sensing operator such as its null-space property,
restricted isometry property, spherical section property, or RIPless property.
To recover a (nearly) sparse vector , minimizing
returns (nearly) the same solution as minimizing
almost whenever . The same relation also
holds between minimizing and minimizing
for recovering a (nearly) low-rank matrix , if . Furthermore, we show that the linearized Bregman algorithm for
minimizing subject to enjoys global
linear convergence as long as a nonzero solution exists, and we give an
explicit rate of convergence. The convergence property does not require a
solution solution or any properties on . To our knowledge, this is the best
known global convergence result for first-order sparse optimization algorithms.Comment: arXiv admin note: text overlap with arXiv:1207.5326 by other author
Fast stochastic dual coordinate descent algorithms for linearly constrained convex optimization
The problem of finding a solution to the linear system with certain
minimization properties arises in numerous scientific and engineering areas. In
the era of big data, the stochastic optimization algorithms become increasingly
significant due to their scalability for problems of unprecedented size. This
paper focuses on the problem of minimizing a strongly convex function subject
to linear constraints. We consider the dual formulation of this problem and
adopt the stochastic coordinate descent to solve it. The proposed algorithmic
framework, called fast stochastic dual coordinate descent, utilizes sampling
matrices sampled from user-defined distributions to extract gradient
information. Moreover, it employs Polyak's heavy ball momentum acceleration
with adaptive parameters learned through iterations, overcoming the limitation
of the heavy ball momentum method that it requires prior knowledge of certain
parameters, such as the singular values of a matrix. With these extensions, the
framework is able to recover many well-known methods in the context, including
the randomized sparse Kaczmarz method, the randomized regularized Kaczmarz
method, the linearized Bregman iteration, and a variant of the conjugate
gradient (CG) method. We prove that, with strongly admissible objective
function, the proposed method converges linearly in expectation. Numerical
experiments are provided to confirm our results.Comment: arXiv admin note: text overlap with arXiv:2305.0548
MAGMA: Multi-level accelerated gradient mirror descent algorithm for large-scale convex composite minimization
Composite convex optimization models arise in several applications, and are
especially prevalent in inverse problems with a sparsity inducing norm and in
general convex optimization with simple constraints. The most widely used
algorithms for convex composite models are accelerated first order methods,
however they can take a large number of iterations to compute an acceptable
solution for large-scale problems. In this paper we propose to speed up first
order methods by taking advantage of the structure present in many applications
and in image processing in particular. Our method is based on multi-level
optimization methods and exploits the fact that many applications that give
rise to large scale models can be modelled using varying degrees of fidelity.
We use Nesterov's acceleration techniques together with the multi-level
approach to achieve convergence rate, where
denotes the desired accuracy. The proposed method has a better
convergence rate than any other existing multi-level method for convex
problems, and in addition has the same rate as accelerated methods, which is
known to be optimal for first-order methods. Moreover, as our numerical
experiments show, on large-scale face recognition problems our algorithm is
several times faster than the state of the art
An Extragradient-Based Alternating Direction Method for Convex Minimization
In this paper, we consider the problem of minimizing the sum of two convex
functions subject to linear linking constraints. The classical alternating
direction type methods usually assume that the two convex functions have
relatively easy proximal mappings. However, many problems arising from
statistics, image processing and other fields have the structure that while one
of the two functions has easy proximal mapping, the other function is smoothly
convex but does not have an easy proximal mapping. Therefore, the classical
alternating direction methods cannot be applied. To deal with the difficulty,
we propose in this paper an alternating direction method based on
extragradients. Under the assumption that the smooth function has a Lipschitz
continuous gradient, we prove that the proposed method returns an
-optimal solution within iterations. We apply the
proposed method to solve a new statistical model called fused logistic
regression. Our numerical experiments show that the proposed method performs
very well when solving the test problems. We also test the performance of the
proposed method through solving the lasso problem arising from statistics and
compare the result with several existing efficient solvers for this problem;
the results are very encouraging indeed
- β¦