84 research outputs found

    Accelerated algorithms for linearly constrained convex minimization

    Get PDF
    ν•™μœ„λ…Όλ¬Έ (박사)-- μ„œμšΈλŒ€ν•™κ΅ λŒ€ν•™μ› : μˆ˜λ¦¬κ³Όν•™λΆ€, 2014. 2. κ°•λͺ…μ£Ό.μ„ ν˜• μ œν•œ 쑰건의 μˆ˜ν•™μ  μ΅œμ ν™”λŠ” λ‹€μ–‘ν•œ μ˜μƒ 처리 문제의 λͺ¨λΈλ‘œμ„œ 사 용되고 μžˆλ‹€. 이 λ…Όλ¬Έμ—μ„œλŠ” 이 μ„ ν˜• μ œν•œ 쑰건의 μˆ˜ν•™μ  μ΅œμ ν™” 문제λ₯Ό ν’€κΈ°μœ„ν•œ λΉ λ₯Έ μ•Œκ³ λ¦¬λ“¬λ“€μ„ μ†Œκ°œν•˜κ³ μž ν•œλ‹€. μš°λ¦¬κ°€ μ œμ•ˆν•˜λŠ” 방법듀 은 κ³΅ν†΅μ μœΌλ‘œ Nesterov에 μ˜ν•΄μ„œ κ°œλ°œλ˜μ—ˆλ˜ κ°€μ†ν™”ν•œ ν”„λ‘μ‹œλ§ κ·Έλ ˆλ”” μ–ΈνŠΈ λ°©λ²•μ—μ„œ μ‚¬μš©λ˜μ—ˆλ˜ 보외법을 기초둜 ν•˜κ³  μžˆλ‹€. μ—¬κΈ°μ—μ„œ μš°λ¦¬λŠ” ν¬κ²Œλ³΄μ•„μ„œ 두가지 μ•Œκ³ λ¦¬λ“¬μ„ μ œμ•ˆν•˜κ³ μž ν•œλ‹€. 첫번째 방법은 κ°€μ†ν™”ν•œ Bregman 방법이며, μ••μΆ•μ„Όμ‹±λ¬Έμ œμ— μ μš©ν•˜μ—¬μ„œ μ›λž˜μ˜ Bregman 방법보닀 κ°€μ†ν™”ν•œ 방법이 더 빠름을 ν™•μΈν•œλ‹€. λ‘λ²ˆμ§Έ 방법은 κ°€μ†ν™”ν•œ μ–΄κ·Έλ¨Όν‹°λ“œ λΌκ·Έλž‘μ§€μ•ˆ 방법을 ν™•μž₯ν•œ 것인데, μ–΄κ·Έλ¨Όν‹°λ“œ λΌκ·Έλž‘μ§€μ•ˆ 방법은 λ‚΄λΆ€ 문제λ₯Ό 가지고 있고, 이런 λ‚΄λΆ€λ¬Έμ œλŠ” 일반적으둜 μ •ν™•ν•œ 닡을 계산할 수 μ—†λ‹€. κ·Έλ ‡κΈ° λ•Œλ¬Έμ— 이런 λ‚΄λΆ€λ¬Έμ œλ₯Ό μ λ‹Ήν•œ 쑰건을 λ§Œμ‘±ν•˜λ„λ‘ λΆ€μ •ν™•ν•˜ 게 풀더라도 κ°€μ†ν™”ν•œ μ–΄κ·Έλ¨Όν‹°λ“œ λΌκ·Έλž‘μ§€ 방법이 μ •ν™•ν•˜κ²Œ λ‚΄λΆ€λ¬Έμ œλ₯Ό ν’€λ•Œμ™€ 같은 μˆ˜λ ΄μ„±μ„ κ°–λŠ” 쑰건을 μ œμ‹œν•œλ‹€. μš°λ¦¬λŠ” λ˜ν•œ κ°€μ†ν™”ν•œ μ–Όν„° λ„€μ΄νŒ… λ””λ ‰μ…˜ 방법데 λŒ€ν•΄μ„œλ„ λΉ„μŠ·ν•œ λ‚΄μš©μ„ μ „κ°œν•œλ‹€.Abstract i 1 Introduction 1 2 Previous Methods 5 2.1 Mathematical Preliminary . . . . . . . . . . . . . . . . . . . . 5 2.2 The algorithms for solving the linearly constrained convex minimization . . . . . . . . . . . . . . . . . . . . . . . . . . . 8 2.2.1 Augmented Lagrangian Method . . . . . . . . . . . . . 8 2.2.2 Bregman Methods . . . . . . . . . . . . . . . . . . . . 9 2.2.3 Alternating direction method of multipliers . . . . . . . 13 2.3 The accelerating algorithms for unconstrained convex minimization problem . . . . . . . . . . . . . . . . . . . . . . . . . 15 2.3.1 Fast inexact iterative shrinkage thresholding algorithm 16 2.3.2 Inexact accelerated proximal point method . . . . . . . 19 3 Proposed Algorithms 23 3.1 Proposed Algorithm 1 : Accelerated Bregman method . . . . . 23 3.1.1 Equivalence to the accelerated augmented Lagrangian method . . . . . . . . . . . . . . . . . . . . . . . . . . 24 3.1.2 Complexity of the accelerated Bregman method . . . . 27 3.2 Proposed Algorithm 2 : I-AALM . . . . . . . . . . . . . . . . 35 3.3 Proposed Algorithm 3 : I-AADMM . . . . . . . . . . . . . . . 43 3.4 Numerical Results . . . . . . . . . . . . . . . . . . . . . . . . . 54 3.4.1 Comparison to Bregman method with accelerated Bregman method . . . . . . . . . . . . . . . . . . . . . . . . 54 3.4.2 Numerical results of inexact accelerated augmented Lagrangian method using various subproblem solvers . . . 60 3.4.3 Comparison to the inexact accelerated augmented Lagrangian method with other methods . . . . . . . . . . 63 3.4.4 Inexact accelerated alternating direction method of multipliers for Multiplicative Noise Removal . . . . . . . . 69 4 Conclusion 86 Abstract (in Korean) 94Docto

    Augmented L1 and Nuclear-Norm Models with a Globally Linearly Convergent Algorithm

    Full text link
    This paper studies the long-existing idea of adding a nice smooth function to "smooth" a non-differentiable objective function in the context of sparse optimization, in particular, the minimization of ∣∣x∣∣1+1/(2Ξ±)∣∣x∣∣22||x||_1+1/(2\alpha)||x||_2^2, where xx is a vector, as well as the minimization of ∣∣Xβˆ£βˆ£βˆ—+1/(2Ξ±)∣∣X∣∣F2||X||_*+1/(2\alpha)||X||_F^2, where XX is a matrix and ∣∣Xβˆ£βˆ£βˆ—||X||_* and ∣∣X∣∣F||X||_F are the nuclear and Frobenius norms of XX, respectively. We show that they can efficiently recover sparse vectors and low-rank matrices. In particular, they enjoy exact and stable recovery guarantees similar to those known for minimizing ∣∣x∣∣1||x||_1 and ∣∣Xβˆ£βˆ£βˆ—||X||_* under the conditions on the sensing operator such as its null-space property, restricted isometry property, spherical section property, or RIPless property. To recover a (nearly) sparse vector x0x^0, minimizing ∣∣x∣∣1+1/(2Ξ±)∣∣x∣∣22||x||_1+1/(2\alpha)||x||_2^2 returns (nearly) the same solution as minimizing ∣∣x∣∣1||x||_1 almost whenever Ξ±β‰₯10∣∣x0∣∣∞\alpha\ge 10||x^0||_\infty. The same relation also holds between minimizing ∣∣Xβˆ£βˆ£βˆ—+1/(2Ξ±)∣∣X∣∣F2||X||_*+1/(2\alpha)||X||_F^2 and minimizing ∣∣Xβˆ£βˆ£βˆ—||X||_* for recovering a (nearly) low-rank matrix X0X^0, if Ξ±β‰₯10∣∣X0∣∣2\alpha\ge 10||X^0||_2. Furthermore, we show that the linearized Bregman algorithm for minimizing ∣∣x∣∣1+1/(2Ξ±)∣∣x∣∣22||x||_1+1/(2\alpha)||x||_2^2 subject to Ax=bAx=b enjoys global linear convergence as long as a nonzero solution exists, and we give an explicit rate of convergence. The convergence property does not require a solution solution or any properties on AA. To our knowledge, this is the best known global convergence result for first-order sparse optimization algorithms.Comment: arXiv admin note: text overlap with arXiv:1207.5326 by other author

    Fast stochastic dual coordinate descent algorithms for linearly constrained convex optimization

    Full text link
    The problem of finding a solution to the linear system Ax=bAx = b with certain minimization properties arises in numerous scientific and engineering areas. In the era of big data, the stochastic optimization algorithms become increasingly significant due to their scalability for problems of unprecedented size. This paper focuses on the problem of minimizing a strongly convex function subject to linear constraints. We consider the dual formulation of this problem and adopt the stochastic coordinate descent to solve it. The proposed algorithmic framework, called fast stochastic dual coordinate descent, utilizes sampling matrices sampled from user-defined distributions to extract gradient information. Moreover, it employs Polyak's heavy ball momentum acceleration with adaptive parameters learned through iterations, overcoming the limitation of the heavy ball momentum method that it requires prior knowledge of certain parameters, such as the singular values of a matrix. With these extensions, the framework is able to recover many well-known methods in the context, including the randomized sparse Kaczmarz method, the randomized regularized Kaczmarz method, the linearized Bregman iteration, and a variant of the conjugate gradient (CG) method. We prove that, with strongly admissible objective function, the proposed method converges linearly in expectation. Numerical experiments are provided to confirm our results.Comment: arXiv admin note: text overlap with arXiv:2305.0548

    MAGMA: Multi-level accelerated gradient mirror descent algorithm for large-scale convex composite minimization

    Full text link
    Composite convex optimization models arise in several applications, and are especially prevalent in inverse problems with a sparsity inducing norm and in general convex optimization with simple constraints. The most widely used algorithms for convex composite models are accelerated first order methods, however they can take a large number of iterations to compute an acceptable solution for large-scale problems. In this paper we propose to speed up first order methods by taking advantage of the structure present in many applications and in image processing in particular. Our method is based on multi-level optimization methods and exploits the fact that many applications that give rise to large scale models can be modelled using varying degrees of fidelity. We use Nesterov's acceleration techniques together with the multi-level approach to achieve O(1/Ο΅)\mathcal{O}(1/\sqrt{\epsilon}) convergence rate, where Ο΅\epsilon denotes the desired accuracy. The proposed method has a better convergence rate than any other existing multi-level method for convex problems, and in addition has the same rate as accelerated methods, which is known to be optimal for first-order methods. Moreover, as our numerical experiments show, on large-scale face recognition problems our algorithm is several times faster than the state of the art

    An Extragradient-Based Alternating Direction Method for Convex Minimization

    Get PDF
    In this paper, we consider the problem of minimizing the sum of two convex functions subject to linear linking constraints. The classical alternating direction type methods usually assume that the two convex functions have relatively easy proximal mappings. However, many problems arising from statistics, image processing and other fields have the structure that while one of the two functions has easy proximal mapping, the other function is smoothly convex but does not have an easy proximal mapping. Therefore, the classical alternating direction methods cannot be applied. To deal with the difficulty, we propose in this paper an alternating direction method based on extragradients. Under the assumption that the smooth function has a Lipschitz continuous gradient, we prove that the proposed method returns an Ο΅\epsilon-optimal solution within O(1/Ο΅)O(1/\epsilon) iterations. We apply the proposed method to solve a new statistical model called fused logistic regression. Our numerical experiments show that the proposed method performs very well when solving the test problems. We also test the performance of the proposed method through solving the lasso problem arising from statistics and compare the result with several existing efficient solvers for this problem; the results are very encouraging indeed
    • …
    corecore