240 research outputs found
On the convergence analysis of the greedy randomized Kaczmarz method
In this paper, we analyze the greedy randomized Kaczmarz (GRK) method
proposed in Bai and Wu (SIAM J. Sci. Comput., 40(1):A592--A606, 2018) for
solving linear systems. We develop more precise greedy probability criteria to
effectively select the working row from the coefficient matrix. Notably, we
prove that the linear convergence of the GRK method is deterministic and
demonstrate that using a tighter threshold parameter can lead to a faster
convergence rate. Our result revises existing convergence analyses, which are
solely based on the expected error by realizing that the iterates of the GRK
method are random variables. Consequently, we obtain an improved iteration
complexity for the GRK method. Moreover, the Polyak's heavy ball momentum
technique is incorporated to improve the performance of the GRK method. We
propose a refined convergence analysis, compared with the technique used in
Loizou and Richt\'{a}rik (Comput. Optim. Appl., 77(3):653--710, 2020), of
momentum variants of randomized iterative methods, which shows that the
proposed GRK method with momentum (mGRK) also enjoys a deterministic linear
convergence. Numerical experiments show that the mGRK method is more efficient
than the GRK method
On adaptive stochastic heavy ball momentum for solving linear systems
The stochastic heavy ball momentum (SHBM) method has gained considerable
popularity as a scalable approach for solving large-scale optimization
problems. However, one limitation of this method is its reliance on prior
knowledge of certain problem parameters, such as singular values of a matrix.
In this paper, we propose an adaptive variant of the SHBM method for solving
stochastic problems that are reformulated from linear systems using
user-defined distributions. Our adaptive SHBM (ASHBM) method utilizes iterative
information to update the parameters, addressing an open problem in the
literature regarding the adaptive learning of momentum parameters. We prove
that our method converges linearly in expectation, with a better convergence
rate compared to the basic method. Notably, we demonstrate that the
deterministic version of our ASHBM algorithm can be reformulated as a variant
of the conjugate gradient (CG) method, inheriting many of its appealing
properties, such as finite-time convergence. Consequently, the ASHBM method can
be further generalized to develop a brand-new framework of the stochastic CG
(SCG) method for solving linear systems. Our theoretical results are supported
by numerical experiments
Fast stochastic dual coordinate descent algorithms for linearly constrained convex optimization
The problem of finding a solution to the linear system with certain
minimization properties arises in numerous scientific and engineering areas. In
the era of big data, the stochastic optimization algorithms become increasingly
significant due to their scalability for problems of unprecedented size. This
paper focuses on the problem of minimizing a strongly convex function subject
to linear constraints. We consider the dual formulation of this problem and
adopt the stochastic coordinate descent to solve it. The proposed algorithmic
framework, called fast stochastic dual coordinate descent, utilizes sampling
matrices sampled from user-defined distributions to extract gradient
information. Moreover, it employs Polyak's heavy ball momentum acceleration
with adaptive parameters learned through iterations, overcoming the limitation
of the heavy ball momentum method that it requires prior knowledge of certain
parameters, such as the singular values of a matrix. With these extensions, the
framework is able to recover many well-known methods in the context, including
the randomized sparse Kaczmarz method, the randomized regularized Kaczmarz
method, the linearized Bregman iteration, and a variant of the conjugate
gradient (CG) method. We prove that, with strongly admissible objective
function, the proposed method converges linearly in expectation. Numerical
experiments are provided to confirm our results.Comment: arXiv admin note: text overlap with arXiv:2305.0548
- …