206 research outputs found
Linearly Convergent First-Order Algorithms for Semi-definite Programming
In this paper, we consider two formulations for Linear Matrix Inequalities
(LMIs) under Slater type constraint qualification assumption, namely, SDP
smooth and non-smooth formulations. We also propose two first-order linearly
convergent algorithms for solving these formulations. Moreover, we introduce a
bundle-level method which converges linearly uniformly for both smooth and
non-smooth problems and does not require any smoothness information. The
convergence properties of these algorithms are also discussed. Finally, we
consider a special case of LMIs, linear system of inequalities, and show that a
linearly convergent algorithm can be obtained under a weaker assumption
Discussion
Peer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/106905/1/insr12033.pd
Large Scale Constrained Linear Regression Revisited: Faster Algorithms via Preconditioning
In this paper, we revisit the large-scale constrained linear regression
problem and propose faster methods based on some recent developments in
sketching and optimization. Our algorithms combine (accelerated) mini-batch SGD
with a new method called two-step preconditioning to achieve an approximate
solution with a time complexity lower than that of the state-of-the-art
techniques for the low precision case. Our idea can also be extended to the
high precision case, which gives an alternative implementation to the Iterative
Hessian Sketch (IHS) method with significantly improved time complexity.
Experiments on benchmark and synthetic datasets suggest that our methods indeed
outperform existing ones considerably in both the low and high precision cases.Comment: Appear in AAAI-1
- …