5 research outputs found

    Fast, Provably convergent IRLS Algorithm for p-norm Linear Regression

    Full text link
    Linear regression in β„“p\ell_p-norm is a canonical optimization problem that arises in several applications, including sparse recovery, semi-supervised learning, and signal processing. Generic convex optimization algorithms for solving β„“p\ell_p-regression are slow in practice. Iteratively Reweighted Least Squares (IRLS) is an easy to implement family of algorithms for solving these problems that has been studied for over 50 years. However, these algorithms often diverge for p > 3, and since the work of Osborne (1985), it has been an open problem whether there is an IRLS algorithm that is guaranteed to converge rapidly for p > 3. We propose p-IRLS, the first IRLS algorithm that provably converges geometrically for any p∈[2,∞).p \in [2,\infty). Our algorithm is simple to implement and is guaranteed to find a (1+Ξ΅)(1+\varepsilon)-approximate solution in O(p3.5mpβˆ’22(pβˆ’1)log⁑mΞ΅)≀Op(mlog⁑mΞ΅)O(p^{3.5} m^{\frac{p-2}{2(p-1)}} \log \frac{m}{\varepsilon}) \le O_p(\sqrt{m} \log \frac{m}{\varepsilon} ) iterations. Our experiments demonstrate that it performs even better than our theoretical bounds, beats the standard Matlab/CVX implementation for solving these problems by 10--50x, and is the fastest among available implementations in the high-accuracy regime.Comment: Code for this work is available at https://github.com/utoronto-theory/pIRL

    Faster p-norm minimizing flows, via smoothed q-norm problems

    Full text link
    We present faster high-accuracy algorithms for computing β„“p\ell_p-norm minimizing flows. On a graph with mm edges, our algorithm can compute a (1+1/poly(m))(1+1/\text{poly}(m))-approximate unweighted β„“p\ell_p-norm minimizing flow with pm1+1pβˆ’1+o(1)pm^{1+\frac{1}{p-1}+o(1)} operations, for any pβ‰₯2,p \ge 2, giving the best bound for all p≳5.24.p\gtrsim 5.24. Combined with the algorithm from the work of Adil et al. (SODA '19), we can now compute such flows for any 2≀p≀mo(1)2\le p\le m^{o(1)} in time at most O(m1.24).O(m^{1.24}). In comparison, the previous best running time was Ξ©(m1.33)\Omega(m^{1.33}) for large constant p.p. For pβˆΌΞ΄βˆ’1log⁑m,p\sim\delta^{-1}\log m, our algorithm computes a (1+Ξ΄)(1+\delta)-approximate maximum flow on undirected graphs using m1+o(1)Ξ΄βˆ’1m^{1+o(1)}\delta^{-1} operations, matching the current best bound, albeit only for unit-capacity graphs. We also give an algorithm for solving general β„“p\ell_{p}-norm regression problems for large p.p. Our algorithm makes pm13+o(1)log⁑2(1/Ξ΅)pm^{\frac{1}{3}+o(1)}\log^2(1/\varepsilon) calls to a linear solver. This gives the first high-accuracy algorithm for computing weighted β„“p\ell_{p}-norm minimizing flows that runs in time o(m1.5)o(m^{1.5}) for some p=mΞ©(1).p=m^{\Omega(1)}. Our key technical contribution is to show that smoothed β„“p\ell_p-norm problems introduced by Adil et al., are interreducible for different values of p.p. No such reduction is known for standard β„“p\ell_p-norm problems.Comment: ACM-SIAM Symposium on Discrete Algorithms (SODA 2020

    Iteratively Reweighted Least Squares for Basis Pursuit with Global Linear Convergence Rate

    Full text link
    The recovery of sparse data is at the core of many applications in machine learning and signal processing. While such problems can be tackled using β„“1\ell_1-regularization as in the LASSO estimator and in the Basis Pursuit approach, specialized algorithms are typically required to solve the corresponding high-dimensional non-smooth optimization for large instances. Iteratively Reweighted Least Squares (IRLS) is a widely used algorithm for this purpose due its excellent numerical performance. However, while existing theory is able to guarantee convergence of this algorithm to the minimizer, it does not provide a global convergence rate. In this paper, we prove that a variant of IRLS converges with a global linear rate to a sparse solution, i.e., with a linear error decrease occurring immediately from any initialization, if the measurements fulfill the usual null space property assumption. We support our theory by numerical experiments showing that our linear rate captures the correct dimension dependence. We anticipate that our theoretical findings will lead to new insights for many other use cases of the IRLS algorithm, such as in low-rank matrix recovery.Comment: 26 pages, 3 figure

    Acceleration with a Ball Optimization Oracle

    Full text link
    Consider an oracle which takes a point xx and returns the minimizer of a convex function ff in an β„“2\ell_2 ball of radius rr around xx. It is straightforward to show that roughly rβˆ’1log⁑1Ο΅r^{-1}\log\frac{1}{\epsilon} calls to the oracle suffice to find an Ο΅\epsilon-approximate minimizer of ff in an β„“2\ell_2 unit ball. Perhaps surprisingly, this is not optimal: we design an accelerated algorithm which attains an Ο΅\epsilon-approximate minimizer with roughly rβˆ’2/3log⁑1Ο΅r^{-2/3} \log \frac{1}{\epsilon} oracle queries, and give a matching lower bound. Further, we implement ball optimization oracles for functions with locally stable Hessians using a variant of Newton's method. The resulting algorithm applies to a number of problems of practical and theoretical import, improving upon previous results for logistic and β„“βˆž\ell_\infty regression and achieving guarantees comparable to the state-of-the-art for β„“p\ell_p regression.Comment: 37 page

    Almost-linear-time Weighted β„“p\ell_p-norm Solvers in Slightly Dense Graphs via Sparsification

    Full text link
    We give almost-linear-time algorithms for constructing sparsifiers with $n\ poly(\log n)edgesthatapproximatelypreserveweighted edges that approximately preserve weighted (\ell^{2}_2 + \ell^{p}_p)floworvoltageobjectivesongraphs.Forflowobjectives,thisisthefirstsparsifierconstructionforsuchmixedobjectivesbeyondunit flow or voltage objectives on graphs. For flow objectives, this is the first sparsifier construction for such mixed objectives beyond unit \ell_pweights,andisbasedonexpanderdecompositions.Forvoltageobjectives,wegivethefirstsparsifierconstructionfortheseobjectives,whichwebuildusinggraphspannersandleveragescoresampling.Togetherwiththeiterativerefinementframeworkof[Adiletal,SODA2019],andanewmultiplicativeβˆ’weightsbasedconstantβˆ’approximationalgorithmformixedβˆ’objectiveflowsorvoltages,weshowhowtofind weights, and is based on expander decompositions. For voltage objectives, we give the first sparsifier construction for these objectives, which we build using graph spanners and leverage score sampling. Together with the iterative refinement framework of [Adil et al, SODA 2019], and a new multiplicative-weights based constant-approximation algorithm for mixed-objective flows or voltages, we show how to find (1+2^{-\text{poly}(\log n)})approximationsforweighted approximations for weighted \ell_pβˆ’normminimizingflowsorvoltagesin-norm minimizing flows or voltages in p(m^{1+o(1)} + n^{4/3 + o(1)})timefor time for p=\omega(1),whichisalmostβˆ’linearforgraphsthatareslightlydense( which is almost-linear for graphs that are slightly dense (m \ge n^{4/3 + o(1)}$)
    corecore