23 research outputs found

    Graph Oracle Models, Lower Bounds, and Gaps for Parallel Stochastic Optimization

    Full text link
    We suggest a general oracle-based framework that captures different parallel stochastic optimization settings described by a dependency graph, and derive generic lower bounds in terms of this graph. We then use the framework and derive lower bounds for several specific parallel optimization settings, including delayed updates and parallel processing with intermittent communication. We highlight gaps between lower and upper bounds on the oracle complexity, and cases where the "natural" algorithms are not known to be optimal

    Parallel Submodular Function Minimization

    Full text link
    We consider the parallel complexity of submodular function minimization (SFM). We provide a pair of methods which obtain two new query versus depth trade-offs a submodular function defined on subsets of nn elements that has integer values between M-M and MM. The first method has depth 22 and query complexity nO(M)n^{O(M)} and the second method has depth O~(n1/3M2/3)\widetilde{O}(n^{1/3} M^{2/3}) and query complexity O(poly(n,M))O(\mathrm{poly}(n, M)). Despite a line of work on improved parallel lower bounds for SFM, prior to our work the only known algorithms for parallel SFM either followed from more general methods for sequential SFM or highly-parallel minimization of convex 2\ell_2-Lipschitz functions. Interestingly, to obtain our second result we provide the first highly-parallel algorithm for minimizing \ell_\infty-Lipschitz function over the hypercube which obtains near-optimal depth for obtaining constant accuracy

    No Quantum Speedup over Gradient Descent for Non-Smooth Convex Optimization

    Get PDF
    We study the first-order convex optimization problem, where we have black-box access to a (not necessarily smooth) function f:RnRf:\mathbb{R}^n \to \mathbb{R} and its (sub)gradient. Our goal is to find an ϵ\epsilon-approximate minimum of ff starting from a point that is distance at most RR from the true minimum. If ff is GG-Lipschitz, then the classic gradient descent algorithm solves this problem with O((GR/ϵ)2)O((GR/\epsilon)^{2}) queries. Importantly, the number of queries is independent of the dimension nn and gradient descent is optimal in this regard: No deterministic or randomized algorithm can achieve better complexity that is still independent of the dimension nn. In this paper we reprove the randomized lower bound of Ω((GR/ϵ)2)\Omega((GR/\epsilon)^{2}) using a simpler argument than previous lower bounds. We then show that although the function family used in the lower bound is hard for randomized algorithms, it can be solved using O(GR/ϵ)O(GR/\epsilon) quantum queries. We then show an improved lower bound against quantum algorithms using a different set of instances and establish our main result that in general even quantum algorithms need Ω((GR/ϵ)2)\Omega((GR/\epsilon)^2) queries to solve the problem. Hence there is no quantum speedup over gradient descent for black-box first-order convex optimization without further assumptions on the function family.Comment: 25 page

    Memory-Query Tradeoffs for Randomized Convex Optimization

    Full text link
    We show that any randomized first-order algorithm which minimizes a dd-dimensional, 11-Lipschitz convex function over the unit ball must either use Ω(d2δ)\Omega(d^{2-\delta}) bits of memory or make Ω(d1+δ/6o(1))\Omega(d^{1+\delta/6-o(1)}) queries, for any constant δ(0,1)\delta\in (0,1) and when the precision ϵ\epsilon is quasipolynomially small in dd. Our result implies that cutting plane methods, which use O~(d2)\tilde{O}(d^2) bits of memory and O~(d)\tilde{O}(d) queries, are Pareto-optimal among randomized first-order algorithms, and quadratic memory is required to achieve optimal query complexity for convex optimization

    Submodular Maximization with Matroid and Packing Constraints in Parallel

    Full text link
    We consider the problem of maximizing the multilinear extension of a submodular function subject a single matroid constraint or multiple packing constraints with a small number of adaptive rounds of evaluation queries. We obtain the first algorithms with low adaptivity for submodular maximization with a matroid constraint. Our algorithms achieve a 11/eϵ1-1/e-\epsilon approximation for monotone functions and a 1/eϵ1/e-\epsilon approximation for non-monotone functions, which nearly matches the best guarantees known in the fully adaptive setting. The number of rounds of adaptivity is O(log2n/ϵ3)O(\log^2{n}/\epsilon^3), which is an exponential speedup over the existing algorithms. We obtain the first parallel algorithm for non-monotone submodular maximization subject to packing constraints. Our algorithm achieves a 1/eϵ1/e-\epsilon approximation using O(log(n/ϵ)log(1/ϵ)log(n+m)/ϵ2)O(\log(n/\epsilon) \log(1/\epsilon) \log(n+m)/ \epsilon^2) parallel rounds, which is again an exponential speedup in parallel time over the existing algorithms. For monotone functions, we obtain a 11/eϵ1-1/e-\epsilon approximation in O(log(n/ϵ)log(m)/ϵ2)O(\log(n/\epsilon)\log(m)/\epsilon^2) parallel rounds. The number of parallel rounds of our algorithm matches that of the state of the art algorithm for solving packing LPs with a linear objective. Our results apply more generally to the problem of maximizing a diminishing returns submodular (DR-submodular) function
    corecore