14 research outputs found

    Smooth Minimization of Nonsmooth Functions with Parallel Coordinate Descent Methods

    No full text
    39 pages, 1 algorithm, 3 figures, 2 tablesInternational audienceWe study the performance of a family of randomized parallel coordinate descent methods for minimizing the sum of a nonsmooth and separable convex functions. The problem class includes as a special case L1-regularized L1 regression and the minimization of the exponential loss ("AdaBoost problem"). We assume the input data defining the loss function is contained in a sparse m×nm\times n matrix AA with at most ω\omega nonzeros in each row. Our methods need O(nβ/τ)O(n \beta/\tau) iterations to find an approximate solution with high probability, where τ\tau is the number of processors and β=1+(ω−1)(τ−1)/(n−1)\beta = 1 + (\omega-1)(\tau-1)/(n-1) for the fastest variant. The notation hides dependence on quantities such as the required accuracy and confidence levels and the distance of the starting iterate from an optimal point. Since β/τ\beta/\tau is a decreasing function of τ\tau, the method needs fewer iterations when more processors are used. Certain variants of our algorithms perform on average only O(\nnz(A)/n) arithmetic operations during a single iteration per processor and, because β\beta decreases when ω\omega does, fewer iterations are needed for sparser problems
    corecore