7 research outputs found

    Delayed Stochastic Algorithms for Distributed Weakly Convex Optimization

    Full text link
    This paper studies delayed stochastic algorithms for weakly convex optimization in a distributed network with workers connected to a master node. More specifically, we consider a structured stochastic weakly convex objective function which is the composition of a convex function and a smooth nonconvex function. Recently, Xu et al. 2022 showed that an inertial stochastic subgradient method converges at a rate of O(Ï„/K)\mathcal{O}(\tau/\sqrt{K}), which suffers a significant penalty from the maximum information delay Ï„\tau. To alleviate this issue, we propose a new delayed stochastic prox-linear (DSPL\texttt{DSPL}) method in which the master performs the proximal update of the parameters and the workers only need to linearly approximate the inner smooth function. Somewhat surprisingly, we show that the delays only affect the high order term in the complexity rate and hence, are negligible after a certain number of DSPL\texttt{DSPL} iterations. Moreover, to further improve the empirical performance, we propose a delayed extrapolated prox-linear (DSEPL\texttt{DSEPL}) method which employs Polyak-type momentum to speed up the algorithm convergence. Building on the tools for analyzing DSPL\texttt{DSPL}, we also develop improved analysis of delayed stochastic subgradient method (DSGD\texttt{DSGD}). In particular, for general weakly convex problems, we show that convergence of DSGD\texttt{DSGD} only depends on the expected delay
    corecore