455 research outputs found

    Accelerated Gossip in Networks of Given Dimension using Jacobi Polynomial Iterations

    Get PDF
    Consider a network of agents connected by communication links, where each agent holds a real value. The gossip problem consists in estimating the average of the values diffused in the network in a distributed manner. We develop a method solving the gossip problem that depends only on the spectral dimension of the network, that is, in the communication network set-up, the dimension of the space in which the agents live. This contrasts with previous work that required the spectral gap of the network as a parameter, or suffered from slow mixing. Our method shows an important improvement over existing algorithms in the non-asymptotic regime, i.e., when the values are far from being fully mixed in the network. Our approach stems from a polynomial-based point of view on gossip algorithms, as well as an approximation of the spectral measure of the graphs with a Jacobi measure. We show the power of the approach with simulations on various graphs, and with performance guarantees on graphs of known spectral dimension, such as grids and random percolation bonds. An extension of this work to distributed Laplacian solvers is discussed. As a side result, we also use the polynomial-based point of view to show the convergence of the message passing algorithm for gossip of Moallemi \& Van Roy on regular graphs. The explicit computation of the rate of the convergence shows that message passing has a slow rate of convergence on graphs with small spectral gap

    Leveraging the two timescale regime to demonstrate convergence of neural networks

    Full text link
    We study the training dynamics of shallow neural networks, in a two-timescale regime in which the stepsizes for the inner layer are much smaller than those for the outer layer. In this regime, we prove convergence of the gradient flow to a global optimum of the non-convex optimization problem in a simple univariate setting. The number of neurons need not be asymptotically large for our result to hold, distinguishing our result from popular recent approaches such as the neural tangent kernel or mean-field regimes. Experimental illustration is provided, showing that the stochastic gradient descent behaves according to our description of the gradient flow and thus converges to a global optimum in the two-timescale regime, but can fail outside of this regime.Comment: 33 pages, 7 figure

    Use of green roofs to solve storm water issues at the basin scale – Study in the Hauts-de-Seine County (France)

    Get PDF
    International audienceAt the building scale, green roof has demonstrated a positive impact on urban runoff (decrease in the peak discharge and runoff volume). This work aims to study if similar impacts can be observed at basin scale. It is particularly focused on the possibility to solve some operational issues caused by storm water.For this purpose, a methodology has been proposed. It combines: a method to estimate the maximum roof area that can be covered by green roof, called green roofing potential, and an urban rainfall-runoff model able to simulate the hydrological behaviour of green roof.This methodology was applied to two urban catchments affected one by flooding and the other one by combined sewage overflow. The results show that green roof can reduce the frequency and the magnitude of such problems depending on the covered roof surface. Combined with other infrastructures, they represent an interesting solution for urban water management

    Accelerated Gossip in Networks of Given Dimension using Jacobi Polynomial Iterations

    Get PDF
    Consider a network of agents connected by communication links, where each agent holds a real value. The gossip problem consists in estimating the average of the values diffused in the network in a distributed manner. We develop a method solving the gossip problem that depends only on the spectral dimension of the network, that is, in the communication network set-up, the dimension of the space in which the agents live. This contrasts with previous work that required the spectral gap of the network as a parameter, or suffered from slow mixing. Our method shows an important improvement over existing algorithms in the non-asymptotic regime, i.e., when the values are far from being fully mixed in the network. Our approach stems from a polynomial-based point of view on gossip algorithms, as well as an approximation of the spectral measure of the graphs with a Jacobi measure. We show the power of the approach with simulations on various graphs, and with performance guarantees on graphs of known spectral dimension, such as grids and random percolation bonds. An extension of this work to distributed Laplacian solvers is discussed. As a side result, we also use the polynomial-based point of view to show the convergence of the message passing algorithm for gossip of Moallemi & Van Roy on regular graphs. The explicit computation of the rate of the convergence shows that message passing has a slow rate of convergence on graphs with small spectral gap

    Tight Nonparametric Convergence Rates for Stochastic Gradient Descent under the Noiseless Linear Model

    Get PDF
    International audienceIn the context of statistical supervised learning, the noiseless linear model assumes that there exists a deterministic linear relation Y=θ,XY = \langle \theta_*, X \rangle between the random output YY and the random feature vector Φ(U)\Phi(U), a potentially non-linear transformation of the inputs UU. We analyze the convergence of single-pass, fixed step-size stochastic gradient descent on the least-square risk under this model. The convergence of the iterates to the optimum θ\theta_* and the decay of the generalization error follow polynomial convergence rates with exponents that both depend on the regularities of the optimum θ\theta_* and of the feature vectors Φ(u)\Phi(u). We interpret our result in the reproducing kernel Hilbert space framework. As a special case, we analyze an online algorithm for estimating a real function on the unit interval from the noiseless observation of its value at randomly sampled points; the convergence depends on the Sobolev smoothness of the function and of a chosen kernel. Finally, we apply our analysis beyond the supervised learning setting to obtain convergence rates for the averaging process (a.k.a. gossip algorithm) on a graph depending on its spectral dimension

    A Continuized View on Nesterov Acceleration

    Get PDF
    We introduce the "continuized" Nesterov acceleration, a close variant of Nesterov acceleration whose variables are indexed by a continuous time parameter. The two variables continuously mix following a linear ordinary differential equation and take gradient steps at random times. This continuized variant benefits from the best of the continuous and the discrete frameworks: as a continuous process, one can use differential calculus to analyze convergence and obtain analytical expressions for the parameters; but a discretization of the continuized process can be computed exactly with convergence rates similar to those of Nesterov original acceleration. We show that the discretization has the same structure as Nesterov acceleration, but with random parameters
    corecore