2,459 research outputs found

    Cover Time and Broadcast Time

    Get PDF
    We introduce a new technique for bounding the cover time of random walks by relating it to the runtime of randomized broadcast. In particular, we strongly confirm for dense graphs the intuition of Chandra et al. (1997) that ``the cover time of the graph is an appropriate metric for the performance of certain kinds of randomized broadcast algorithms\u27\u27. In more detail, our results are as follows: begin{itemize} item For any graph G=(V,E)G=(V,E) of size nn and minimum degree deltadelta, we have mathcalR(G)=mathcalO(fracEdeltacdotlogn)mathcal{R}(G)= mathcal{O}(frac{|E|}{delta} cdot log n), where mathcalR(G)mathcal{R}(G) denotes the quotient of the cover time and broadcast time. This bound is tight for binary trees and tight up to logarithmic factors for many graphs including hypercubes, expanders and lollipop graphs. item For any deltadelta-regular (or almost deltadelta-regular) graph GG it holds that mathcalR(G)=Omega(fracdelta2ncdotfrac1logn)mathcal{R}(G) = Omega(frac{delta^2}{n} cdot frac{1}{log n}). Together with our upper bound on mathcalR(G)mathcal{R}(G), this lower bound strongly confirms the intuition of Chandra et al.~for graphs with minimum degree Theta(n)Theta(n), since then the cover time equals the broadcast time multiplied by nn (neglecting logarithmic factors). item Conversely, for any deltadelta we construct almost deltadelta-regular graphs that satisfy mathcalR(G)=mathcalO(maxsqrtn,deltacdotlog2n)mathcal{R}(G) = mathcal{O}(max { sqrt{n},delta } cdot log^2 n). Since any regular expander satisfies mathcalR(G)=Theta(n)mathcal{R}(G) = Theta(n), the strong relationship given above does not hold if deltadelta is polynomially smaller than nn. end{itemize} Our bounds also demonstrate that the relationship between cover time and broadcast time is much stronger than the known relationships between any of them and the mixing time (or the closely related spectral gap)

    The coalescing-branching random walk on expanders and the dual epidemic process

    Get PDF
    Information propagation on graphs is a fundamental topic in distributed computing. One of the simplest models of information propagation is the push protocol in which at each round each agent independently pushes the current knowledge to a random neighbour. In this paper we study the so-called coalescing-branching random walk (COBRA), in which each vertex pushes the information to kk randomly selected neighbours and then stops passing information until it receives the information again. The aim of COBRA is to propagate information fast but with a limited number of transmissions per vertex per step. In this paper we study the cover time of the COBRA process defined as the minimum time until each vertex has received the information at least once. Our main result says that if GG is an nn-vertex rr-regular graph whose transition matrix has second eigenvalue λ\lambda, then the COBRA cover time of GG is O(logn)\mathcal O(\log n ), if 1λ1-\lambda is greater than a positive constant, and O((logn)/(1λ)3))\mathcal O((\log n)/(1-\lambda)^3)), if 1λlog(n)/n1-\lambda \gg \sqrt{\log( n)/n}. These bounds are independent of rr and hold for 3rn13 \le r \le n-1. They improve the previous bound of O(log2n)O(\log^2 n) for expander graphs. Our main tool in analysing the COBRA process is a novel duality relation between this process and a discrete epidemic process, which we call a biased infection with persistent source (BIPS). A fixed vertex vv is the source of an infection and remains permanently infected. At each step each vertex uu other than vv selects kk neighbours, independently and uniformly, and uu is infected in this step if and only if at least one of the selected neighbours has been infected in the previous step. We show the duality between COBRA and BIPS which says that the time to infect the whole graph in the BIPS process is of the same order as the cover time of the COBRA proces

    Towards Resistance Sparsifiers

    Get PDF
    We study resistance sparsification of graphs, in which the goal is to find a sparse subgraph (with reweighted edges) that approximately preserves the effective resistances between every pair of nodes. We show that every dense regular expander admits a (1+ϵ)(1+\epsilon)-resistance sparsifier of size O~(n/ϵ)\tilde O(n/\epsilon), and conjecture this bound holds for all graphs on nn nodes. In comparison, spectral sparsification is a strictly stronger notion and requires Ω(n/ϵ2)\Omega(n/\epsilon^2) edges even on the complete graph. Our approach leads to the following structural question on graphs: Does every dense regular expander contain a sparse regular expander as a subgraph? Our main technical contribution, which may of independent interest, is a positive answer to this question in a certain setting of parameters. Combining this with a recent result of von Luxburg, Radl, and Hein~(JMLR, 2014) leads to the aforementioned resistance sparsifiers

    Linear-time list recovery of high-rate expander codes

    Full text link
    We show that expander codes, when properly instantiated, are high-rate list recoverable codes with linear-time list recovery algorithms. List recoverable codes have been useful recently in constructing efficiently list-decodable codes, as well as explicit constructions of matrices for compressive sensing and group testing. Previous list recoverable codes with linear-time decoding algorithms have all had rate at most 1/2; in contrast, our codes can have rate 1ϵ1 - \epsilon for any ϵ>0\epsilon > 0. We can plug our high-rate codes into a construction of Meir (2014) to obtain linear-time list recoverable codes of arbitrary rates, which approach the optimal trade-off between the number of non-trivial lists provided and the rate of the code. While list-recovery is interesting on its own, our primary motivation is applications to list-decoding. A slight strengthening of our result would implies linear-time and optimally list-decodable codes for all rates, and our work is a step in the direction of solving this important problem

    Towards a better approximation for sparsest cut?

    Full text link
    We give a new (1+ϵ)(1+\epsilon)-approximation for sparsest cut problem on graphs where small sets expand significantly more than the sparsest cut (sets of size n/rn/r expand by a factor lognlogr\sqrt{\log n\log r} bigger, for some small rr; this condition holds for many natural graph families). We give two different algorithms. One involves Guruswami-Sinop rounding on the level-rr Lasserre relaxation. The other is combinatorial and involves a new notion called {\em Small Set Expander Flows} (inspired by the {\em expander flows} of ARV) which we show exists in the input graph. Both algorithms run in time 2O(r)poly(n)2^{O(r)} \mathrm{poly}(n). We also show similar approximation algorithms in graphs with genus gg with an analogous local expansion condition. This is the first algorithm we know of that achieves (1+ϵ)(1+\epsilon)-approximation on such general family of graphs

    Derandomization and Group Testing

    Full text link
    The rapid development of derandomization theory, which is a fundamental area in theoretical computer science, has recently led to many surprising applications outside its initial intention. We will review some recent such developments related to combinatorial group testing. In its most basic setting, the aim of group testing is to identify a set of "positive" individuals in a population of items by taking groups of items and asking whether there is a positive in each group. In particular, we will discuss explicit constructions of optimal or nearly-optimal group testing schemes using "randomness-conducting" functions. Among such developments are constructions of error-correcting group testing schemes using randomness extractors and condensers, as well as threshold group testing schemes from lossless condensers.Comment: Invited Paper in Proceedings of 48th Annual Allerton Conference on Communication, Control, and Computing, 201

    On Fortification of Projection Games

    Get PDF
    A recent result of Moshkovitz \cite{Moshkovitz14} presented an ingenious method to provide a completely elementary proof of the Parallel Repetition Theorem for certain projection games via a construction called fortification. However, the construction used in \cite{Moshkovitz14} to fortify arbitrary label cover instances using an arbitrary extractor is insufficient to prove parallel repetition. In this paper, we provide a fix by using a stronger graph that we call fortifiers. Fortifiers are graphs that have both 1\ell_1 and 2\ell_2 guarantees on induced distributions from large subsets. We then show that an expander with sufficient spectral gap, or a bi-regular extractor with stronger parameters (the latter is also the construction used in an independent update \cite{Moshkovitz15} of \cite{Moshkovitz14} with an alternate argument), is a good fortifier. We also show that using a fortifier (in particular 2\ell_2 guarantees) is necessary for obtaining the robustness required for fortification.Comment: 19 page
    corecore