52 research outputs found

    Lower Bounds on the Chromatic Number of Random Graphs

    Full text link
    We prove that a formula predicted on the basis of non-rigorous physics arguments [Zdeborová and Krzakala: Phys. Rev. E (2007)] provides a lower bound on the chromatic number of sparse random graphs. The proof is based on the interpolation method from mathematical physics. In the case of random regular graphs the lower bound can be expressed algebraically, while in the case of the binomial random we obtain a variational formula. As an application we calculate improved explicit lower bounds on the chromatic number of random graphs for small (average) degrees. Additionally, we show how asymptotic formulas for large degrees that were previously obtained by lengthy and complicated combinatorial arguments can be re-derived easily from these new results

    Smoothed Complexity Theory

    Get PDF
    Smoothed analysis is a new way of analyzing algorithms introduced by Spielman and Teng (J. ACM, 2004). Classical methods like worst-case or average-case analysis have accompanying complexity classes, like P and AvgP, respectively. While worst-case or average-case analysis give us a means to talk about the running time of a particular algorithm, complexity classes allows us to talk about the inherent difficulty of problems. Smoothed analysis is a hybrid of worst-case and average-case analysis and compensates some of their drawbacks. Despite its success for the analysis of single algorithms and problems, there is no embedding of smoothed analysis into computational complexity theory, which is necessary to classify problems according to their intrinsic difficulty. We propose a framework for smoothed complexity theory, define the relevant classes, and prove some first hardness results (of bounded halting and tiling) and tractability results (binary optimization problems, graph coloring, satisfiability). Furthermore, we discuss extensions and shortcomings of our model and relate it to semi-random models.Comment: to be presented at MFCS 201

    Computing Diffusion State Distance using Green's Function and Heat Kernel on Graphs

    Full text link
    The diffusion state distance (DSD) was introduced by Cao-Zhang-Park-Daniels-Crovella-Cowen-Hescott [{\em PLoS ONE, 2013}] to capture functional similarity in protein-protein interaction networks. They proved the convergence of DSD for non-bipartite graphs. In this paper, we extend the DSD to bipartite graphs using lazy-random walks and consider the general LqL_q-version of DSD. We discovered the connection between the DSD LqL_q-distance and Green's function, which was studied by Chung and Yau [{\em J. Combinatorial Theory (A), 2000}]. Based on that, we computed the DSD LqL_q-distance for Paths, Cycles, Hypercubes, as well as random graphs G(n,p)G(n,p) and G(w1,...,wn)G(w_1,..., w_n). We also examined the DSD distances of two biological networks.Comment: Accepted by the 11th Workshop on Algorithms and Models for the Web Graph (WAW2014

    On the cavity method for decimated random constraint satisfaction problems and the analysis of belief propagation guided decimation algorithms

    Full text link
    We introduce a version of the cavity method for diluted mean-field spin models that allows the computation of thermodynamic quantities similar to the Franz-Parisi quenched potential in sparse random graph models. This method is developed in the particular case of partially decimated random constraint satisfaction problems. This allows to develop a theoretical understanding of a class of algorithms for solving constraint satisfaction problems, in which elementary degrees of freedom are sequentially assigned according to the results of a message passing procedure (belief-propagation). We confront this theoretical analysis to the results of extensive numerical simulations.Comment: 32 pages, 24 figure

    Sparser Random 3SAT Refutation Algorithms and the Interpolation Problem:Extended Abstract

    Get PDF
    We formalize a combinatorial principle, called the 3XOR principle, due to Feige, Kim and Ofek [12], as a family of unsatisfiable propositional formulas for which refutations of small size in any propo-sitional proof system that possesses the feasible interpolation property imply an efficient deterministic refutation algorithm for random 3SAT with n variables and Ω(n1.4) clauses. Such small size refutations would improve the state of the art (with respect to the clause density) efficient refutation algorithm, which works only for Ω(n1.5) many clauses [13]. We demonstrate polynomial-size refutations of the 3XOR principle in resolution operating with disjunctions of quadratic equations with small integer coefficients, denoted R(quad); this is a weak extension of cutting planes with small coefficients. We show that R(quad) is weakly autom-atizable iff R(lin) is weakly automatizable, where R(lin) is similar to R(quad) but with linear instead of quadratic equations (introduced in [25]). This reduces the problem of refuting random 3CNF with n vari-ables and Ω(n1.4) clauses to the interpolation problem of R(quad) and to the weak automatizability of R(lin)

    On Eigenvalues of Random Complexes

    Full text link
    We consider higher-dimensional generalizations of the normalized Laplacian and the adjacency matrix of graphs and study their eigenvalues for the Linial-Meshulam model Xk(n,p)X^k(n,p) of random kk-dimensional simplicial complexes on nn vertices. We show that for p=Ω(logn/n)p=\Omega(\log n/n), the eigenvalues of these matrices are a.a.s. concentrated around two values. The main tool, which goes back to the work of Garland, are arguments that relate the eigenvalues of these matrices to those of graphs that arise as links of (k2)(k-2)-dimensional faces. Garland's result concerns the Laplacian; we develop an analogous result for the adjacency matrix. The same arguments apply to other models of random complexes which allow for dependencies between the choices of kk-dimensional simplices. In the second part of the paper, we apply this to the question of possible higher-dimensional analogues of the discrete Cheeger inequality, which in the classical case of graphs relates the eigenvalues of a graph and its edge expansion. It is very natural to ask whether this generalizes to higher dimensions and, in particular, whether the higher-dimensional Laplacian spectra capture the notion of coboundary expansion - a generalization of edge expansion that arose in recent work of Linial and Meshulam and of Gromov. We show that this most straightforward version of a higher-dimensional discrete Cheeger inequality fails, in quite a strong way: For every k2k\geq 2 and nNn\in \mathbb{N}, there is a kk-dimensional complex YnkY^k_n on nn vertices that has strong spectral expansion properties (all nontrivial eigenvalues of the normalised kk-dimensional Laplacian lie in the interval [1O(1/n),1+O(1/n)][1-O(1/\sqrt{n}),1+O(1/\sqrt{n})]) but whose coboundary expansion is bounded from above by O(logn/n)O(\log n/n) and so tends to zero as nn\rightarrow \infty; moreover, YnkY^k_n can be taken to have vanishing integer homology in dimension less than kk.Comment: Extended full version of an extended abstract that appeared at SoCG 2012, to appear in Israel Journal of Mathematic

    Analysis of an exhaustive search algorithm in random graphs and the n^{c\log n} -asymptotics

    Full text link
    We analyze the cost used by a naive exhaustive search algorithm for finding a maximum independent set in random graphs under the usual G_{n,p} -model where each possible edge appears independently with the same probability p. The expected cost turns out to be of the less common asymptotic order n^{c\log n}, which we explore from several different perspectives. Also we collect many instances where such an order appears, from algorithmics to analysis, from probability to algebra. The limiting distribution of the cost required by the algorithm under a purely idealized random model is proved to be normal. The approach we develop is of some generality and is amenable for other graph algorithms.Comment: 35 page
    corecore