397 research outputs found

    A probabilistic approach to reducing the algebraic complexity of computing Delaunay triangulations

    Get PDF
    Computing Delaunay triangulations in Rd\mathbb{R}^d involves evaluating the so-called in\_sphere predicate that determines if a point xx lies inside, on or outside the sphere circumscribing d+1d+1 points p0,,pdp_0,\ldots ,p_d. This predicate reduces to evaluating the sign of a multivariate polynomial of degree d+2d+2 in the coordinates of the points x,p0,,pdx, \, p_0,\, \ldots,\, p_d. Despite much progress on exact geometric computing, the fact that the degree of the polynomial increases with dd makes the evaluation of the sign of such a polynomial problematic except in very low dimensions. In this paper, we propose a new approach that is based on the witness complex, a weak form of the Delaunay complex introduced by Carlsson and de Silva. The witness complex Wit(L,W)\mathrm{Wit} (L,W) is defined from two sets LL and WW in some metric space XX: a finite set of points LL on which the complex is built, and a set WW of witnesses that serves as an approximation of XX. A fundamental result of de Silva states that Wit(L,W)=Del(L)\mathrm{Wit}(L,W)=\mathrm{Del} (L) if W=X=RdW=X=\mathbb{R}^d. In this paper, we give conditions on LL that ensure that the witness complex and the Delaunay triangulation coincide when WW is a finite set, and we introduce a new perturbation scheme to compute a perturbed set LL' close to LL such that Del(L)=wit(L,W)\mathrm{Del} (L')= \mathrm{wit} (L', W). Our perturbation algorithm is a geometric application of the Moser-Tardos constructive proof of the Lov\'asz local lemma. The only numerical operations we use are (squared) distance comparisons (i.e., predicates of degree 2). The time-complexity of the algorithm is sublinear in W|W|. Interestingly, although the algorithm does not compute any measure of simplex quality, a lower bound on the thickness of the output simplices can be guaranteed.Comment: 24 page

    A Probabilistic Approach to Reducing Algebraic Complexity of Delaunay Triangulations

    Get PDF
    International audienceWe propose algorithms to compute the Delaunay triangulation of a point set L using only (squared) distance comparisons (i.e., predicates of degree 2). Our approach is based on the witness complex, a weak form of the Delaunay complex introduced by Carlsson and de Silva. We give conditions that ensure that the witness complex and the Delaunay triangulation coincide and we introduce a new perturbation scheme to compute a perturbed set L′ close to L such that the Delaunay triangulation and the witness complex coincide. Our perturbation algorithm is a geometric application of the Moser-Tardos constructive proof of the Lovász local lemma

    Fast Local Computation Algorithms

    Full text link
    For input xx, let F(x)F(x) denote the set of outputs that are the "legal" answers for a computational problem FF. Suppose xx and members of F(x)F(x) are so large that there is not time to read them in their entirety. We propose a model of {\em local computation algorithms} which for a given input xx, support queries by a user to values of specified locations yiy_i in a legal output yF(x)y \in F(x). When more than one legal output yy exists for a given xx, the local computation algorithm should output in a way that is consistent with at least one such yy. Local computation algorithms are intended to distill the common features of several concepts that have appeared in various algorithmic subfields, including local distributed computation, local algorithms, locally decodable codes, and local reconstruction. We develop a technique, based on known constructions of small sample spaces of kk-wise independent random variables and Beck's analysis in his algorithmic approach to the Lov{\'{a}}sz Local Lemma, which under certain conditions can be applied to construct local computation algorithms that run in {\em polylogarithmic} time and space. We apply this technique to maximal independent set computations, scheduling radio network broadcasts, hypergraph coloring and satisfying kk-SAT formulas.Comment: A preliminary version of this paper appeared in ICS 2011, pp. 223-23

    The Complexity of Distributed Edge Coloring with Small Palettes

    Full text link
    The complexity of distributed edge coloring depends heavily on the palette size as a function of the maximum degree Δ\Delta. In this paper we explore the complexity of edge coloring in the LOCAL model in different palette size regimes. 1. We simplify the \emph{round elimination} technique of Brandt et al. and prove that (2Δ2)(2\Delta-2)-edge coloring requires Ω(logΔlogn)\Omega(\log_\Delta \log n) time w.h.p. and Ω(logΔn)\Omega(\log_\Delta n) time deterministically, even on trees. The simplified technique is based on two ideas: the notion of an irregular running time and some general observations that transform weak lower bounds into stronger ones. 2. We give a randomized edge coloring algorithm that can use palette sizes as small as Δ+O~(Δ)\Delta + \tilde{O}(\sqrt{\Delta}), which is a natural barrier for randomized approaches. The running time of the algorithm is at most O(logΔTLLL)O(\log\Delta \cdot T_{LLL}), where TLLLT_{LLL} is the complexity of a permissive version of the constructive Lovasz local lemma. 3. We develop a new distributed Lovasz local lemma algorithm for tree-structured dependency graphs, which leads to a (1+ϵ)Δ(1+\epsilon)\Delta-edge coloring algorithm for trees running in O(loglogn)O(\log\log n) time. This algorithm arises from two new results: a deterministic O(logn)O(\log n)-time LLL algorithm for tree-structured instances, and a randomized O(loglogn)O(\log\log n)-time graph shattering method for breaking the dependency graph into independent O(logn)O(\log n)-size LLL instances. 4. A natural approach to computing (Δ+1)(\Delta+1)-edge colorings (Vizing's theorem) is to extend partial colorings by iteratively re-coloring parts of the graph. We prove that this approach may be viable, but in the worst case requires recoloring subgraphs of diameter Ω(Δlogn)\Omega(\Delta\log n). This stands in contrast to distributed algorithms for Brooks' theorem, which exploit the existence of O(logΔn)O(\log_\Delta n)-length augmenting paths
    corecore