61,184 research outputs found

    Fully dynamic maintenance of k-connectivity in parallel

    Get PDF
    Ā©2001 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.Given a graph G=(V, E) with n vertices and m edges, the k-connectivity of G denotes either the k-edge connectivity or the k-vertex connectivity of G. In this paper, we deal with the fully dynamic maintenance of k-connectivity of G in the parallel setting for k=2, 3. We study the problem of maintaining k-edge/vertex connected components of a graph undergoing repeatedly dynamic updates, such as edge insertions and deletions, and answering the query of whether two vertices are included in the same k-edge/vertex connected component. Our major results are the following: (1) An NC algorithm for the 2-edge connectivity problem is proposed, which runs in O(log n log(m/n)) time using O(n3/4) processors per update and query. (2) It is shown that the biconnectivity problem can be solved in O(log2 n ) time using O(nĪ±(2n, n)/logn) processors per update and O(1) time with a single processor per query or in O(log n logn/m) time using O(nĪ±(2n, n)/log n) processors per update and O(logn) time using O(nĪ±(2n, n)/logn) processors per query, where Ī±(.,.) is the inverse of Ackermann's function. (3) An NC algorithm for the triconnectivity problem is also derived, which takes O(log n logn/m+logn log log n/Ī±(3n, n)) time using O(nĪ±(3n, n)/log n) processors per update and O(1) time with a single processor per query. (4) An NC algorithm for the 3-edge connectivity problem is obtained, which has the same time and processor complexities as the algorithm for the triconnectivity problem. To the best of our knowledge, the proposed algorithms are the first NC algorithms for the problems using O(n) processors in contrast to Ī©(m) processors for solving them from scratch. In particular, the proposed NC algorithm for the 2-edge connectivity problem uses only O(n3/4) processors. All the proposed algorithms run on a CRCW PRAMWeifa Liang, Brent, R.P., Hong She

    Randomized fully dynamic graph algorithms with polylogarithmic time per operation

    Get PDF
    This paper solves a longstanding open problem in fully dynamic algorithms: We present the first fully dynamic algorithms that maintain connectivity, bipartiteness, and approximate minimum spanning trees in polylogarithmic time per edge insertion or deletion. The algorithms are designed using a new dynamic technique that combines a novel graph decomposition with randomization. They are Las-Vegas type randomized algorithms which use simple data structures and have a small constant factor. Let n denote the number of nodes in the graph. For a sequence of _0_(m0) operations, where m0 is the number of edges in the initial graph, the expected time for p updates is O(p log3 n) (Throughout the paper the logarithms are base 2.) for connectivity and bipartiteness. The worst-case time for one query is O(log n/log log n). For the k-edge witness problem ("Does the removal of k given edges disconnect the graph?") the expected time for p updates is O(p.pow(log(n),3)) and the expected time for q queries is O(p_k.pow(log(n),3)). Given a graph with k different weights, the minimum spanning tree can be maintained during a sequence of p updates in expected time O(p_k.pow(log(n),3)). This implies an algorithm to maintain a 1 + e- approximation of the minimum spanning tree in expected time O((p.pow(log(n),3).log U)/e) for p updates, where the weights of the edges are between 1 and U

    Optimal decremental connectivity in planar graphs

    Get PDF
    We show an algorithm for dynamic maintenance of connectivity information in an undirected planar graph subject to edge deletions. Our algorithm may answer connectivity queries of the form `Are vertices uu and vv connected with a path?' in constant time. The queries can be intermixed with any sequence of edge deletions, and the algorithm handles all updates in O(n)O(n) time. This results improves over previously known O(nlogā”n)O(n \log n) time algorithm
    • ā€¦
    corecore