21 research outputs found

    Incremental Network Design with Minimum Spanning Trees

    Full text link
    Given an edge-weighted graph G=(V,E)G=(V,E) and a set E0⊂EE_0\subset E, the incremental network design problem with minimum spanning trees asks for a sequence of edges e1′,…,eT′∈E∖E0e'_1,\ldots,e'_T\in E\setminus E_0 minimizing ∑t=1Tw(Xt)\sum_{t=1}^Tw(X_t) where w(Xt)w(X_t) is the weight of a minimum spanning tree XtX_t for the subgraph (V,E0∪{e1′,…,et′})(V,E_0\cup\{e'_1,\ldots,e'_t\}) and T=∣E∖E0∣T=\lvert E\setminus E_0\rvert. We prove that this problem can be solved by a greedy algorithm.Comment: 9 pages, minor revision based on reviewer comment

    A Linear Time Algorithm for Finding Minimum Spanning Tree Replacement Edges

    Full text link
    Given an undirected, weighted graph, the minimum spanning tree (MST) is a tree that connects all of the vertices of the graph with minimum sum of edge weights. In real world applications, network designers often seek to quickly find a replacement edge for each edge in the MST. For example, when a traffic accident closes a road in a transportation network, or a line goes down in a communication network, the replacement edge may reconnect the MST at lowest cost. In the paper, we consider the case of finding the lowest cost replacement edge for each edge of the MST. A previous algorithm by Tarjan takes O(mα(m,n))O(m \alpha(m, n)) time, where α(m,n)\alpha(m, n) is the inverse Ackermann's function. Given the MST and sorted non-tree edges, our algorithm is the first that runs in O(m+n)O(m+n) time and O(m+n)O(m+n) space to find all replacement edges. Moreover, it is easy to implement and our experimental study demonstrates fast performance on several types of graphs. Additionally, since the most vital edge is the tree edge whose removal causes the highest cost, our algorithm finds it in linear time

    Sensitivity Analysis of the Economic Lot-Sizing Problem

    Get PDF
    In this paper we study sensitivity analysis of the uncapacitated single level economic lot-sizing problem, which was introduced by Wagner and Whitin about thirty years ago. In particular we are concerned with the computation of the maximal ranges in which the numerical problem parameters may vary individually, such that a solution already obtained remains optimal. Only recently it was discovered that faster algorithms than the Wagner-Whitin algorithm exist to solve the economic lot-sizing problem. Moreover, these algorithms reveal that the problem has more structure than was recognized so far. When performing the sensitivity analysis we exploit these newly obtained insights

    Randomized fully dynamic graph algorithms with polylogarithmic time per operation

    Get PDF
    This paper solves a longstanding open problem in fully dynamic algorithms: We present the first fully dynamic algorithms that maintain connectivity, bipartiteness, and approximate minimum spanning trees in polylogarithmic time per edge insertion or deletion. The algorithms are designed using a new dynamic technique that combines a novel graph decomposition with randomization. They are Las-Vegas type randomized algorithms which use simple data structures and have a small constant factor. Let n denote the number of nodes in the graph. For a sequence of _0_(m0) operations, where m0 is the number of edges in the initial graph, the expected time for p updates is O(p log3 n) (Throughout the paper the logarithms are base 2.) for connectivity and bipartiteness. The worst-case time for one query is O(log n/log log n). For the k-edge witness problem ("Does the removal of k given edges disconnect the graph?") the expected time for p updates is O(p.pow(log(n),3)) and the expected time for q queries is O(p_k.pow(log(n),3)). Given a graph with k different weights, the minimum spanning tree can be maintained during a sequence of p updates in expected time O(p_k.pow(log(n),3)). This implies an algorithm to maintain a 1 + e- approximation of the minimum spanning tree in expected time O((p.pow(log(n),3).log U)/e) for p updates, where the weights of the edges are between 1 and U

    Parallel Real-Time Computation: Sometimes Quantity Means Quality

    Get PDF
    The primary purpose of parallel computation is the fast execution of computational tasks that require an inordinate amount of time to perform sequentially. As a consequence, interest in parallel computation to date has naturally focused on the speedup provided by parallel algorithms over their sequential counterparts. The thesis of this paper is that a second equally important motivation for using parallel computers exists. Specifically, the following question is posed: Can parallel computers, thanks to their multiple processors, do more than simply speed up the solution to a problem? We show that within the paradigm of real-time computation, some classes of problems have the property that a solution to a problem in the class, when computed in parallel, is far superior in quality than the best one obtained on a sequential computer. What constitutes a better solution depends on the problem under consideration. Thus, `better' means `closer to optimal' for optimization problems, `more secure' for cryptographic problems, and `more accurate' for numerical problems. Examples from these classes are presented. In each case, the solution obtained in parallel is significantly, provably, and consistently better than a sequential one. It is important to note that the purpose of this paper is not to demonstrate merely that a parallel computer can obtain a better solution to a computational problem than one derived sequentially. The latter is an interesting (and often surprising) observation in its own right, but we wish to go further. It is shown here that the improvement in quality can be arbitrarily high (and certainly superlinear in the number of processors used by the parallel computer). This result is akin to superlinear speedup --- a phenomenon itself originally thought to be impossible
    corecore