20 research outputs found

    Exact Localisations of Feedback Sets

    Full text link
    The feedback arc (vertex) set problem, shortened FASP (FVSP), is to transform a given multi digraph G=(V,E)G=(V,E) into an acyclic graph by deleting as few arcs (vertices) as possible. Due to the results of Richard M. Karp in 1972 it is one of the classic NP-complete problems. An important contribution of this paper is that the subgraphs Gel(e)G_{\mathrm{el}}(e), Gsi(e)G_{\mathrm{si}}(e) of all elementary cycles or simple cycles running through some arc eEe \in E, can be computed in O(E2)\mathcal{O}\big(|E|^2\big) and O(E4)\mathcal{O}(|E|^4), respectively. We use this fact and introduce the notion of the essential minor and isolated cycles, which yield a priori problem size reductions and in the special case of so called resolvable graphs an exact solution in O(VE3)\mathcal{O}(|V||E|^3). We show that weighted versions of the FASP and FVSP possess a Bellman decomposition, which yields exact solutions using a dynamic programming technique in times O(2mE4log(V))\mathcal{O}\big(2^{m}|E|^4\log(|V|)\big) and O(2nΔ(G)4V4log(E))\mathcal{O}\big(2^{n}\Delta(G)^4|V|^4\log(|E|)\big), where mEV+1m \leq |E|-|V| +1, n(Δ(G)1)VE+1n \leq (\Delta(G)-1)|V|-|E| +1, respectively. The parameters m,nm,n can be computed in O(E3)\mathcal{O}(|E|^3), O(Δ(G)3V3)\mathcal{O}(\Delta(G)^3|V|^3), respectively and denote the maximal dimension of the cycle space of all appearing meta graphs, decoding the intersection behavior of the cycles. Consequently, m,nm,n equal zero if all meta graphs are trees. Moreover, we deliver several heuristics and discuss how to control their variation from the optimum. Summarizing, the presented results allow us to suggest a strategy for an implementation of a fast and accurate FASP/FVSP-SOLVER

    Qubit Allocation for Noisy Intermediate-Scale Quantum Computers

    Get PDF
    In the era of noisy-intermediate-scale quantum computers, we expect to see quantum devices with increasing numbers of qubits emerge in the foreseeable future. To practically run quantum programs, logical qubits have to be mapped to the physical qubits by a qubit allocation algorithm. However, on present day devices, qubits differ by their error rate and connectivity. Here, we establish and demonstrate on current experimental devices a new allocation algorithm that combines the simulated annealing method with local search of the solution space using Dijkstra's algorithm. Our algorithm takes into account the weighted connectivity constraints of both the quantum hardware and the quantum program being compiled. New quantum programs will enable unprecedented developments in physics, chemistry, and materials science and our work offers an important new pathway toward optimizing compilers for quantum programs.Comment: 6 pages, 3 figure

    High-Performance Reachability Query Processing under Index Size Restrictions

    Full text link
    In this paper, we propose a scalable and highly efficient index structure for the reachability problem over graphs. We build on the well-known node interval labeling scheme where the set of vertices reachable from a particular node is compactly encoded as a collection of node identifier ranges. We impose an explicit bound on the size of the index and flexibly assign approximate reachability ranges to nodes of the graph such that the number of index probes to answer a query is minimized. The resulting tunable index structure generates a better range labeling if the space budget is increased, thus providing a direct control over the trade off between index size and the query processing performance. By using a fast recursive querying method in conjunction with our index structure, we show that in practice, reachability queries can be answered in the order of microseconds on an off-the-shelf computer - even for the case of massive-scale real world graphs. Our claims are supported by an extensive set of experimental results using a multitude of benchmark and real-world web-scale graph datasets.Comment: 30 page

    Routing schemes for hybrid communication networks

    Get PDF
    We consider the problem of computing routing schemes in the HYBRID model of distributed computing where nodes have access to two fundamentally different communication modes. In this problem nodes have to compute small labels and routing tables that allow for efficient routing of messages in the local network, which typically offers the majority of the throughput. Recent work has shown that using the HYBRID model admits a significant speed-up compared to what would be possible if either communication mode were used in isolation. Nonetheless, if general graphs are used as the input graph the computation of routing schemes still takes polynomial rounds in the HYBRID model. We bypass this lower bound by restricting the local graph to unit-disc-graphs and solve the problem deterministically with running time O(|H|2+log⁡n), label size O(log⁡n), and size of routing tables O(|H|2⋅log⁡n) where |H| is the number of “radio holes” in the network. Our work builds on recent work by Coy et al., who obtain this result in the much simpler setting where the input graph has no radio holes. We develop new techniques to achieve this, including a decomposition of the local graph into path-convex regions, where each region contains a shortest path for any pair of nodes in it

    Multiple Allocation Hub Interdiction and Protection Problems: Model Formulations and Solution Approaches

    Get PDF
    In this paper, we present computationally efficient formulations for the multiple allocation hub interdiction and hub protection problems, which are bilevel and trilevel mixed integer linear programs, respectively. In the hub interdiction problem, the aim is to identify a subset of r critical hubs from an existing set of p hubs that when interdicted results in the maximum post-interdiction cost of routing flows. We present two alternate ways of reducing the bilevel hub interdiction model to a single level optimization problem. The first approach uses the dual formulation of the lower level problem. The second approach exploits the structure of the lower level problem to replace it by a set of closest assignment constraints (CACs). We present alternate sets of CACs, study their dominance relationships, and report their computational performances. Further, we propose refinements to CACs that offer computational advantages of an order-of-magnitude compared to the one existing in the literature. Further, our proposed modifications offer structural advantages for Benders decomposition, which lead to substantial computational savings, particularly for large problem instances. Finally, we study and solve large scale instances of the trilevel hub protection problem exactly by utilizing the ideas developed for the hub interdiction problem

    Optimal Trees

    Get PDF
    Not Availabl

    Design, implementation and evaluation of a distributed CDCL framework

    Get PDF
    The primary subject of this dissertation is practically solving instances of the Boolean satisfiability problem (SAT) that arise from industrial applications. The invention of the conflict-driven clause-learning (CDCL) algorithm led to enormous progress in this field. CDCL has been augmented with effective pre- and inprocessing techniques that boost its effectiveness. While a considerable amount of work has been done on applying shared-memory parallelism to enhance the performance of CDCL, solving SAT on distributed architectures is studied less thoroughly. In this work, we develop a distributed, CDCL-based framework for SAT solving. This framework consists of three main components: 1. An implementation of the CDCL algorithm that we have written from scratch, 2. a novel, parallel SAT algorithm that builds upon this CDCL implementation and 3. a collection of parallel simplification techniques for SAT instances. We call our resulting framework satUZK; our parallel solving algorithm is called the distributed divide-and-conquer (DDC) algorithm. The DDC algorithm employs a parallel lookahead procedure to dynamically partition the search space. Load balancing is used to ensure that all computational resources are utilized during lookahead. This procedure results in a divide-and-conquer tree that is distributed over all processors. Individual threads are routed through this tree until they arrive at unsolved leaf vertices. Upon arrival, the lookahead procedure is invoked again or the leaf vertex is solved via CDCL. Several extensions to the DDC algorithm are proposed. These include clause sharing and a scheme to locally adjust the LBD score relative to the current search tree vertex. LBD is a measure for the usefulness of clauses that participate in a CDCL search. We evaluate our DDC algorithm empirically and benchmark it against the best distributed SAT algorithms. In this experiment, our DDC algorithm is faster than other distributed, state-of-the-art solvers and solves at least as many instances. In addition to running a parallel algorithm for SAT solving we also consider parallel simplifcation. Here, we first develop a theoretical foundation that allows us to prove the correctness of parallel simplification techniques. Using this as a basis, we examine established simplification algorithms for their parallelizability. It turns out that several well-known simplification techniques can be parallelized efficiently. We provide parallel implementation of the techniques and test their effectiveness in empirical experiments. This evaluation finds several combinations of simplification techniques that can solve instances which could not be solved by the DDC algorithm alone
    corecore