1,256 research outputs found

    Mapping constrained optimization problems to quantum annealing with application to fault diagnosis

    Get PDF
    Current quantum annealing (QA) hardware suffers from practical limitations such as finite temperature, sparse connectivity, small qubit numbers, and control error. We propose new algorithms for mapping boolean constraint satisfaction problems (CSPs) onto QA hardware mitigating these limitations. In particular we develop a new embedding algorithm for mapping a CSP onto a hardware Ising model with a fixed sparse set of interactions, and propose two new decomposition algorithms for solving problems too large to map directly into hardware. The mapping technique is locally-structured, as hardware compatible Ising models are generated for each problem constraint, and variables appearing in different constraints are chained together using ferromagnetic couplings. In contrast, global embedding techniques generate a hardware independent Ising model for all the constraints, and then use a minor-embedding algorithm to generate a hardware compatible Ising model. We give an example of a class of CSPs for which the scaling performance of D-Wave's QA hardware using the local mapping technique is significantly better than global embedding. We validate the approach by applying D-Wave's hardware to circuit-based fault-diagnosis. For circuits that embed directly, we find that the hardware is typically able to find all solutions from a min-fault diagnosis set of size N using 1000N samples, using an annealing rate that is 25 times faster than a leading SAT-based sampling method. Further, we apply decomposition algorithms to find min-cardinality faults for circuits that are up to 5 times larger than can be solved directly on current hardware.Comment: 22 pages, 4 figure

    Preprocessing under uncertainty

    Get PDF
    In this work we study preprocessing for tractable problems when part of the input is unknown or uncertain. This comes up naturally if, e.g., the load of some machines or the congestion of some roads is not known far enough in advance, or if we have to regularly solve a problem over instances that are largely similar, e.g., daily airport scheduling with few charter flights. Unlike robust optimization, which also studies settings like this, our goal lies not in computing solutions that are (approximately) good for every instantiation. Rather, we seek to preprocess the known parts of the input, to speed up finding an optimal solution once the missing data is known. We present efficient algorithms that given an instance with partially uncertain input generate an instance of size polynomial in the amount of uncertain data that is equivalent for every instantiation of the unknown part. Concretely, we obtain such algorithms for Minimum Spanning Tree, Minimum Weight Matroid Basis, and Maximum Cardinality Bipartite Maxing, where respectively the weight of edges, weight of elements, and the availability of vertices is unknown for part of the input. Furthermore, we show that there are tractable problems, such as Small Connected Vertex Cover, for which one cannot hope to obtain similar results.Comment: 18 page

    Secluded Connectivity Problems

    Full text link
    Consider a setting where possibly sensitive information sent over a path in a network is visible to every {neighbor} of the path, i.e., every neighbor of some node on the path, thus including the nodes on the path itself. The exposure of a path PP can be measured as the number of nodes adjacent to it, denoted by N[P]N[P]. A path is said to be secluded if its exposure is small. A similar measure can be applied to other connected subgraphs, such as Steiner trees connecting a given set of terminals. Such subgraphs may be relevant due to considerations of privacy, security or revenue maximization. This paper considers problems related to minimum exposure connectivity structures such as paths and Steiner trees. It is shown that on unweighted undirected nn-node graphs, the problem of finding the minimum exposure path connecting a given pair of vertices is strongly inapproximable, i.e., hard to approximate within a factor of O(2log1ϵn)O(2^{\log^{1-\epsilon}n}) for any ϵ>0\epsilon>0 (under an appropriate complexity assumption), but is approximable with ratio Δ+3\sqrt{\Delta}+3, where Δ\Delta is the maximum degree in the graph. One of our main results concerns the class of bounded-degree graphs, which is shown to exhibit the following interesting dichotomy. On the one hand, the minimum exposure path problem is NP-hard on node-weighted or directed bounded-degree graphs (even when the maximum degree is 4). On the other hand, we present a polynomial algorithm (based on a nontrivial dynamic program) for the problem on unweighted undirected bounded-degree graphs. Likewise, the problem is shown to be polynomial also for the class of (weighted or unweighted) bounded-treewidth graphs

    35th Symposium on Theoretical Aspects of Computer Science: STACS 2018, February 28-March 3, 2018, Caen, France

    Get PDF

    Solving two-stage stochastic network design problems to optimality

    Get PDF
    The Steiner tree problem (STP) is a central and well-studied graph-theoretical combinatorial optimization problem which plays an important role in various applications. It can be stated as follows: Given a weighted graph and a set of terminal vertices, find a subset of edges which connects the terminals at minimum cost. However, in real-world applications the input data might not be given with certainty or it might depend on future decisions. For the STP, for example, edge costs representing the costs of establishing links may be subject to inflations and price deviations. In this thesis we tackle data uncertainty by using the concept of stochastic programming and we study the two-stage stochastic version of the Steiner tree problem (SSTP). Thereby, a set of scenarios defines the possible outcomes of a random variable; each scenario is given by its realization probability and defines a set of terminals and edge costs. A feasible solution consists of a subset of edges in the first stage and edge subsets for all scenarios (second stage) such that each terminal set is connected. The objective is to find a solution that minimizes the expected cost. We consider two approaches for solving the SSTP to optimality: combinatorial algorithms, in particular fixed-parameter tractable (FPT) algorithms, and methods from mathematical programming. Regarding the combinatorial algorithms we develop a linear-time algorithm for trees, an FPT algorithm parameterized by the number of terminals, and we consider treewidth-bounded graphs where we give the first FPT algorithm parameterized by the combination of treewidth and number of scenarios. The second approach is based on deriving strong integer programming (IP) formulations for the SSTP. By using orientation properties we introduce new semi-directed cut- and flow-based IP formulations which are shown to be stronger than the undirected models from the literature. To solve these models to optimality we use a decomposition-based two-stage branch&cut algorithm, which is improved by a fast and efficient method for strengthening the optimality cuts. Moreover, we develop new and stronger integer optimality cuts. The computational performance is evaluated in a comprehensive computational study, which shows the superiority of the new formulations, the benefit of the decomposition, and the advantage of using the strengthened optimality cuts. The Steiner forest problem (SFP) is a related problem where sets of terminals need to be connected. On the one hand, the SFP is a generalization of the STP and on the other hand, we show that the SFP is a special case of the SSTP. Therefore, our results are transferable to the SFP and we present the first FPT algorithm for treewidth-bounded graphs and we model new and stronger (semi-)directed cut- and flow-based IP formulations for the SFP. In the second part of this thesis we consider the two-stage stochastic survivable network design problem, an extension of the SSTP where pairs of vertices may demand a higher connectivity. Similarly to the first part we introduce new and stronger semi-directed cut-based models, apply the same decomposition along with the cut strengthening technique, and argue the validity of the newly introduced integer optimality cuts. A computational study shows the benefit, robustness, and good performance of the decomposition and the cut strengthening method
    corecore