1,256 research outputs found
Mapping constrained optimization problems to quantum annealing with application to fault diagnosis
Current quantum annealing (QA) hardware suffers from practical limitations
such as finite temperature, sparse connectivity, small qubit numbers, and
control error. We propose new algorithms for mapping boolean constraint
satisfaction problems (CSPs) onto QA hardware mitigating these limitations. In
particular we develop a new embedding algorithm for mapping a CSP onto a
hardware Ising model with a fixed sparse set of interactions, and propose two
new decomposition algorithms for solving problems too large to map directly
into hardware.
The mapping technique is locally-structured, as hardware compatible Ising
models are generated for each problem constraint, and variables appearing in
different constraints are chained together using ferromagnetic couplings. In
contrast, global embedding techniques generate a hardware independent Ising
model for all the constraints, and then use a minor-embedding algorithm to
generate a hardware compatible Ising model. We give an example of a class of
CSPs for which the scaling performance of D-Wave's QA hardware using the local
mapping technique is significantly better than global embedding.
We validate the approach by applying D-Wave's hardware to circuit-based
fault-diagnosis. For circuits that embed directly, we find that the hardware is
typically able to find all solutions from a min-fault diagnosis set of size N
using 1000N samples, using an annealing rate that is 25 times faster than a
leading SAT-based sampling method. Further, we apply decomposition algorithms
to find min-cardinality faults for circuits that are up to 5 times larger than
can be solved directly on current hardware.Comment: 22 pages, 4 figure
Preprocessing under uncertainty
In this work we study preprocessing for tractable problems when part of the
input is unknown or uncertain. This comes up naturally if, e.g., the load of
some machines or the congestion of some roads is not known far enough in
advance, or if we have to regularly solve a problem over instances that are
largely similar, e.g., daily airport scheduling with few charter flights.
Unlike robust optimization, which also studies settings like this, our goal
lies not in computing solutions that are (approximately) good for every
instantiation. Rather, we seek to preprocess the known parts of the input, to
speed up finding an optimal solution once the missing data is known.
We present efficient algorithms that given an instance with partially
uncertain input generate an instance of size polynomial in the amount of
uncertain data that is equivalent for every instantiation of the unknown part.
Concretely, we obtain such algorithms for Minimum Spanning Tree, Minimum Weight
Matroid Basis, and Maximum Cardinality Bipartite Maxing, where respectively the
weight of edges, weight of elements, and the availability of vertices is
unknown for part of the input. Furthermore, we show that there are tractable
problems, such as Small Connected Vertex Cover, for which one cannot hope to
obtain similar results.Comment: 18 page
Secluded Connectivity Problems
Consider a setting where possibly sensitive information sent over a path in a
network is visible to every {neighbor} of the path, i.e., every neighbor of
some node on the path, thus including the nodes on the path itself. The
exposure of a path can be measured as the number of nodes adjacent to it,
denoted by . A path is said to be secluded if its exposure is small. A
similar measure can be applied to other connected subgraphs, such as Steiner
trees connecting a given set of terminals. Such subgraphs may be relevant due
to considerations of privacy, security or revenue maximization. This paper
considers problems related to minimum exposure connectivity structures such as
paths and Steiner trees. It is shown that on unweighted undirected -node
graphs, the problem of finding the minimum exposure path connecting a given
pair of vertices is strongly inapproximable, i.e., hard to approximate within a
factor of for any (under an
appropriate complexity assumption), but is approximable with ratio
, where is the maximum degree in the graph. One of
our main results concerns the class of bounded-degree graphs, which is shown to
exhibit the following interesting dichotomy. On the one hand, the minimum
exposure path problem is NP-hard on node-weighted or directed bounded-degree
graphs (even when the maximum degree is 4). On the other hand, we present a
polynomial algorithm (based on a nontrivial dynamic program) for the problem on
unweighted undirected bounded-degree graphs. Likewise, the problem is shown to
be polynomial also for the class of (weighted or unweighted) bounded-treewidth
graphs
Recommended from our members
Stochastic Network Design: Models and Scalable Algorithms
Many natural and social phenomena occur in networks. Examples include the spread of information, ideas, and opinions through a social network, the propagation of an infectious disease among people, and the spread of species within an interconnected habitat network. The ability to modify a phenomenon towards some desired outcomes has widely recognized benefits to our society and the economy. The outcome of a phenomenon is largely determined by the topology or properties of its underlying network. A decision maker can take management actions to modify a network and, therefore, change the outcome of the phenomenon. A management action is an activity that changes the topology or other properties of a network. For example, species that live in a small area may expand their population and gradually spread into an interconnected habitat network. However, human development of various structures such as highways and factories may destroy natural habitats or block paths connecting different habitat patches, which results in a population decline. To facilitate the dispersal of species and help the population recover, artificial corridors (e.g., a wildlife highway crossing) can be built to restore connectivity of isolated habitats, and conservation areas can be established to restore historical habitats of species, both of which are examples of management actions. The set of management actions that can be taken is restricted by a budget, so we must find cost-effective allocations of limited funding resources.
In the thesis, the problem of finding the (nearly) optimal set of management actions is formulated as a discrete and stochastic optimization problem. Specifically, a general decision-making framework called stochastic network design is defined to model a broad range of similar real-world problems. The framework is defined upon a stochastic network, in which edges are either present or absent with certain probabilities. It defines several metrics to measure the outcome of the underlying phenomenon and a set of management actions that modify the network or its parameters in specific ways. The goal is to select a subset of management actions, subject to a budget constraint, to maximize a specified metric.
The major contribution of the thesis is to develop scalable algorithms to find high- quality solutions for different problems within the framework. In general, these problems are NP-hard, and their objective functions are neither submodular nor super-modular. Existing algorithms, such as greedy algorithms and heuristic search algorithms, either lack theoretical guarantees or have limited scalability. In the thesis, fast approximate algorithms are developed under three different settings that are gradually more general. The most restricted setting is when a network is tree-structured. For this case, fully polynomial-time approximation schemes (FPTAS) are developed using dynamic programming algorithms and rounding techniques. A more general setting is when networks are general directed graphs. We use a sampling technique to convert the original stochastic optimization problem into a deterministic optimization problem and develop a primal-dual algorithm to solve it efficiently. In the previous two problem settings, the goal is to maximize connectivity of networks. In the most general setting, the goal is to maximize the number of nodes being connected and minimize the distance between these connected nodes. For example, we do not only want the species to reach a large number of habitat areas but also want them to be able to get there within a reasonable amount of time. The scalable algorithms for this setting combine a fast primal-dual algorithm and a sampling procedure.
Three real-world problems from the areas of computational sustainability and emergency response are used to evaluate these algorithms. They are the barrier removal problem aimed to determine which instream barriers to remove to help fish access their historical habitats in a river network, the spatial conservation planning problem to determine which habitat units to set as conservation areas to encourage the dispersal of endangered species in a landscape, and the pre-disaster preparation problem aimed to minimize the disruption of emergency medical services by natural disasters. In these three problems, the developed algorithms are much more scalable than the existing state-of-the-arts and produce high-quality solutions
Solving two-stage stochastic network design problems to optimality
The Steiner tree problem (STP) is a central and well-studied graph-theoretical combinatorial optimization problem which plays an important role in various applications. It can be stated as follows: Given a weighted graph and a set of terminal vertices, find a subset of edges which connects the terminals at minimum cost. However, in real-world applications the input data might not be given with certainty or it might depend on future decisions. For the STP, for example, edge costs representing the costs of establishing links may be subject to inflations and price deviations. In this thesis we tackle data uncertainty by using the concept of stochastic programming and we study the two-stage stochastic version of the Steiner tree problem (SSTP). Thereby, a set of scenarios defines the possible outcomes of a random variable; each scenario is given by its realization probability and defines a set of terminals and edge costs. A feasible solution consists of a subset of edges in the first stage and edge subsets for all scenarios (second stage) such that each terminal set is connected. The objective is to find a solution that minimizes the expected cost. We consider two approaches for solving the SSTP to optimality: combinatorial algorithms, in particular fixed-parameter tractable (FPT) algorithms, and methods from mathematical programming.
Regarding the combinatorial algorithms we develop a linear-time algorithm for trees, an FPT algorithm parameterized by the number of terminals, and we consider treewidth-bounded graphs where we give the first FPT algorithm parameterized by the combination of treewidth and number of scenarios.
The second approach is based on deriving strong integer programming (IP) formulations for the SSTP. By using orientation properties we introduce new semi-directed cut- and flow-based IP formulations which are shown to be stronger than the undirected models from the literature. To solve these models to optimality we use a decomposition-based two-stage branch&cut algorithm, which is improved by a fast and efficient method for strengthening the optimality cuts. Moreover, we develop new and stronger integer optimality cuts. The computational performance is evaluated in a comprehensive computational study, which shows the superiority of the new formulations, the benefit of the decomposition, and the advantage of using the strengthened optimality cuts.
The Steiner forest problem (SFP) is a related problem where sets of terminals need to be connected. On the one hand, the SFP is a generalization of the STP and on the other hand, we show that the SFP is a special case of the SSTP. Therefore, our results are transferable to the SFP and we present the first FPT algorithm for treewidth-bounded graphs and we model new and stronger (semi-)directed cut- and flow-based IP formulations for the SFP.
In the second part of this thesis we consider the two-stage stochastic survivable network design problem, an extension of the SSTP where pairs of vertices may demand a higher connectivity. Similarly to the first part we introduce new and stronger semi-directed cut-based models, apply the same decomposition along with the cut strengthening technique, and argue the validity of the newly introduced integer optimality cuts. A computational study shows the benefit, robustness, and good performance of the decomposition and the cut strengthening method
- …