522 research outputs found
The Price of Information in Combinatorial Optimization
Consider a network design application where we wish to lay down a
minimum-cost spanning tree in a given graph; however, we only have stochastic
information about the edge costs. To learn the precise cost of any edge, we
have to conduct a study that incurs a price. Our goal is to find a spanning
tree while minimizing the disutility, which is the sum of the tree cost and the
total price that we spend on the studies. In a different application, each edge
gives a stochastic reward value. Our goal is to find a spanning tree while
maximizing the utility, which is the tree reward minus the prices that we pay.
Situations such as the above two often arise in practice where we wish to
find a good solution to an optimization problem, but we start with only some
partial knowledge about the parameters of the problem. The missing information
can be found only after paying a probing price, which we call the price of
information. What strategy should we adopt to optimize our expected
utility/disutility?
A classical example of the above setting is Weitzman's "Pandora's box"
problem where we are given probability distributions on values of
independent random variables. The goal is to choose a single variable with a
large value, but we can find the actual outcomes only after paying a price. Our
work is a generalization of this model to other combinatorial optimization
problems such as matching, set cover, facility location, and prize-collecting
Steiner tree. We give a technique that reduces such problems to their non-price
counterparts, and use it to design exact/approximation algorithms to optimize
our utility/disutility. Our techniques extend to situations where there are
additional constraints on what parameters can be probed or when we can
simultaneously probe a subset of the parameters.Comment: SODA 201
Recommended from our members
Stochastic Network Design: Models and Scalable Algorithms
Many natural and social phenomena occur in networks. Examples include the spread of information, ideas, and opinions through a social network, the propagation of an infectious disease among people, and the spread of species within an interconnected habitat network. The ability to modify a phenomenon towards some desired outcomes has widely recognized benefits to our society and the economy. The outcome of a phenomenon is largely determined by the topology or properties of its underlying network. A decision maker can take management actions to modify a network and, therefore, change the outcome of the phenomenon. A management action is an activity that changes the topology or other properties of a network. For example, species that live in a small area may expand their population and gradually spread into an interconnected habitat network. However, human development of various structures such as highways and factories may destroy natural habitats or block paths connecting different habitat patches, which results in a population decline. To facilitate the dispersal of species and help the population recover, artificial corridors (e.g., a wildlife highway crossing) can be built to restore connectivity of isolated habitats, and conservation areas can be established to restore historical habitats of species, both of which are examples of management actions. The set of management actions that can be taken is restricted by a budget, so we must find cost-effective allocations of limited funding resources.
In the thesis, the problem of finding the (nearly) optimal set of management actions is formulated as a discrete and stochastic optimization problem. Specifically, a general decision-making framework called stochastic network design is defined to model a broad range of similar real-world problems. The framework is defined upon a stochastic network, in which edges are either present or absent with certain probabilities. It defines several metrics to measure the outcome of the underlying phenomenon and a set of management actions that modify the network or its parameters in specific ways. The goal is to select a subset of management actions, subject to a budget constraint, to maximize a specified metric.
The major contribution of the thesis is to develop scalable algorithms to find high- quality solutions for different problems within the framework. In general, these problems are NP-hard, and their objective functions are neither submodular nor super-modular. Existing algorithms, such as greedy algorithms and heuristic search algorithms, either lack theoretical guarantees or have limited scalability. In the thesis, fast approximate algorithms are developed under three different settings that are gradually more general. The most restricted setting is when a network is tree-structured. For this case, fully polynomial-time approximation schemes (FPTAS) are developed using dynamic programming algorithms and rounding techniques. A more general setting is when networks are general directed graphs. We use a sampling technique to convert the original stochastic optimization problem into a deterministic optimization problem and develop a primal-dual algorithm to solve it efficiently. In the previous two problem settings, the goal is to maximize connectivity of networks. In the most general setting, the goal is to maximize the number of nodes being connected and minimize the distance between these connected nodes. For example, we do not only want the species to reach a large number of habitat areas but also want them to be able to get there within a reasonable amount of time. The scalable algorithms for this setting combine a fast primal-dual algorithm and a sampling procedure.
Three real-world problems from the areas of computational sustainability and emergency response are used to evaluate these algorithms. They are the barrier removal problem aimed to determine which instream barriers to remove to help fish access their historical habitats in a river network, the spatial conservation planning problem to determine which habitat units to set as conservation areas to encourage the dispersal of endangered species in a landscape, and the pre-disaster preparation problem aimed to minimize the disruption of emergency medical services by natural disasters. In these three problems, the developed algorithms are much more scalable than the existing state-of-the-arts and produce high-quality solutions
Approximation Algorithms for Union and Intersection Covering Problems
In a classical covering problem, we are given a set of requests that we need
to satisfy (fully or partially), by buying a subset of items at minimum cost.
For example, in the k-MST problem we want to find the cheapest tree spanning at
least k nodes of an edge-weighted graph. Here nodes and edges represent
requests and items, respectively.
In this paper, we initiate the study of a new family of multi-layer covering
problems. Each such problem consists of a collection of h distinct instances of
a standard covering problem (layers), with the constraint that all layers share
the same set of requests. We identify two main subfamilies of these problems: -
in a union multi-layer problem, a request is satisfied if it is satisfied in at
least one layer; - in an intersection multi-layer problem, a request is
satisfied if it is satisfied in all layers. To see some natural applications,
consider both generalizations of k-MST. Union k-MST can model a problem where
we are asked to connect a set of users to at least one of two communication
networks, e.g., a wireless and a wired network. On the other hand, intersection
k-MST can formalize the problem of connecting a subset of users to both
electricity and water.
We present a number of hardness and approximation results for union and
intersection versions of several standard optimization problems: MST, Steiner
tree, set cover, facility location, TSP, and their partial covering variants
Deliver or hold: Approximation Algorithms for the Periodic Inventory Routing Problem
The inventory routing problem involves trading off inventory holding
costs at client locations with vehicle routing costs to deliver
frequently from a single central depot to meet deterministic client demands over a finite planing horizon. In this paper, we consider periodic solutions that visit clients in one of several specified frequencies, and focus on the case when the frequencies of visiting nodes are nested. We give the first constant-factor approximation algorithms for designing optimum nested periodic schedules for the problem with no limit on vehicle capacities by simple reductions to prize-collecting network design problems. For instance, we present a 2.55-approximation algorithm for the minimum-cost nested periodic
schedule where the vehicle routes are modeled as minimum Steiner trees. We also show a general reduction from the capacitated
problem where all vehicles have the same capacity to the uncapacitated
version with a slight loss in performance. This reduction gives a
4.55-approximation for the capacitated problem. In addition, we prove several structural results relating the values of optimal policies of various types
Improving Christofides' Algorithm for the s-t Path TSP
We present a deterministic (1+sqrt(5))/2-approximation algorithm for the s-t
path TSP for an arbitrary metric. Given a symmetric metric cost on n vertices
including two prespecified endpoints, the problem is to find a shortest
Hamiltonian path between the two endpoints; Hoogeveen showed that the natural
variant of Christofides' algorithm is a 5/3-approximation algorithm for this
problem, and this asymptotically tight bound in fact has been the best
approximation ratio known until now. We modify this algorithm so that it
chooses the initial spanning tree based on an optimal solution to the Held-Karp
relaxation rather than a minimum spanning tree; we prove this simple but
crucial modification leads to an improved approximation ratio, surpassing the
20-year-old barrier set by the natural Christofides' algorithm variant. Our
algorithm also proves an upper bound of (1+sqrt(5))/2 on the integrality gap of
the path-variant Held-Karp relaxation. The techniques devised in this paper can
be applied to other optimization problems as well: these applications include
improved approximation algorithms and improved LP integrality gap upper bounds
for the prize-collecting s-t path problem and the unit-weight graphical metric
s-t path TSP.Comment: 31 pages, 5 figure
Greedy Algorithms for Online Survivable Network Design
In an instance of the network design problem, we are given a graph G=(V,E), an edge-cost function c:E -> R^{>= 0}, and a connectivity criterion. The goal is to find a minimum-cost subgraph H of G that meets the connectivity requirements. An important family of this class is the survivable network design problem (SNDP): given non-negative integers r_{u v} for each pair u,v in V, the solution subgraph H should contain r_{u v} edge-disjoint paths for each pair u and v.
While this problem is known to admit good approximation algorithms in the offline case, the problem is much harder in the online setting. Gupta, Krishnaswamy, and Ravi [Gupta et al., 2012] (STOC\u2709) are the first to consider the online survivable network design problem. They demonstrate an algorithm with competitive ratio of O(k log^3 n), where k=max_{u,v} r_{u v}. Note that the competitive ratio of the algorithm by Gupta et al. grows linearly in k. Since then, an important open problem in the online community [Naor et al., 2011; Gupta et al., 2012] is whether the linear dependence on k can be reduced to a logarithmic dependency.
Consider an online greedy algorithm that connects every demand by adding a minimum cost set of edges to H. Surprisingly, we show that this greedy algorithm significantly improves the competitive ratio when a congestion of 2 is allowed on the edges or when the model is stochastic. While our algorithm is fairly simple, our analysis requires a deep understanding of k-connected graphs. In particular, we prove that the greedy algorithm is O(log^2 n log k)-competitive if one satisfies every demand between u and v by r_{uv}/2 edge-disjoint paths. The spirit of our result is similar to the work of Chuzhoy and Li [Chuzhoy and Li, 2012] (FOCS\u2712), in which the authors give a polylogarithmic approximation algorithm for edge-disjoint paths with congestion 2.
Moreover, we study the greedy algorithm in the online stochastic setting. We consider the i.i.d. model, where each online demand is drawn from a single probability distribution, the unknown i.i.d. model, where every demand is drawn from a single but unknown probability distribution, and the prophet model in which online demands are drawn from (possibly) different probability distributions. Through a different analysis, we prove that a similar greedy algorithm is constant competitive for the i.i.d. and the prophet models. Also, the greedy algorithm is O(log n)-competitive for the unknown i.i.d. model, which is almost tight due to the lower bound of [Garg et al., 2008] for single connectivity
- …