8 research outputs found
Approximation Algorithms for Covering/Packing Integer Programs
Given matrices A and B and vectors a, b, c and d, all with non-negative
entries, we consider the problem of computing min {c.x: x in Z^n_+, Ax > a, Bx
< b, x < d}. We give a bicriteria-approximation algorithm that, given epsilon
in (0, 1], finds a solution of cost O(ln(m)/epsilon^2) times optimal, meeting
the covering constraints (Ax > a) and multiplicity constraints (x < d), and
satisfying Bx < (1 + epsilon)b + beta, where beta is the vector of row sums
beta_i = sum_j B_ij. Here m denotes the number of rows of A.
This gives an O(ln m)-approximation algorithm for CIP -- minimum-cost
covering integer programs with multiplicity constraints, i.e., the special case
when there are no packing constraints Bx < b. The previous best approximation
ratio has been O(ln(max_j sum_i A_ij)) since 1982. CIP contains the set cover
problem as a special case, so O(ln m)-approximation is the best possible unless
P=NP.Comment: Preliminary version appeared in IEEE Symposium on Foundations of
Computer Science (2001). To appear in Journal of Computer and System Science
Maximum n-times Coverage for Vaccine Design
We introduce the maximum -times coverage problem that selects overlays
to maximize the summed coverage of weighted elements, where each element must
be covered at least times. We also define the min-cost -times coverage
problem where the objective is to select the minimum set of overlays such that
the sum of the weights of elements that are covered at least times is at
least . Maximum -times coverage is a generalization of the multi-set
multi-cover problem, is NP-complete, and is not submodular. We introduce two
new practical solutions for -times coverage based on integer linear
programming and sequential greedy optimization. We show that maximum -times
coverage is a natural way to frame peptide vaccine design, and find that it
produces a pan-strain COVID-19 vaccine design that is superior to 29 other
published designs in predicted population coverage and the expected number of
peptides displayed by each individual's HLA molecules.Comment: 10 pages, 5 figure
Probabilistic methods in combinatorial and stochastic optimization
Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Mathematics, 2005.Includes bibliographical references (leaves 103-106).(cont.) Packing/Covering problems, we prove upper and lower bounds on the adaptivity gap depending on the dimension. We also design polynomial-time algorithms achieving near-optimal approximation guarantees with respect to the adaptive optimum. Finally, we prove complexity-theoretic results regarding optimal adaptive policies. These results are based on a connection between adaptive policies and Arthur-Merlin games which yields PSPACE-hardness results for numerous questions regarding adaptive policies.In this thesis we study a variety of combinatorial problems with inherent randomness. In the first part of the thesis, we study the possibility of covering the solutions of an optimization problem on random subgraphs. The motivation for this approach is a situation where an optimization problem needs to be solved repeatedly for random instances. Then we seek a pre-processing stage which would speed-up subsequent queries by finding a fixed sparse subgraph covering the solution for a random subgraph with high probability. The first problem that we investigate is the minimum spanning tree. Our essential result regarding this problem is that for every graph with edge weights, there is a set of O(n log n) edges which contains the minimum spanning tree of a random subgraph with high probability. More generally, we extend this result to matroids. Further, we consider optimization problems based on the shortest path metric and we find covering sets of size 0(n(Ì1+2/c) log2Ì n) that approximate the shortest path metric of a random vertex-induced subgraph within a constant factor of c with high probability. In the second part, we turn to a model of stochastic optimization, where a solution is built sequentially by selecting a collection of "items". We distinguish between adaptive and non-adaptive strategies, where adaptivity means being able to perceive the precise characteristics of chosen items and use this knowledge in subsequent decisions. The benefit of adaptivity is our central concept which we investigate for a variety of specific problems. For the Stochastic Knapsack problem, we prove constant upper and lower bounds on the "adaptivity gap" between optimal adaptive and non-adaptive policies. For more general Stochasticby Jan Vondrák.Ph.D
Identical parallel machine scheduling problems: structural patterns, bounding techniques and solution procedures
The work is about fundamental parallel machine scheduling problems which occur in manufacturing systems where a set of jobs with individual processing times has to be assigned to a set of machines with respect to several workload objective functions like makespan minimization, machine covering or workload balancing. In the first chapter of the work an up-to-date survey on the most relevant literature for these problems is given, since the last review dealing with these problems has been published almost 20 years ago. We also give an insight into the relevant literature contributed by the Artificial Intelligence community, where the problem is known as number partitioning. The core of the work is a universally valid characterization of optimal makespan and machine-covering solutions where schedules are evaluated independently from the processing times of the jobs. Based on these novel structural insights we derive several strong dominance criteria. Implemented in a branch-and-bound algorithm these criteria have proved to be effective in limiting the solution space, particularly in the case of small ratios of the number of jobs to the number of machines. Further, we provide a counter-example to a central result by Ho et al. (2009) who proved that a schedule which minimizes the normalized sum of squared workload deviations is necessarily a makespan-optimal one. We explain why their proof is incorrect and present computational results revealing the difference between workload balancing and makespan minimization. The last chapter of the work is about the minimum cardinality bin covering problem which is a dual problem of machine-covering with respect to bounding techniques. We discuss reduction criteria, derive several lower bound arguments and propose construction heuristics as well as a subset sum-based improvement algorithm. Moreover, we present a tailored branch-and-bound method which is able to solve instances with up to 20 bins