8 research outputs found

    Approximation Algorithms for Covering/Packing Integer Programs

    Get PDF
    Given matrices A and B and vectors a, b, c and d, all with non-negative entries, we consider the problem of computing min {c.x: x in Z^n_+, Ax > a, Bx < b, x < d}. We give a bicriteria-approximation algorithm that, given epsilon in (0, 1], finds a solution of cost O(ln(m)/epsilon^2) times optimal, meeting the covering constraints (Ax > a) and multiplicity constraints (x < d), and satisfying Bx < (1 + epsilon)b + beta, where beta is the vector of row sums beta_i = sum_j B_ij. Here m denotes the number of rows of A. This gives an O(ln m)-approximation algorithm for CIP -- minimum-cost covering integer programs with multiplicity constraints, i.e., the special case when there are no packing constraints Bx < b. The previous best approximation ratio has been O(ln(max_j sum_i A_ij)) since 1982. CIP contains the set cover problem as a special case, so O(ln m)-approximation is the best possible unless P=NP.Comment: Preliminary version appeared in IEEE Symposium on Foundations of Computer Science (2001). To appear in Journal of Computer and System Science

    Maximum n-times Coverage for Vaccine Design

    Full text link
    We introduce the maximum nn-times coverage problem that selects kk overlays to maximize the summed coverage of weighted elements, where each element must be covered at least nn times. We also define the min-cost nn-times coverage problem where the objective is to select the minimum set of overlays such that the sum of the weights of elements that are covered at least nn times is at least τ\tau. Maximum nn-times coverage is a generalization of the multi-set multi-cover problem, is NP-complete, and is not submodular. We introduce two new practical solutions for nn-times coverage based on integer linear programming and sequential greedy optimization. We show that maximum nn-times coverage is a natural way to frame peptide vaccine design, and find that it produces a pan-strain COVID-19 vaccine design that is superior to 29 other published designs in predicted population coverage and the expected number of peptides displayed by each individual's HLA molecules.Comment: 10 pages, 5 figure

    Probabilistic methods in combinatorial and stochastic optimization

    Get PDF
    Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Mathematics, 2005.Includes bibliographical references (leaves 103-106).(cont.) Packing/Covering problems, we prove upper and lower bounds on the adaptivity gap depending on the dimension. We also design polynomial-time algorithms achieving near-optimal approximation guarantees with respect to the adaptive optimum. Finally, we prove complexity-theoretic results regarding optimal adaptive policies. These results are based on a connection between adaptive policies and Arthur-Merlin games which yields PSPACE-hardness results for numerous questions regarding adaptive policies.In this thesis we study a variety of combinatorial problems with inherent randomness. In the first part of the thesis, we study the possibility of covering the solutions of an optimization problem on random subgraphs. The motivation for this approach is a situation where an optimization problem needs to be solved repeatedly for random instances. Then we seek a pre-processing stage which would speed-up subsequent queries by finding a fixed sparse subgraph covering the solution for a random subgraph with high probability. The first problem that we investigate is the minimum spanning tree. Our essential result regarding this problem is that for every graph with edge weights, there is a set of O(n log n) edges which contains the minimum spanning tree of a random subgraph with high probability. More generally, we extend this result to matroids. Further, we consider optimization problems based on the shortest path metric and we find covering sets of size 0(n(Ì1+2/c) log2Ì n) that approximate the shortest path metric of a random vertex-induced subgraph within a constant factor of c with high probability. In the second part, we turn to a model of stochastic optimization, where a solution is built sequentially by selecting a collection of "items". We distinguish between adaptive and non-adaptive strategies, where adaptivity means being able to perceive the precise characteristics of chosen items and use this knowledge in subsequent decisions. The benefit of adaptivity is our central concept which we investigate for a variety of specific problems. For the Stochastic Knapsack problem, we prove constant upper and lower bounds on the "adaptivity gap" between optimal adaptive and non-adaptive policies. For more general Stochasticby Jan Vondrák.Ph.D

    Identical parallel machine scheduling problems: structural patterns, bounding techniques and solution procedures

    Get PDF
    The work is about fundamental parallel machine scheduling problems which occur in manufacturing systems where a set of jobs with individual processing times has to be assigned to a set of machines with respect to several workload objective functions like makespan minimization, machine covering or workload balancing. In the first chapter of the work an up-to-date survey on the most relevant literature for these problems is given, since the last review dealing with these problems has been published almost 20 years ago. We also give an insight into the relevant literature contributed by the Artificial Intelligence community, where the problem is known as number partitioning. The core of the work is a universally valid characterization of optimal makespan and machine-covering solutions where schedules are evaluated independently from the processing times of the jobs. Based on these novel structural insights we derive several strong dominance criteria. Implemented in a branch-and-bound algorithm these criteria have proved to be effective in limiting the solution space, particularly in the case of small ratios of the number of jobs to the number of machines. Further, we provide a counter-example to a central result by Ho et al. (2009) who proved that a schedule which minimizes the normalized sum of squared workload deviations is necessarily a makespan-optimal one. We explain why their proof is incorrect and present computational results revealing the difference between workload balancing and makespan minimization. The last chapter of the work is about the minimum cardinality bin covering problem which is a dual problem of machine-covering with respect to bounding techniques. We discuss reduction criteria, derive several lower bound arguments and propose construction heuristics as well as a subset sum-based improvement algorithm. Moreover, we present a tailored branch-and-bound method which is able to solve instances with up to 20 bins
    corecore