28 research outputs found

    On a generalization of iterated and randomized rounding

    Get PDF
    We give a general method for rounding linear programs that combines the commonly used iterated rounding and randomized rounding techniques. In particular, we show that whenever iterated rounding can be applied to a problem with some slack, there is a randomized procedure that returns an integral solution that satisfies the guarantees of iterated rounding and also has concentration properties. We use this to give new results for several classic problems where iterated rounding has been useful

    Proximity results and faster algorithms for Integer Programming using the Steinitz Lemma

    Full text link
    We consider integer programming problems in standard form max{cTx:Ax=b,x0,xZn}\max \{c^Tx : Ax = b, \, x\geq 0, \, x \in Z^n\} where AZm×nA \in Z^{m \times n}, bZmb \in Z^m and cZnc \in Z^n. We show that such an integer program can be solved in time (mΔ)O(m)b2(m \Delta)^{O(m)} \cdot \|b\|_\infty^2, where Δ\Delta is an upper bound on each absolute value of an entry in AA. This improves upon the longstanding best bound of Papadimitriou (1981) of (mΔ)O(m2)(m\cdot \Delta)^{O(m^2)}, where in addition, the absolute values of the entries of bb also need to be bounded by Δ\Delta. Our result relies on a lemma of Steinitz that states that a set of vectors in RmR^m that is contained in the unit ball of a norm and that sum up to zero can be ordered such that all partial sums are of norm bounded by mm. We also use the Steinitz lemma to show that the 1\ell_1-distance of an optimal integer and fractional solution, also under the presence of upper bounds on the variables, is bounded by m(2mΔ+1)mm \cdot (2\,m \cdot \Delta+1)^m. Here Δ\Delta is again an upper bound on the absolute values of the entries of AA. The novel strength of our bound is that it is independent of nn. We provide evidence for the significance of our bound by applying it to general knapsack problems where we obtain structural and algorithmic results that improve upon the recent literature.Comment: We achieve much milder dependence of the running time on the largest entry in $b

    On a generalization of iterated and randomized rounding

    Get PDF
    We give a general method for rounding linear programs that combines the commonly used iterated rounding and randomized rounding techniques. In particular, we show that whenever iterated rounding can be applied to a problem with some slack, there is a randomized procedure that returns an integral solution that satisfies the guarantees of iterated rounding and also has concentration properties. We use this to give new results for several classic problems such as rounding column-sparse LPs, makespan minimization on unrelated machines, degree-bounded spanning trees and multi-budgeted matchings

    The Gram-Schmidt Walk: A Cure for the Banaszczyk Blues

    Get PDF
    A classic result of Banaszczyk (Random Str. & Algor. 1997) states that given any n vectors in Rm with ℓ2-norm at most 1 and any convex body K in Rm of Gaussian measure at least half, there exists a ±1 combination of these vectors that lies in 5K. Banaszczyk’s proof of this result was non-constructive and it was open how to find such a ±1 combination in polynomial time. In this paper, we give an efficient randomized algorithm to find a ±1 combination of the vectors which lies in cK for some fixed constant c > 0. This leads to new efficient algorithms for several problems in discrepancy theory

    Approximating Bin Packing within O(log OPT * log log OPT) bins

    Full text link
    For bin packing, the input consists of n items with sizes s_1,...,s_n in [0,1] which have to be assigned to a minimum number of bins of size 1. The seminal Karmarkar-Karp algorithm from '82 produces a solution with at most OPT + O(log^2 OPT) bins. We provide the first improvement in now 3 decades and show that one can find a solution of cost OPT + O(log OPT * log log OPT) in polynomial time. This is achieved by rounding a fractional solution to the Gilmore-Gomory LP relaxation using the Entropy Method from discrepancy theory. The result is constructive via algorithms of Bansal and Lovett-Meka
    corecore