190 research outputs found

    On the Approximability of the Traveling Salesman Problem with Line Neighborhoods

    Get PDF
    We study the variant of the Euclidean Traveling Salesman problem where instead of a set of points, we are given a set of lines as input, and the goal is to find the shortest tour that visits each line. The best known upper and lower bounds for the problem in Rd\mathbb{R}^d, with d3d\ge 3, are NP\mathrm{NP}-hardness and an O(log3n)O(\log^3 n)-approximation algorithm which is based on a reduction to the group Steiner tree problem. We show that TSP with lines in Rd\mathbb{R}^d is APX-hard for any d3d\ge 3. More generally, this implies that TSP with kk-dimensional flats does not admit a PTAS for any 1kd21\le k \leq d-2 unless P=NP\mathrm{P}=\mathrm{NP}, which gives a complete classification of the approximability of these problems, as there are known PTASes for k=0k=0 (i.e., points) and k=d1k=d-1 (hyperplanes). We are able to give a stronger inapproximability factor for d=O(logn)d=O(\log n) by showing that TSP with lines does not admit a (2ϵ)(2-\epsilon)-approximation in dd dimensions under the unique games conjecture. On the positive side, we leverage recent results on restricted variants of the group Steiner tree problem in order to give an O(log2n)O(\log^2 n)-approximation algorithm for the problem, albeit with a running time of nO(loglogn)n^{O(\log\log n)}

    Parameterized Approximation Schemes for Independent Set of Rectangles and Geometric Knapsack

    Get PDF
    The area of parameterized approximation seeks to combine approximation and parameterized algorithms to obtain, e.g., (1+epsilon)-approximations in f(k,epsilon)n^O(1) time where k is some parameter of the input. The goal is to overcome lower bounds from either of the areas. We obtain the following results on parameterized approximability: - In the maximum independent set of rectangles problem (MISR) we are given a collection of n axis parallel rectangles in the plane. Our goal is to select a maximum-cardinality subset of pairwise non-overlapping rectangles. This problem is NP-hard and also W[1]-hard [Marx, ESA\u2705]. The best-known polynomial-time approximation factor is O(log log n) [Chalermsook and Chuzhoy, SODA\u2709] and it admits a QPTAS [Adamaszek and Wiese, FOCS\u2713; Chuzhoy and Ene, FOCS\u2716]. Here we present a parameterized approximation scheme (PAS) for MISR, i.e. an algorithm that, for any given constant epsilon>0 and integer k>0, in time f(k,epsilon)n^g(epsilon), either outputs a solution of size at least k/(1+epsilon), or declares that the optimum solution has size less than k. - In the (2-dimensional) geometric knapsack problem (2DK) we are given an axis-aligned square knapsack and a collection of axis-aligned rectangles in the plane (items). Our goal is to translate a maximum cardinality subset of items into the knapsack so that the selected items do not overlap. In the version of 2DK with rotations (2DKR), we are allowed to rotate items by 90 degrees. Both variants are NP-hard, and the best-known polynomial-time approximation factor is 2+epsilon [Jansen and Zhang, SODA\u2704]. These problems admit a QPTAS for polynomially bounded item sizes [Adamaszek and Wiese, SODA\u2715]. We show that both variants are W[1]-hard. Furthermore, we present a PAS for 2DKR. For all considered problems, getting time f(k,epsilon)n^O(1), rather than f(k,epsilon)n^g(epsilon), would give FPT time f\u27(k)n^O(1) exact algorithms by setting epsilon=1/(k+1), contradicting W[1]-hardness. Instead, for each fixed epsilon>0, our PASs give (1+epsilon)-approximate solutions in FPT time. For both MISR and 2DKR our techniques also give rise to preprocessing algorithms that take n^g(epsilon) time and return a subset of at most k^g(epsilon) rectangles/items that contains a solution of size at least k/(1+epsilon) if a solution of size k exists. This is a special case of the recently introduced notion of a polynomial-size approximate kernelization scheme [Lokshtanov et al., STOC\u2717]

    Covering Problems via Structural Approaches

    Get PDF
    The minimum set cover problem is, without question, among the most ubiquitous and well-studied problems in computer science. Its theoretical hardness has been fully characterized--logarithmic approximability has been established, and no sublogarithmic approximation exists unless P=NP. However, the gap between real-world instances and the theoretical worst case is often immense--many covering problems of practical relevance admit much better approximations, or even solvability in polynomial time. Simple combinatorial or geometric structure can often be exploited to obtain improved algorithms on a problem-by-problem basis, but there is no general method of determining the extent to which this is possible. In this thesis, we aim to shed light on the relationship between the structure and the hardness of covering problems. We discuss several measures of structural complexity of set cover instances and prove new algorithmic and hardness results linking the approximability of a set cover problem to its underlying structure. In particular, we provide: - An APX-hardness proof for a wide family of problems that encode a simple covering problem known as Special-3SC. - A class of polynomial dynamic programming algorithms for a group of weighted geometric set cover problems having simple structure. - A simplified quasi-uniform sampling algorithm that yields improved approximations for weighted covering problems having low cell complexity or geometric union complexity. - Applications of the above to various capacitated covering problems via linear programming strengthening and rounding. In total, we obtain new results for dozens of covering problems exhibiting geometric or combinatorial structure. We tabulate these problems and classify them according to their approximability

    On the Approximability of the Traveling Salesman Problem with Line Neighborhoods

    Get PDF
    We study the variant of the Euclidean Traveling Salesman problem where instead of a set of points, we are given a set of lines as input, and the goal is to find the shortest tour that visits each line. The best known upper and lower bounds for the problem in Rd\mathbb{R}^d, with d3d\ge 3, are NP\mathrm{NP}-hardness and an O(log3n)O(\log^3 n)-approximation algorithm which is based on a reduction to the group Steiner tree problem. We show that TSP with lines in Rd\mathbb{R}^d is APX-hard for any d3d\ge 3. More generally, this implies that TSP with kk-dimensional flats does not admit a PTAS for any 1kd21\le k \leq d-2 unless P=NP\mathrm{P}=\mathrm{NP}, which gives a complete classification of the approximability of these problems, as there are known PTASes for k=0k=0 (i.e., points) and k=d1k=d-1 (hyperplanes). We are able to give a stronger inapproximability factor for d=O(logn)d=O(\log n) by showing that TSP with lines does not admit a (2ϵ)(2-\epsilon)-approximation in dd dimensions under the unique games conjecture. On the positive side, we leverage recent results on restricted variants of the group Steiner tree problem in order to give an O(log2n)O(\log^2 n)-approximation algorithm for the problem, albeit with a running time of nO(loglogn)n^{O(\log\log n)}

    The Bane of Low-Dimensionality Clustering

    Get PDF
    In this paper, we give a conditional lower bound of nΩ(k)n^{\Omega(k)} on running time for the classic k-median and k-means clustering objectives (where n is the size of the input), even in low-dimensional Euclidean space of dimension four, assuming the Exponential Time Hypothesis (ETH). We also consider k-median (and k-means) with penalties where each point need not be assigned to a center, in which case it must pay a penalty, and extend our lower bound to at least three-dimensional Euclidean space. This stands in stark contrast to many other geometric problems such as the traveling salesman problem, or computing an independent set of unit spheres. While these problems benefit from the so-called (limited) blessing of dimensionality, as they can be solved in time nO(k11/d)n^{O(k^{1-1/d})} or 2n11/d2^{n^{1-1/d}} in d dimensions, our work shows that widely-used clustering objectives have a lower bound of nΩ(k)n^{\Omega(k)}, even in dimension four. We complete the picture by considering the two-dimensional case: we show that there is no algorithm that solves the penalized version in time less than no(k)n^{o(\sqrt{k})}, and provide a matching upper bound of nO(k)n^{O(\sqrt{k})}. The main tool we use to establish these lower bounds is the placement of points on the moment curve, which takes its inspiration from constructions of point sets yielding Delaunay complexes of high complexity
    corecore