65,087 research outputs found
Parameterized Approximation Schemes for Independent Set of Rectangles and Geometric Knapsack
The area of parameterized approximation seeks to combine approximation and parameterized algorithms to obtain, e.g., (1+epsilon)-approximations in f(k,epsilon)n^O(1) time where k is some parameter of the input. The goal is to overcome lower bounds from either of the areas. We obtain the following results on parameterized approximability:
- In the maximum independent set of rectangles problem (MISR) we are given a collection of n axis parallel rectangles in the plane. Our goal is to select a maximum-cardinality subset of pairwise non-overlapping rectangles. This problem is NP-hard and also W[1]-hard [Marx, ESA\u2705]. The best-known polynomial-time approximation factor is O(log log n) [Chalermsook and Chuzhoy, SODA\u2709] and it admits a QPTAS [Adamaszek and Wiese, FOCS\u2713; Chuzhoy and Ene, FOCS\u2716]. Here we present a parameterized approximation scheme (PAS) for MISR, i.e. an algorithm that, for any given constant epsilon>0 and integer k>0, in time f(k,epsilon)n^g(epsilon), either outputs a solution of size at least k/(1+epsilon), or declares that the optimum solution has size less than k.
- In the (2-dimensional) geometric knapsack problem (2DK) we are given an axis-aligned square knapsack and a collection of axis-aligned rectangles in the plane (items). Our goal is to translate a maximum cardinality subset of items into the knapsack so that the selected items do not overlap. In the version of 2DK with rotations (2DKR), we are allowed to rotate items by 90 degrees. Both variants are NP-hard, and the best-known polynomial-time approximation factor is 2+epsilon [Jansen and Zhang, SODA\u2704]. These problems admit a QPTAS for polynomially bounded item sizes [Adamaszek and Wiese, SODA\u2715]. We show that both variants are W[1]-hard. Furthermore, we present a PAS for 2DKR.
For all considered problems, getting time f(k,epsilon)n^O(1), rather than f(k,epsilon)n^g(epsilon), would give FPT time f\u27(k)n^O(1) exact algorithms by setting epsilon=1/(k+1), contradicting W[1]-hardness. Instead, for each fixed epsilon>0, our PASs give (1+epsilon)-approximate solutions in FPT time.
For both MISR and 2DKR our techniques also give rise to preprocessing algorithms that take n^g(epsilon) time and return a subset of at most k^g(epsilon) rectangles/items that contains a solution of size at least k/(1+epsilon) if a solution of size k exists. This is a special case of the recently introduced notion of a polynomial-size approximate kernelization scheme [Lokshtanov et al., STOC\u2717]
Efficient Parameterized Algorithms for Computing All-Pairs Shortest Paths
Computing all-pairs shortest paths is a fundamental and much-studied problem
with many applications. Unfortunately, despite intense study, there are still
no significantly faster algorithms for it than the time
algorithm due to Floyd and Warshall (1962). Somewhat faster algorithms exist
for the vertex-weighted version if fast matrix multiplication may be used.
Yuster (SODA 2009) gave an algorithm running in time ,
but no combinatorial, truly subcubic algorithm is known.
Motivated by the recent framework of efficient parameterized algorithms (or
"FPT in P"), we investigate the influence of the graph parameters clique-width
() and modular-width () on the running times of algorithms for solving
All-Pairs Shortest Paths. We obtain efficient (and combinatorial) parameterized
algorithms on non-negative vertex-weighted graphs of times
, resp. . If fast matrix
multiplication is allowed then the latter can be improved to
using the algorithm of Yuster as a black box.
The algorithm relative to modular-width is adaptive, meaning that the running
time matches the best unparameterized algorithm for parameter value equal
to , and they outperform them already for for any
On space efficiency of algorithms working on structural decompositions of graphs
Dynamic programming on path and tree decompositions of graphs is a technique
that is ubiquitous in the field of parameterized and exponential-time
algorithms. However, one of its drawbacks is that the space usage is
exponential in the decomposition's width. Following the work of Allender et al.
[Theory of Computing, '14], we investigate whether this space complexity
explosion is unavoidable. Using the idea of reparameterization of Cai and
Juedes [J. Comput. Syst. Sci., '03], we prove that the question is closely
related to a conjecture that the Longest Common Subsequence problem
parameterized by the number of input strings does not admit an algorithm that
simultaneously uses XP time and FPT space. Moreover, we complete the complexity
landscape sketched for pathwidth and treewidth by Allender et al. by
considering the parameter tree-depth. We prove that computations on tree-depth
decompositions correspond to a model of non-deterministic machines that work in
polynomial time and logarithmic space, with access to an auxiliary stack of
maximum height equal to the decomposition's depth. Together with the results of
Allender et al., this describes a hierarchy of complexity classes for
polynomial-time non-deterministic machines with different restrictions on the
access to working space, which mirrors the classic relations between treewidth,
pathwidth, and tree-depth.Comment: An extended abstract appeared in the proceedings of STACS'16. The new
version is augmented with a space-efficient algorithm for Dominating Set
using the Chinese remainder theore
- …