25,632 research outputs found
Covering Many (Or Few) Edges with k Vertices in Sparse Graphs
We study the following two fixed-cardinality optimization problems (a maximization and a minimization variant). For a fixed ? between zero and one we are given a graph and two numbers k ? ? and t ? ?. The task is to find a vertex subset S of exactly k vertices that has value at least (resp. at most for minimization) t. Here, the value of a vertex set computes as ? times the number of edges with exactly one endpoint in S plus 1-? times the number of edges with both endpoints in S. These two problems generalize many prominent graph problems, such as Densest k-Subgraph, Sparsest k-Subgraph, Partial Vertex Cover, and Max (k,n-k)-Cut.
In this work, we complete the picture of their parameterized complexity on several types of sparse graphs that are described by structural parameters. In particular, we provide kernelization algorithms and kernel lower bounds for these problems. A somewhat surprising consequence of our kernelizations is that Partial Vertex Cover and Max (k,n-k)-Cut not only behave in the same way but that the kernels for both problems can be obtained by the same algorithms
Solving Connectivity Problems Parameterized by Treedepth in Single-Exponential Time and Polynomial Space
A breakthrough result of Cygan et al. (FOCS 2011) showed that connectivity problems parameterized by treewidth can be solved much faster than the previously best known time ?^*(2^{?(twlog tw)}). Using their inspired Cut&Count technique, they obtained ?^*(?^tw) time algorithms for many such problems. Moreover, they proved these running times to be optimal assuming the Strong Exponential-Time Hypothesis. Unfortunately, like other dynamic programming algorithms on tree decompositions, these algorithms also require exponential space, and this is widely believed to be unavoidable. In contrast, for the slightly larger parameter called treedepth, there are already several examples of matching the time bounds obtained for treewidth, but using only polynomial space. Nevertheless, this has remained open for connectivity problems.
In the present work, we close this knowledge gap by applying the Cut&Count technique to graphs of small treedepth. While the general idea is unchanged, we have to design novel procedures for counting consistently cut solution candidates using only polynomial space. Concretely, we obtain time ?^*(3^d) and polynomial space for Connected Vertex Cover, Feedback Vertex Set, and Steiner Tree on graphs of treedepth d. Similarly, we obtain time ?^*(4^d) and polynomial space for Connected Dominating Set and Connected Odd Cycle Transversal
Fast Parallel Fixed-Parameter Algorithms via Color Coding
Fixed-parameter algorithms have been successfully applied to solve numerous
difficult problems within acceptable time bounds on large inputs. However, most
fixed-parameter algorithms are inherently \emph{sequential} and, thus, make no
use of the parallel hardware present in modern computers. We show that parallel
fixed-parameter algorithms do not only exist for numerous parameterized
problems from the literature -- including vertex cover, packing problems,
cluster editing, cutting vertices, finding embeddings, or finding matchings --
but that there are parallel algorithms working in \emph{constant} time or at
least in time \emph{depending only on the parameter} (and not on the size of
the input) for these problems. Phrased in terms of complexity classes, we place
numerous natural parameterized problems in parameterized versions of AC. On
a more technical level, we show how the \emph{color coding} method can be
implemented in constant time and apply it to embedding problems for graphs of
bounded tree-width or tree-depth and to model checking first-order formulas in
graphs of bounded degree
Parameterized Approximation Schemes using Graph Widths
Combining the techniques of approximation algorithms and parameterized
complexity has long been considered a promising research area, but relatively
few results are currently known. In this paper we study the parameterized
approximability of a number of problems which are known to be hard to solve
exactly when parameterized by treewidth or clique-width. Our main contribution
is to present a natural randomized rounding technique that extends well-known
ideas and can be used for both of these widths. Applying this very generic
technique we obtain approximation schemes for a number of problems, evading
both polynomial-time inapproximability and parameterized intractability bounds
Parameterized complexity of machine scheduling: 15 open problems
Machine scheduling problems are a long-time key domain of algorithms and
complexity research. A novel approach to machine scheduling problems are
fixed-parameter algorithms. To stimulate this thriving research direction, we
propose 15 open questions in this area whose resolution we expect to lead to
the discovery of new approaches and techniques both in scheduling and
parameterized complexity theory.Comment: Version accepted to Computers & Operations Researc
Parameterized Algorithmics for Computational Social Choice: Nine Research Challenges
Computational Social Choice is an interdisciplinary research area involving
Economics, Political Science, and Social Science on the one side, and
Mathematics and Computer Science (including Artificial Intelligence and
Multiagent Systems) on the other side. Typical computational problems studied
in this field include the vulnerability of voting procedures against attacks,
or preference aggregation in multi-agent systems. Parameterized Algorithmics is
a subfield of Theoretical Computer Science seeking to exploit meaningful
problem-specific parameters in order to identify tractable special cases of in
general computationally hard problems. In this paper, we propose nine of our
favorite research challenges concerning the parameterized complexity of
problems appearing in this context
Lossy Kernelization
In this paper we propose a new framework for analyzing the performance of
preprocessing algorithms. Our framework builds on the notion of kernelization
from parameterized complexity. However, as opposed to the original notion of
kernelization, our definitions combine well with approximation algorithms and
heuristics. The key new definition is that of a polynomial size
-approximate kernel. Loosely speaking, a polynomial size
-approximate kernel is a polynomial time pre-processing algorithm that
takes as input an instance to a parameterized problem, and outputs
another instance to the same problem, such that . Additionally, for every , a -approximate solution
to the pre-processed instance can be turned in polynomial time into a
-approximate solution to the original instance .
Our main technical contribution are -approximate kernels of
polynomial size for three problems, namely Connected Vertex Cover, Disjoint
Cycle Packing and Disjoint Factors. These problems are known not to admit any
polynomial size kernels unless . Our approximate
kernels simultaneously beat both the lower bounds on the (normal) kernel size,
and the hardness of approximation lower bounds for all three problems. On the
negative side we prove that Longest Path parameterized by the length of the
path and Set Cover parameterized by the universe size do not admit even an
-approximate kernel of polynomial size, for any , unless
. In order to prove this lower bound we need to combine
in a non-trivial way the techniques used for showing kernelization lower bounds
with the methods for showing hardness of approximationComment: 58 pages. Version 2 contain new results: PSAKS for Cycle Packing and
approximate kernel lower bounds for Set Cover and Hitting Set parameterized
by universe siz
Parameterized Complexity of Edge Interdiction Problems
We study the parameterized complexity of interdiction problems in graphs. For
an optimization problem on graphs, one can formulate an interdiction problem as
a game consisting of two players, namely, an interdictor and an evader, who
compete on an objective with opposing interests. In edge interdiction problems,
every edge of the input graph has an interdiction cost associated with it and
the interdictor interdicts the graph by modifying the edges in the graph, and
the number of such modifications is constrained by the interdictor's budget.
The evader then solves the given optimization problem on the modified graph.
The action of the interdictor must impede the evader as much as possible. We
focus on edge interdiction problems related to minimum spanning tree, maximum
matching and shortest paths. These problems arise in different real world
scenarios. We derive several fixed-parameter tractability and W[1]-hardness
results for these interdiction problems with respect to various parameters.
Next, we show close relation between interdiction problems and partial cover
problems on bipartite graphs where the goal is not to cover all elements but to
minimize/maximize the number of covered elements with specific number of sets.
Hereby, we investigate the parameterized complexity of several partial cover
problems on bipartite graphs
- …