7,567 research outputs found
Kernelization Lower Bounds By Cross-Composition
We introduce the cross-composition framework for proving kernelization lower
bounds. A classical problem L AND/OR-cross-composes into a parameterized
problem Q if it is possible to efficiently construct an instance of Q with
polynomially bounded parameter value that expresses the logical AND or OR of a
sequence of instances of L. Building on work by Bodlaender et al. (ICALP 2008)
and using a result by Fortnow and Santhanam (STOC 2008) with a refinement by
Dell and van Melkebeek (STOC 2010), we show that if an NP-hard problem
OR-cross-composes into a parameterized problem Q then Q does not admit a
polynomial kernel unless NP \subseteq coNP/poly and the polynomial hierarchy
collapses. Similarly, an AND-cross-composition for Q rules out polynomial
kernels for Q under Bodlaender et al.'s AND-distillation conjecture.
Our technique generalizes and strengthens the recent techniques of using
composition algorithms and of transferring the lower bounds via polynomial
parameter transformations. We show its applicability by proving kernelization
lower bounds for a number of important graphs problems with structural
(non-standard) parameterizations, e.g., Clique, Chromatic Number, Weighted
Feedback Vertex Set, and Weighted Odd Cycle Transversal do not admit polynomial
kernels with respect to the vertex cover number of the input graphs unless the
polynomial hierarchy collapses, contrasting the fact that these problems are
trivially fixed-parameter tractable for this parameter.
After learning of our results, several teams of authors have successfully
applied the cross-composition framework to different parameterized problems.
For completeness, our presentation of the framework includes several extensions
based on this follow-up work. For example, we show how a relaxed version of
OR-cross-compositions may be used to give lower bounds on the degree of the
polynomial in the kernel size.Comment: A preliminary version appeared in the proceedings of the 28th
International Symposium on Theoretical Aspects of Computer Science (STACS
2011) under the title "Cross-Composition: A New Technique for Kernelization
Lower Bounds". Several results have been strengthened compared to the
preliminary version (http://arxiv.org/abs/1011.4224). 29 pages, 2 figure
Cross-Composition: A New Technique for Kernelization Lower Bounds
We introduce a new technique for proving kernelization lower bounds, called
cross-composition. A classical problem L cross-composes into a parameterized
problem Q if an instance of Q with polynomially bounded parameter value can
express the logical OR of a sequence of instances of L. Building on work by
Bodlaender et al. (ICALP 2008) and using a result by Fortnow and Santhanam
(STOC 2008) we show that if an NP-complete problem cross-composes into a
parameterized problem Q then Q does not admit a polynomial kernel unless the
polynomial hierarchy collapses. Our technique generalizes and strengthens the
recent techniques of using OR-composition algorithms and of transferring the
lower bounds via polynomial parameter transformations. We show its
applicability by proving kernelization lower bounds for a number of important
graphs problems with structural (non-standard) parameterizations, e.g.,
Chromatic Number, Clique, and Weighted Feedback Vertex Set do not admit
polynomial kernels with respect to the vertex cover number of the input graphs
unless the polynomial hierarchy collapses, contrasting the fact that these
problems are trivially fixed-parameter tractable for this parameter. We have
similar lower bounds for Feedback Vertex Set.Comment: Updated information based on final version submitted to STACS 201
Completeness of classical spin models and universal quantum computation
We study mappings between distinct classical spin systems that leave the
partition function invariant. As recently shown in [Phys. Rev. Lett. 100,
110501 (2008)], the partition function of the 2D square lattice Ising model in
the presence of an inhomogeneous magnetic field, can specialize to the
partition function of any Ising system on an arbitrary graph. In this sense the
2D Ising model is said to be "complete". However, in order to obtain the above
result, the coupling strengths on the 2D lattice must assume complex values,
and thus do not allow for a physical interpretation. Here we show how a
complete model with real -and, hence, "physical"- couplings can be obtained if
the 3D Ising model is considered. We furthermore show how to map general
q-state systems with possibly many-body interactions to the 2D Ising model with
complex parameters, and give completeness results for these models with real
parameters. We also demonstrate that the computational overhead in these
constructions is in all relevant cases polynomial. These results are proved by
invoking a recently found cross-connection between statistical mechanics and
quantum information theory, where partition functions are expressed as quantum
mechanical amplitudes. Within this framework, there exists a natural
correspondence between many-body quantum states that allow universal quantum
computation via local measurements only, and complete classical spin systems.Comment: 43 pages, 28 figure
Fast Dynamic Graph Algorithms for Parameterized Problems
Fully dynamic graph is a data structure that (1) supports edge insertions and
deletions and (2) answers problem specific queries. The time complexity of (1)
and (2) are referred to as the update time and the query time respectively.
There are many researches on dynamic graphs whose update time and query time
are , that is, sublinear in the graph size. However, almost all such
researches are for problems in P. In this paper, we investigate dynamic graphs
for NP-hard problems exploiting the notion of fixed parameter tractability
(FPT).
We give dynamic graphs for Vertex Cover and Cluster Vertex Deletion
parameterized by the solution size . These dynamic graphs achieve almost the
best possible update time and the query time
, where is the time complexity of any static
graph algorithm for the problems. We obtain these results by dynamically
maintaining an approximate solution which can be used to construct a small
problem kernel. Exploiting the dynamic graph for Cluster Vertex Deletion, as a
corollary, we obtain a quasilinear-time (polynomial) kernelization algorithm
for Cluster Vertex Deletion. Until now, only quadratic time kernelization
algorithms are known for this problem.
We also give a dynamic graph for Chromatic Number parameterized by the
solution size of Cluster Vertex Deletion, and a dynamic graph for
bounded-degree Feedback Vertex Set parameterized by the solution size. Assuming
the parameter is a constant, each dynamic graph can be updated in
time and can compute a solution in time. These results are obtained by
another approach.Comment: SWAT 2014 to appea
Structural parameterizations for boxicity
The boxicity of a graph is the least integer such that has an
intersection model of axis-aligned -dimensional boxes. Boxicity, the problem
of deciding whether a given graph has boxicity at most , is NP-complete
for every fixed . We show that boxicity is fixed-parameter tractable
when parameterized by the cluster vertex deletion number of the input graph.
This generalizes the result of Adiga et al., that boxicity is fixed-parameter
tractable in the vertex cover number.
Moreover, we show that boxicity admits an additive -approximation when
parameterized by the pathwidth of the input graph.
Finally, we provide evidence in favor of a conjecture of Adiga et al. that
boxicity remains NP-complete when parameterized by the treewidth.Comment: 19 page
Fast Biclustering by Dual Parameterization
We study two clustering problems, Starforest Editing, the problem of adding
and deleting edges to obtain a disjoint union of stars, and the generalization
Bicluster Editing. We show that, in addition to being NP-hard, none of the
problems can be solved in subexponential time unless the exponential time
hypothesis fails.
Misra, Panolan, and Saurabh (MFCS 2013) argue that introducing a bound on the
number of connected components in the solution should not make the problem
easier: In particular, they argue that the subexponential time algorithm for
editing to a fixed number of clusters (p-Cluster Editing) by Fomin et al. (J.
Comput. Syst. Sci., 80(7) 2014) is an exception rather than the rule. Here, p
is a secondary parameter, bounding the number of components in the solution.
However, upon bounding the number of stars or bicliques in the solution, we
obtain algorithms which run in time for p-Starforest
Editing and for p-Bicluster Editing. We
obtain a similar result for the more general case of t-Partite p-Cluster
Editing. This is subexponential in k for fixed number of clusters, since p is
then considered a constant.
Our results even out the number of multivariate subexponential time
algorithms and give reasons to believe that this area warrants further study.Comment: Accepted for presentation at IPEC 201
- …