1,233 research outputs found
Solving Hard Computational Problems Efficiently: Asymptotic Parametric Complexity 3-Coloring Algorithm
Many practical problems in almost all scientific and technological
disciplines have been classified as computationally hard (NP-hard or even
NP-complete). In life sciences, combinatorial optimization problems frequently
arise in molecular biology, e.g., genome sequencing; global alignment of
multiple genomes; identifying siblings or discovery of dysregulated pathways.In
almost all of these problems, there is the need for proving a hypothesis about
certain property of an object that can be present only when it adopts some
particular admissible structure (an NP-certificate) or be absent (no admissible
structure), however, none of the standard approaches can discard the hypothesis
when no solution can be found, since none can provide a proof that there is no
admissible structure. This article presents an algorithm that introduces a
novel type of solution method to "efficiently" solve the graph 3-coloring
problem; an NP-complete problem. The proposed method provides certificates
(proofs) in both cases: present or absent, so it is possible to accept or
reject the hypothesis on the basis of a rigorous proof. It provides exact
solutions and is polynomial-time (i.e., efficient) however parametric. The only
requirement is sufficient computational power, which is controlled by the
parameter . Nevertheless, here it is proved that the
probability of requiring a value of to obtain a solution for a
random graph decreases exponentially: , making
tractable almost all problem instances. Thorough experimental analyses were
performed. The algorithm was tested on random graphs, planar graphs and
4-regular planar graphs. The obtained experimental results are in accordance
with the theoretical expected results.Comment: Working pape
A New Look at the Easy-Hard-Easy Pattern of Combinatorial Search Difficulty
The easy-hard-easy pattern in the difficulty of combinatorial search problems
as constraints are added has been explained as due to a competition between the
decrease in number of solutions and increased pruning. We test the generality
of this explanation by examining one of its predictions: if the number of
solutions is held fixed by the choice of problems, then increased pruning
should lead to a monotonic decrease in search cost. Instead, we find the
easy-hard-easy pattern in median search cost even when the number of solutions
is held constant, for some search methods. This generalizes previous
observations of this pattern and shows that the existing theory does not
explain the full range of the peak in search cost. In these cases the pattern
appears to be due to changes in the size of the minimal unsolvable subproblems,
rather than changing numbers of solutions.Comment: See http://www.jair.org/ for any accompanying file
The backtracking survey propagation algorithm for solving random K-SAT problems
Discrete combinatorial optimization has a central role in many scientific
disciplines, however, for hard problems we lack linear time algorithms that
would allow us to solve very large instances. Moreover, it is still unclear
what are the key features that make a discrete combinatorial optimization
problem hard to solve. Here we study random K-satisfiability problems with
, which are known to be very hard close to the SAT-UNSAT threshold,
where problems stop having solutions. We show that the backtracking survey
propagation algorithm, in a time practically linear in the problem size, is
able to find solutions very close to the threshold, in a region unreachable by
any other algorithm. All solutions found have no frozen variables, thus
supporting the conjecture that only unfrozen solutions can be found in linear
time, and that a problem becomes impossible to solve in linear time when all
solutions contain frozen variables.Comment: 11 pages, 10 figures. v2: data largely improved and manuscript
rewritte
- …