4,003 research outputs found
Phase Transitions of the Typical Algorithmic Complexity of the Random Satisfiability Problem Studied with Linear Programming
Here we study the NP-complete -SAT problem. Although the worst-case
complexity of NP-complete problems is conjectured to be exponential, there
exist parametrized random ensembles of problems where solutions can typically
be found in polynomial time for suitable ranges of the parameter. In fact,
random -SAT, with as control parameter, can be solved quickly
for small enough values of . It shows a phase transition between a
satisfiable phase and an unsatisfiable phase. For branch and bound algorithms,
which operate in the space of feasible Boolean configurations, the empirically
hardest problems are located only close to this phase transition. Here we study
-SAT () and the related optimization problem MAX-SAT by a linear
programming approach, which is widely used for practical problems and allows
for polynomial run time. In contrast to branch and bound it operates outside
the space of feasible configurations. On the other hand, finding a solution
within polynomial time is not guaranteed. We investigated several variants like
including artificial objective functions, so called cutting-plane approaches,
and a mapping to the NP-complete vertex-cover problem. We observed several
easy-hard transitions, from where the problems are typically solvable (in
polynomial time) using the given algorithms, respectively, to where they are
not solvable in polynomial time. For the related vertex-cover problem on random
graphs these easy-hard transitions can be identified with structural properties
of the graphs, like percolation transitions. For the present random -SAT
problem we have investigated numerous structural properties also exhibiting
clear transitions, but they appear not be correlated to the here observed
easy-hard transitions. This renders the behaviour of random -SAT more
complex than, e.g., the vertex-cover problem.Comment: 11 pages, 5 figure
Quantum adiabatic optimization and combinatorial landscapes
In this paper we analyze the performance of the Quantum Adiabatic Evolution
algorithm on a variant of Satisfiability problem for an ensemble of random
graphs parametrized by the ratio of clauses to variables, . We
introduce a set of macroscopic parameters (landscapes) and put forward an
ansatz of universality for random bit flips. We then formulate the problem of
finding the smallest eigenvalue and the excitation gap as a statistical
mechanics problem. We use the so-called annealing approximation with a
refinement that a finite set of macroscopic variables (versus only energy) is
used, and are able to show the existence of a dynamic threshold
starting with some value of K -- the number of variables in
each clause. Beyond dynamic threshold, the algorithm should take exponentially
long time to find a solution. We compare the results for extended and
simplified sets of landscapes and provide numerical evidence in support of our
universality ansatz. We have been able to map the ensemble of random graphs
onto another ensemble with fluctuations significantly reduced. This enabled us
to obtain tight upper bounds on satisfiability transition and to recompute the
dynamical transition using the extended set of landscapes.Comment: 41 pages, 10 figures; added a paragraph on paper's organization to
the introduction, fixed reference
Taming a non-convex landscape with dynamical long-range order: memcomputing Ising benchmarks
Recent work on quantum annealing has emphasized the role of collective
behavior in solving optimization problems. By enabling transitions of clusters
of variables, such solvers are able to navigate their state space and locate
solutions more efficiently despite having only local connections between
elements. However, collective behavior is not exclusive to quantum annealers,
and classical solvers that display collective dynamics should also possess an
advantage in navigating a non-convex landscape. Here, we give evidence that a
benchmark derived from quantum annealing studies is solvable in polynomial time
using digital memcomputing machines, which utilize a collection of dynamical
components with memory to represent the structure of the underlying
optimization problem. To illustrate the role of memory and clarify the
structure of these solvers we propose a simple model of these machines that
demonstrates the emergence of long-range order. This model, when applied to
finding the ground state of the Ising frustrated-loop benchmarks, undergoes a
transient phase of avalanches which can span the entire lattice and
demonstrates a connection between long-range behavior and their probability of
success. These results establish the advantages of computational approaches
based on collective dynamics of continuous dynamical systems
ASlib: A Benchmark Library for Algorithm Selection
The task of algorithm selection involves choosing an algorithm from a set of
algorithms on a per-instance basis in order to exploit the varying performance
of algorithms over a set of instances. The algorithm selection problem is
attracting increasing attention from researchers and practitioners in AI. Years
of fruitful applications in a number of domains have resulted in a large amount
of data, but the community lacks a standard format or repository for this data.
This situation makes it difficult to share and compare different approaches
effectively, as is done in other, more established fields. It also
unnecessarily hinders new researchers who want to work in this area. To address
this problem, we introduce a standardized format for representing algorithm
selection scenarios and a repository that contains a growing number of data
sets from the literature. Our format has been designed to be able to express a
wide variety of different scenarios. Demonstrating the breadth and power of our
platform, we describe a set of example experiments that build and evaluate
algorithm selection models through a common interface. The results display the
potential of algorithm selection to achieve significant performance
improvements across a broad range of problems and algorithms.Comment: Accepted to be published in Artificial Intelligence Journa
Problem Understanding through Landscape Theory
In order to understand the structure of a problem we need to measure some features of the problem. Some examples of measures suggested in the past are autocorrelation and fitness-distance correlation. Landscape theory, developed in the last years in the field of combinatorial optimization, provides mathematical expressions to efficiently compute statistics on optimization problems. In this paper we discuss how can we use optimizaciĂłn combinatoria in the context of problem understanding and present two software tools that can be used to efficiently compute the mentioned measures.Ministerio de EconomĂa y Competitividad (TIN2011-28194
Random subcubes as a toy model for constraint satisfaction problems
We present an exactly solvable random-subcube model inspired by the structure
of hard constraint satisfaction and optimization problems. Our model reproduces
the structure of the solution space of the random k-satisfiability and
k-coloring problems, and undergoes the same phase transitions as these
problems. The comparison becomes quantitative in the large-k limit. Distance
properties, as well the x-satisfiability threshold, are studied. The model is
also generalized to define a continuous energy landscape useful for studying
several aspects of glassy dynamics.Comment: 21 pages, 4 figure
- …