636,330 research outputs found
Complexity, parallel computation and statistical physics
The intuition that a long history is required for the emergence of complexity
in natural systems is formalized using the notion of depth. The depth of a
system is defined in terms of the number of parallel computational steps needed
to simulate it. Depth provides an objective, irreducible measure of history
applicable to systems of the kind studied in statistical physics. It is argued
that physical complexity cannot occur in the absence of substantial depth and
that depth is a useful proxy for physical complexity. The ideas are illustrated
for a variety of systems in statistical physics.Comment: 21 pages, 7 figure
Particle Computation: Complexity, Algorithms, and Logic
We investigate algorithmic control of a large swarm of mobile particles (such
as robots, sensors, or building material) that move in a 2D workspace using a
global input signal (such as gravity or a magnetic field). We show that a maze
of obstacles to the environment can be used to create complex systems. We
provide a wide range of results for a wide range of questions. These can be
subdivided into external algorithmic problems, in which particle configurations
serve as input for computations that are performed elsewhere, and internal
logic problems, in which the particle configurations themselves are used for
carrying out computations. For external algorithms, we give both negative and
positive results. If we are given a set of stationary obstacles, we prove that
it is NP-hard to decide whether a given initial configuration of unit-sized
particles can be transformed into a desired target configuration. Moreover, we
show that finding a control sequence of minimum length is PSPACE-complete. We
also work on the inverse problem, providing constructive algorithms to design
workspaces that efficiently implement arbitrary permutations between different
configurations. For internal logic, we investigate how arbitrary computations
can be implemented. We demonstrate how to encode dual-rail logic to build a
universal logic gate that concurrently evaluates and, nand, nor, and or
operations. Using many of these gates and appropriate interconnects, we can
evaluate any logical expression. However, we establish that simulating the full
range of complex interactions present in arbitrary digital circuits encounters
a fundamental difficulty: a fan-out gate cannot be generated. We resolve this
missing component with the help of 2x1 particles, which can create fan-out
gates that produce multiple copies of the inputs. Using these gates we provide
rules for replicating arbitrary digital circuits.Comment: 27 pages, 19 figures, full version that combines three previous
conference article
Universality of Entanglement and Quantum Computation Complexity
We study the universality of scaling of entanglement in Shor's factoring
algorithm and in adiabatic quantum algorithms across a quantum phase transition
for both the NP-complete Exact Cover problem as well as the Grover's problem.
The analytic result for Shor's algorithm shows a linear scaling of the entropy
in terms of the number of qubits, therefore difficulting the possibility of an
efficient classical simulation protocol. A similar result is obtained
numerically for the quantum adiabatic evolution Exact Cover algorithm, which
also shows universality of the quantum phase transition the system evolves
nearby. On the other hand, entanglement in Grover's adiabatic algorithm remains
a bounded quantity even at the critical point. A classification of scaling of
entanglement appears as a natural grading of the computational complexity of
simulating quantum phase transitions.Comment: 30 pages, 17 figures, accepted for publication in PR
Sensor networks and distributed CSP: communication, computation and complexity
We introduce SensorDCSP, a naturally distributed benchmark based on a real-world application that arises in the context of networked distributed systems. In order to study the performance of Distributed CSP (DisCSP) algorithms in a truly distributed setting, we use a discrete-event network simulator, which allows us to model the impact of different network traffic conditions on the performance of the algorithms. We consider two complete DisCSP algorithms: asynchronous backtracking (ABT) and asynchronous weak commitment search (AWC), and perform performance comparison for these algorithms on both satisfiable and unsatisfiable instances of SensorDCSP. We found that random delays (due to network traffic or in some cases actively introduced by the agents) combined with a dynamic decentralized restart strategy can improve the performance of DisCSP algorithms. In addition, we introduce GSensorDCSP, a plain-embedded version of SensorDCSP that is closely related to various real-life dynamic tracking systems. We perform both analytical and empirical study of this benchmark domain. In particular, this benchmark allows us to study the attractiveness of solution repairing for solving a sequence of DisCSPs that represent the dynamic tracking of a set of moving objects.This work was supported in part by AFOSR (F49620-01-1-0076, Intelligent Information Systems Institute and MURI F49620-01-1-0361), CICYT (TIC2001-1577-C03-03 and TIC2003-00950), DARPA (F30602-00-2- 0530), an NSF CAREER award (IIS-9734128), and an Alfred P. Sloan Research Fellowship. The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily representing the official policies or endorsements, either expressed or implied, of the US Government
- …