371 research outputs found
Deterministic Computations Whose History Is Independent of the Order of Asynchronous Updating
Consider a network of processors (sites) in which each site x has a finite set N(x) of neighbors. There is a transition function f that for each site x computes the next state ξ(x) from the states in N(x). But these transitions (updates) are applied in arbitrary order, one or many at a time. If the state of site x at time t is η(x; t) then let us define the sequence ζ(x; 0); ζ(x; 1), ... by taking the sequence η(x; 0),η(x; 1), ... , and deleting each repetition, i.e. each element equal to the preceding one. The function f is said to have invariant histories if the sequence ζ(x; i), (while it lasts, in case it is finite) depends only on the initial configuration, not on the order of updates.
This paper shows that though the invariant history property is typically undecidable, there is a useful simple sufficient condition, called commutativity: For any configuration, for any pair x; y of neighbors, if the updating would change both ξ(x) and ξ(y) then the result of updating first x and then y is the same as the result of doing this in the reverse order. This fact is derivable from known results on the confluence of term-rewriting systems but the self-contained proof given here may be justifiable.National Science Foundation (CCR-920484
Clairvoyant embedding in one dimension
Let v, w be infinite 0-1 sequences, and m a positive integer. We say that w
is m-embeddable in v, if there exists an increasing sequence n_{i} of integers
with n_{0}=0, such that 0 0.
Let X and Y be independent coin-tossing sequences. We will show that there is
an m with the property that Y is m-embeddable into X with positive probability.
This answers a question that was open for a while. The proof generalizes
somewhat the hierarchical method of an earlier paper of the author on dependent
percolation.Comment: 49 pages. Some errors corrected. arXiv admin note: substantial text
overlap with arXiv:math/010915
Deterministic computations whose history is independent of the order of asynchronous updating
Consider a network of processors (sites) in which each site x has a finite
set N(x) of neighbors. There is a transition function f that for each site x
computes the next state \xi(x) from the states in N(x). But these transitions
(updates) are applied in arbitrary order, one or many at a time. If the state
of site x at time t is \eta(x,t) then let us define the sequence \zeta(x,0),
\zeta(x,1), ... by taking the sequence \eta(x,0), \eta(x,1), ..., and deleting
repetitions. The function f is said to have invariant histories if the sequence
\zeta(x,i), (while it lasts, in case it is finite) depends only on the initial
configuration, not on the order of updates.
This paper shows that though the invariant history property is typically
undecidable, there is a useful simple sufficient condition, called
commutativity: For any configuration, for any pair x,y of neighbors, if the
updating would change both \xi(x) and \xi(y) then the result of updating first
x and then y is the same as the result of doing this in the reverse order
The angel wins
The angel-devil game is played on an infinite two-dimensional ``chessboard''.
The squares of the board are all white at the beginning. The players called
angel and devil take turns in their steps. When it is the devil's turn, he can
turn a square black. The angel always stays on a white square, and when it is
her turn she can fly at a distance of at most J steps (each of which can be
horizontal, vertical or diagonal) to a new white square. Here J is a constant.
The devil wins if the angel does not find any more white squares to land on.
The result of the paper is that if J is sufficiently large then the angel has a
strategy such that the devil will never capture her. This deceptively
easy-sounding result has been a conjecture, surprisingly, for about thirty
years. Several other independent solutions have appeared simultaneously, some
of them prove that J=2 is sufficient (see the Wikipedia on the angel problem).
Still, it is hoped that the hierarchical solution presented here may prove
useful for some generalizations.Comment: 28 pages, 8 figure
Compatible sequences and a slow Winkler percolation
Two infinite 0-1 sequences are called compatible when it is possible to cast
out 0's from both in such a way that they become complementary to each other.
Answering a question of Peter Winkler, we show that if the two 0-1-sequences
are random i.i.d. and independent from each other, with probability p of 1's,
then if p is sufficiently small they are compatible with positive probability.
The question is equivalent to a certain dependent percolation with a power-law
behavior: the probability that the origin is blocked at distance n but not
closer decreases only polynomially fast and not, as usual, exponentially.Comment: 33 pages, 8 figures. Submitted to Combinatorics, Probability and
Computing. Some errors correcte
Uniform test of algorithmic randomness over a general space
The algorithmic theory of randomness is well developed when the underlying
space is the set of finite or infinite sequences and the underlying probability
distribution is the uniform distribution or a computable distribution. These
restrictions seem artificial. Some progress has been made to extend the theory
to arbitrary Bernoulli distributions (by Martin-Loef), and to arbitrary
distributions (by Levin). We recall the main ideas and problems of Levin's
theory, and report further progress in the same framework.
- We allow non-compact spaces (like the space of continuous functions,
underlying the Brownian motion).
- The uniform test (deficiency of randomness) d_P(x) (depending both on the
outcome x and the measure P should be defined in a general and natural way.
- We see which of the old results survive: existence of universal tests,
conservation of randomness, expression of tests in terms of description
complexity, existence of a universal measure, expression of mutual information
as "deficiency of independence.
- The negative of the new randomness test is shown to be a generalization of
complexity in continuous spaces; we show that the addition theorem survives.
The paper's main contribution is introducing an appropriate framework for
studying these questions and related ones (like statistics for a general family
of distributions).Comment: 40 pages. Journal reference and a slight correction in the proof of
Theorem 7 adde
The clairvoyant demon has a hard task
Consider the integer lattice L = ℤ2. For some m [ges ] 4, let us colour each column of this lattice independently and uniformly with one of m colours. We do the same for the rows, independently of the columns. A point of L will be called blocked if its row and column have the same colour. We say that this random configuration percolates if there is a path in L starting at the origin, consisting of rightward and upward unit steps, avoiding the blocked points. As a problem arising in distributed computing, it has been conjectured that for m [ges ] 4 the configuration percolates with positive probability. This question remains open, but we prove that the probability that there is percolation to distance n but not to infinity is not exponentially small in n. This narrows the range of methods available for proving the conjecture
Algorithmic Statistics
While Kolmogorov complexity is the accepted absolute measure of information
content of an individual finite object, a similarly absolute notion is needed
for the relation between an individual data sample and an individual model
summarizing the information in the data, for example, a finite set (or
probability distribution) where the data sample typically came from. The
statistical theory based on such relations between individual objects can be
called algorithmic statistics, in contrast to classical statistical theory that
deals with relations between probabilistic ensembles. We develop the
algorithmic theory of statistic, sufficient statistic, and minimal sufficient
statistic. This theory is based on two-part codes consisting of the code for
the statistic (the model summarizing the regularity, the meaningful
information, in the data) and the model-to-data code. In contrast to the
situation in probabilistic statistical theory, the algorithmic relation of
(minimal) sufficiency is an absolute relation between the individual model and
the individual data sample. We distinguish implicit and explicit descriptions
of the models. We give characterizations of algorithmic (Kolmogorov) minimal
sufficient statistic for all data samples for both description modes--in the
explicit mode under some constraints. We also strengthen and elaborate earlier
results on the ``Kolmogorov structure function'' and ``absolutely
non-stochastic objects''--those rare objects for which the simplest models that
summarize their relevant information (minimal sufficient statistics) are at
least as complex as the objects themselves. We demonstrate a close relation
between the probabilistic notions and the algorithmic ones.Comment: LaTeX, 22 pages, 1 figure, with correction to the published journal
versio
- …