76 research outputs found
Communication Efficiency in Self-stabilizing Silent Protocols
Self-stabilization is a general paradigm to provide forward recovery
capabilities to distributed systems and networks. Intuitively, a protocol is
self-stabilizing if it is able to recover without external intervention from
any catastrophic transient failure. In this paper, our focus is to lower the
communication complexity of self-stabilizing protocols \emph{below} the need of
checking every neighbor forever. In more details, the contribution of the paper
is threefold: (i) We provide new complexity measures for communication
efficiency of self-stabilizing protocols, especially in the stabilized phase or
when there are no faults, (ii) On the negative side, we show that for
non-trivial problems such as coloring, maximal matching, and maximal
independent set, it is impossible to get (deterministic or probabilistic)
self-stabilizing solutions where every participant communicates with less than
every neighbor in the stabilized phase, and (iii) On the positive side, we
present protocols for coloring, maximal matching, and maximal independent set
such that a fraction of the participants communicates with exactly one neighbor
in the stabilized phase
Weak vs. Self vs. Probabilistic Stabilization
Self-stabilization is a strong property that guarantees that a network always
resume correct behavior starting from an arbitrary initial state. Weaker
guarantees have later been introduced to cope with impossibility results:
probabilistic stabilization only gives probabilistic convergence to a correct
behavior. Also, weak stabilization only gives the possibility of convergence.
In this paper, we investigate the relative power of weak, self, and
probabilistic stabilization, with respect to the set of problems that can be
solved. We formally prove that in that sense, weak stabilization is strictly
stronger that self-stabilization. Also, we refine previous results on weak
stabilization to prove that, for practical schedule instances, a deterministic
weak-stabilizing protocol can be turned into a probabilistic self-stabilizing
one. This latter result hints at more practical use of weak-stabilization, as
such algorthms are easier to design and prove than their (probabilistic)
self-stabilizing counterparts
Randomization Adaptive Self-Stabilization
We present a scheme to convert self-stabilizing algorithms that use
randomization during and following convergence to self-stabilizing algorithms
that use randomization only during convergence. We thus reduce the number of
random bits from an infinite number to a bounded number. The scheme is
applicable to the cases in which there exits a local predicate for each node,
such that global consistency is implied by the union of the local predicates.
We demonstrate our scheme over the token circulation algorithm of Herman and
the recent constant time Byzantine self-stabilizing clock synchronization
algorithm by Ben-Or, Dolev and Hoch. The application of our scheme results in
the first constant time Byzantine self-stabilizing clock synchronization
algorithm that uses a bounded number of random bits
Minimizing Message Size in Stochastic Communication Patterns: Fast Self-Stabilizing Protocols with 3 bits
This paper considers the basic model of communication, in
which in each round, each agent extracts information from few randomly chosen
agents. We seek to identify the smallest amount of information revealed in each
interaction (message size) that nevertheless allows for efficient and robust
computations of fundamental information dissemination tasks. We focus on the
Majority Bit Dissemination problem that considers a population of agents,
with a designated subset of source agents. Each source agent holds an input bit
and each agent holds an output bit. The goal is to let all agents converge
their output bits on the most frequent input bit of the sources (the majority
bit). Note that the particular case of a single source agent corresponds to the
classical problem of Broadcast. We concentrate on the severe fault-tolerant
context of self-stabilization, in which a correct configuration must be reached
eventually, despite all agents starting the execution with arbitrary initial
states.
We first design a general compiler which can essentially transform any
self-stabilizing algorithm with a certain property that uses -bits
messages to one that uses only -bits messages, while paying only a
small penalty in the running time. By applying this compiler recursively we
then obtain a self-stabilizing Clock Synchronization protocol, in which agents
synchronize their clocks modulo some given integer , within rounds w.h.p., and using messages that contain bits only.
We then employ the new Clock Synchronization tool to obtain a
self-stabilizing Majority Bit Dissemination protocol which converges in time, w.h.p., on every initial configuration, provided that the
ratio of sources supporting the minority opinion is bounded away from half.
Moreover, this protocol also uses only 3 bits per interaction.Comment: 28 pages, 4 figure
Leader Election in Anonymous Rings: Franklin Goes Probabilistic
We present a probabilistic leader election algorithm for anonymous, bidirectional, asynchronous rings. It is based on an algorithm from Franklin, augmented with random identity selection, hop counters to detect identity clashes, and round numbers modulo 2. As a result, the algorithm is finite-state, so that various model checking techniques can be employed to verify its correctness, that is, eventually a unique leader is elected with probability one. We also sketch a formal correctness proof of the algorithm for rings with arbitrary size
Generalized solution for the Herman Protocol Conjecture
We have a cycle of nodes and there is a token on an odd number of nodes.
At each step, each token independently moves to its clockwise neighbor or stays
at its position with probability . If two tokens arrive to the
same node, then we remove both of them. The process ends when only one token
remains. The question is that for a fixed , which is the initial
configuration that maximizes the expected number of steps . The Herman
Protocol Conjecture says that the -token configuration with distances
and maximizes . We
present a proof of this conjecture not only for but also for
for some function
which method applies for different generalizations of the problem
Self-Stabilizing Repeated Balls-into-Bins
We study the following synchronous process that we call "repeated
balls-into-bins". The process is started by assigning balls to bins in
an arbitrary way. In every subsequent round, from each non-empty bin one ball
is chosen according to some fixed strategy (random, FIFO, etc), and re-assigned
to one of the bins uniformly at random.
We define a configuration "legitimate" if its maximum load is
. We prove that, starting from any configuration, the
process will converge to a legitimate configuration in linear time and then it
will only take on legitimate configurations over a period of length bounded by
any polynomial in , with high probability (w.h.p.). This implies that the
process is self-stabilizing and that every ball traverses all bins in
rounds, w.h.p
Bounded Model Checking for Probabilistic Programs
In this paper we investigate the applicability of standard model checking
approaches to verifying properties in probabilistic programming. As the
operational model for a standard probabilistic program is a potentially
infinite parametric Markov decision process, no direct adaption of existing
techniques is possible. Therefore, we propose an on-the-fly approach where the
operational model is successively created and verified via a step-wise
execution of the program. This approach enables to take key features of many
probabilistic programs into account: nondeterminism and conditioning. We
discuss the restrictions and demonstrate the scalability on several benchmarks
- âŠ