1,037 research outputs found
Wildcard dimensions, coding theory and fault-tolerant meshes and hypercubes
Hypercubes, meshes and tori are well known interconnection networks for parallel computers. The sets of edges in those graphs can be partitioned to dimensions. It is well known that the hypercube can be extended by adding a wildcard dimension resulting in a folded hypercube that has better fault-tolerant and communication capabilities. First we prove that the folded hypercube is optimal in the sense that only a single wildcard dimension can be added to the hypercube. We then investigate the idea of adding wildcard dimensions to d-dimensional meshes and tori. Using techniques from error correcting codes we construct d-dimensional meshes and tori with wildcard dimensions. Finally, we show how these constructions can be used to tolerate edge and node faults in mesh and torus networks
Quasirandom Rumor Spreading: An Experimental Analysis
We empirically analyze two versions of the well-known "randomized rumor
spreading" protocol to disseminate a piece of information in networks. In the
classical model, in each round each informed node informs a random neighbor. In
the recently proposed quasirandom variant, each node has a (cyclic) list of its
neighbors. Once informed, it starts at a random position of the list, but from
then on informs its neighbors in the order of the list. While for sparse random
graphs a better performance of the quasirandom model could be proven, all other
results show that, independent of the structure of the lists, the same
asymptotic performance guarantees hold as for the classical model. In this
work, we compare the two models experimentally. This not only shows that the
quasirandom model generally is faster, but also that the runtime is more
concentrated around the mean. This is surprising given that much fewer random
bits are used in the quasirandom process. These advantages are also observed in
a lossy communication model, where each transmission does not reach its target
with a certain probability, and in an asynchronous model, where nodes send at
random times drawn from an exponential distribution. We also show that
typically the particular structure of the lists has little influence on the
efficiency.Comment: 14 pages, appeared in ALENEX'0
Optimal Networks from Error Correcting Codes
To address growth challenges facing large Data Centers and supercomputing
clusters a new construction is presented for scalable, high throughput, low
latency networks. The resulting networks require 1.5-5 times fewer switches,
2-6 times fewer cables, have 1.2-2 times lower latency and correspondingly
lower congestion and packet losses than the best present or proposed networks
providing the same number of ports at the same total bisection. These advantage
ratios increase with network size. The key new ingredient is the exact
equivalence discovered between the problem of maximizing network bisection for
large classes of practically interesting Cayley graphs and the problem of
maximizing codeword distance for linear error correcting codes. Resulting
translation recipe converts existent optimal error correcting codes into
optimal throughput networks.Comment: 14 pages, accepted at ANCS 2013 conferenc
Quasirandom Load Balancing
We propose a simple distributed algorithm for balancing indivisible tokens on
graphs. The algorithm is completely deterministic, though it tries to imitate
(and enhance) a random algorithm by keeping the accumulated rounding errors as
small as possible.
Our new algorithm surprisingly closely approximates the idealized process
(where the tokens are divisible) on important network topologies. On
d-dimensional torus graphs with n nodes it deviates from the idealized process
only by an additive constant. In contrast to that, the randomized rounding
approach of Friedrich and Sauerwald (2009) can deviate up to Omega(polylog(n))
and the deterministic algorithm of Rabani, Sinclair and Wanka (1998) has a
deviation of Omega(n^{1/d}). This makes our quasirandom algorithm the first
known algorithm for this setting which is optimal both in time and achieved
smoothness. We further show that also on the hypercube our algorithm has a
smaller deviation from the idealized process than the previous algorithms.Comment: 25 page
- …