4,053 research outputs found
A New Characterization of Tree Medians with Applications to Distributed Algorithms
A new characterization of tree medians is presented: we show that a vertex m is a median of a tree T with n vertices iff there exists a partition of the vertex set into [n/2] disjoint pairs (excluding m when n is odd), such that all the paths connecting the two vertices in any of the pairs pass through m. We show that in this case this sum is the largest possible among all such partitions, and we use this fact to discuss lower bounds on the message complexity of the distributed sorting problem. This lower bound implies that, given a network of a tree topology, choosing a median and then route all the information through it is the best possible strategy, in terms of worst-case number of messages sent during any execution of any distributed sorting algorithm. We also discuss the implications for networks of a general topology and for the distributed ranking problem
Generalized centrality in trees
In 1982, Slater defined path subgraph analogues to the center, median, and (branch or branchweight) centroid of a tree. We define three families of central substructures of trees, including three types of central subtrees of degree at most D that yield the center, median, and centroid for D = 0 and Slater's path analogues for D = 2. We generalize these results concerning paths and include proofs that each type of generalized center and generalized centroid is unique. We also present algorithms for finding one or all generalized central substructures of each type.
Observer Placement for Source Localization: The Effect of Budgets and Transmission Variance
When an epidemic spreads in a network, a key question is where was its
source, i.e., the node that started the epidemic. If we know the time at which
various nodes were infected, we can attempt to use this information in order to
identify the source. However, maintaining observer nodes that can provide their
infection time may be costly, and we may have a budget on the number of
observer nodes we can maintain. Moreover, some nodes are more informative than
others due to their location in the network. Hence, a pertinent question
arises: Which nodes should we select as observers in order to maximize the
probability that we can accurately identify the source? Inspired by the simple
setting in which the node-to-node delays in the transmission of the epidemic
are deterministic, we develop a principled approach for addressing the problem
even when transmission delays are random. We show that the optimal
observer-placement differs depending on the variance of the transmission delays
and propose approaches in both low- and high-variance settings. We validate our
methods by comparing them against state-of-the-art observer-placements and show
that, in both settings, our approach identifies the source with higher
accuracy.Comment: Accepted for presentation at the 54th Annual Allerton Conference on
Communication, Control, and Computin
Implications of Selfish Neighbor Selection in Overlay Networks
In a typical overlay network for routing or content sharing, each node must select a fixed number of immediate overlay neighbors for routing traffic or content queries. A selfish node entering such a network would select neighbors so as to minimize the weighted sum of expected access costs to all its destinations. Previous work on selfish neighbor selection has built intuition with simple models where edges are undirected, access costs are modeled by hop-counts, and nodes have potentially unbounded degrees. However, in practice, important constraints not captured by these models lead to richer games with substantively and fundamentally different outcomes. Our work models neighbor selection as a game involving directed links, constraints on the number of allowed neighbors, and costs reflecting both network latency and node preference. We express a node's "best response" wiring strategy as a k-median problem on asymmetric distance, and use this formulation to obtain pure Nash equilibria. We experimentally examine the properties of such stable wirings on synthetic topologies, as well as on real topologies and maps constructed from PlanetLab and AS-level Internet measurements. Our results indicate that selfish nodes can reap substantial performance benefits when connecting to overlay networks composed of non-selfish nodes. On the other hand, in overlays that are dominated by selfish nodes, the resulting stable wirings are optimized to such great extent that even non-selfish newcomers can extract near-optimal performance through naive wiring strategies.Marie Curie Outgoing International Fellowship of the EU (MOIF-CT-2005-007230); National Science Foundation (CNS Cybertrust 0524477, CNS NeTS 0520166, CNS ITR 0205294, EIA RI 020206
Spectra: Robust Estimation of Distribution Functions in Networks
Distributed aggregation allows the derivation of a given global aggregate
property from many individual local values in nodes of an interconnected
network system. Simple aggregates such as minima/maxima, counts, sums and
averages have been thoroughly studied in the past and are important tools for
distributed algorithms and network coordination. Nonetheless, this kind of
aggregates may not be comprehensive enough to characterize biased data
distributions or when in presence of outliers, making the case for richer
estimates of the values on the network. This work presents Spectra, a
distributed algorithm for the estimation of distribution functions over large
scale networks. The estimate is available at all nodes and the technique
depicts important properties, namely: robust when exposed to high levels of
message loss, fast convergence speed and fine precision in the estimate. It can
also dynamically cope with changes of the sampled local property, not requiring
algorithm restarts, and is highly resilient to node churn. The proposed
approach is experimentally evaluated and contrasted to a competing state of the
art distribution aggregation technique.Comment: Full version of the paper published at 12th IFIP International
Conference on Distributed Applications and Interoperable Systems (DAIS),
Stockholm (Sweden), June 201
Validation of Dunbar's number in Twitter conversations
Modern society's increasing dependency on online tools for both work and
recreation opens up unique opportunities for the study of social interactions.
A large survey of online exchanges or conversations on Twitter, collected
across six months involving 1.7 million individuals is presented here. We test
the theoretical cognitive limit on the number of stable social relationships
known as Dunbar's number. We find that users can entertain a maximum of 100-200
stable relationships in support for Dunbar's prediction. The "economy of
attention" is limited in the online world by cognitive and biological
constraints as predicted by Dunbar's theory. Inspired by this empirical
evidence we propose a simple dynamical mechanism, based on finite priority
queuing and time resources, that reproduces the observed social behavior.Comment: 8 pages, 6 figure
Exact Computation of a Manifold Metric, via Lipschitz Embeddings and Shortest Paths on a Graph
Data-sensitive metrics adapt distances locally based the density of data
points with the goal of aligning distances and some notion of similarity. In
this paper, we give the first exact algorithm for computing a data-sensitive
metric called the nearest neighbor metric. In fact, we prove the surprising
result that a previously published -approximation is an exact algorithm.
The nearest neighbor metric can be viewed as a special case of a
density-based distance used in machine learning, or it can be seen as an
example of a manifold metric. Previous computational research on such metrics
despaired of computing exact distances on account of the apparent difficulty of
minimizing over all continuous paths between a pair of points. We leverage the
exact computation of the nearest neighbor metric to compute sparse spanners and
persistent homology. We also explore the behavior of the metric built from
point sets drawn from an underlying distribution and consider the more general
case of inputs that are finite collections of path-connected compact sets.
The main results connect several classical theories such as the conformal
change of Riemannian metrics, the theory of positive definite functions of
Schoenberg, and screw function theory of Schoenberg and Von Neumann. We develop
novel proof techniques based on the combination of screw functions and
Lipschitz extensions that may be of independent interest.Comment: 15 page
- …