31,191 research outputs found
Fast approximation of centrality and distances in hyperbolic graphs
We show that the eccentricities (and thus the centrality indices) of all
vertices of a -hyperbolic graph can be computed in linear
time with an additive one-sided error of at most , i.e., after a
linear time preprocessing, for every vertex of one can compute in
time an estimate of its eccentricity such that
for a small constant . We
prove that every -hyperbolic graph has a shortest path tree,
constructible in linear time, such that for every vertex of ,
. These results are based on an
interesting monotonicity property of the eccentricity function of hyperbolic
graphs: the closer a vertex is to the center of , the smaller its
eccentricity is. We also show that the distance matrix of with an additive
one-sided error of at most can be computed in
time, where is a small constant. Recent empirical studies show that
many real-world graphs (including Internet application networks, web networks,
collaboration networks, social networks, biological networks, and others) have
small hyperbolicity. So, we analyze the performance of our algorithms for
approximating centrality and distance matrix on a number of real-world
networks. Our experimental results show that the obtained estimates are even
better than the theoretical bounds.Comment: arXiv admin note: text overlap with arXiv:1506.01799 by other author
Approximating the Diameter of Planar Graphs in Near Linear Time
We present a -approximation algorithm running in
time for finding the diameter of an undirected
planar graph with non-negative edge lengths
Far-Field Compression for Fast Kernel Summation Methods in High Dimensions
We consider fast kernel summations in high dimensions: given a large set of
points in dimensions (with ) and a pair-potential function (the
{\em kernel} function), we compute a weighted sum of all pairwise kernel
interactions for each point in the set. Direct summation is equivalent to a
(dense) matrix-vector multiplication and scales quadratically with the number
of points. Fast kernel summation algorithms reduce this cost to log-linear or
linear complexity.
Treecodes and Fast Multipole Methods (FMMs) deliver tremendous speedups by
constructing approximate representations of interactions of points that are far
from each other. In algebraic terms, these representations correspond to
low-rank approximations of blocks of the overall interaction matrix. Existing
approaches require an excessive number of kernel evaluations with increasing
and number of points in the dataset.
To address this issue, we use a randomized algebraic approach in which we
first sample the rows of a block and then construct its approximate, low-rank
interpolative decomposition. We examine the feasibility of this approach
theoretically and experimentally. We provide a new theoretical result showing a
tighter bound on the reconstruction error from uniformly sampling rows than the
existing state-of-the-art. We demonstrate that our sampling approach is
competitive with existing (but prohibitively expensive) methods from the
literature. We also construct kernel matrices for the Laplacian, Gaussian, and
polynomial kernels -- all commonly used in physics and data analysis. We
explore the numerical properties of blocks of these matrices, and show that
they are amenable to our approach. Depending on the data set, our randomized
algorithm can successfully compute low rank approximations in high dimensions.
We report results for data sets with ambient dimensions from four to 1,000.Comment: 43 pages, 21 figure
Distributed Approximation Algorithms for Weighted Shortest Paths
A distributed network is modeled by a graph having nodes (processors) and
diameter . We study the time complexity of approximating {\em weighted}
(undirected) shortest paths on distributed networks with a {\em
bandwidth restriction} on edges (the standard synchronous \congest model). The
question whether approximation algorithms help speed up the shortest paths
(more precisely distance computation) was raised since at least 2004 by Elkin
(SIGACT News 2004). The unweighted case of this problem is well-understood
while its weighted counterpart is fundamental problem in the area of
distributed approximation algorithms and remains widely open. We present new
algorithms for computing both single-source shortest paths (\sssp) and
all-pairs shortest paths (\apsp) in the weighted case.
Our main result is an algorithm for \sssp. Previous results are the classic
-time Bellman-Ford algorithm and an -time
-approximation algorithm, for any integer
, which follows from the result of Lenzen and Patt-Shamir (STOC 2013).
(Note that Lenzen and Patt-Shamir in fact solve a harder problem, and we use
to hide the O(\poly\log n) term.) We present an -time -approximation algorithm for \sssp. This
algorithm is {\em sublinear-time} as long as is sublinear, thus yielding a
sublinear-time algorithm with almost optimal solution. When is small, our
running time matches the lower bound of by Das Sarma
et al. (SICOMP 2012), which holds even when , up to a
\poly\log n factor.Comment: Full version of STOC 201
Approximate Hamming distance in a stream
We consider the problem of computing a -approximation of the
Hamming distance between a pattern of length and successive substrings of a
stream. We first look at the one-way randomised communication complexity of
this problem, giving Alice the first half of the stream and Bob the second
half. We show the following: (1) If Alice and Bob both share the pattern then
there is an bit randomised one-way communication
protocol. (2) If only Alice has the pattern then there is an
bit randomised one-way communication protocol.
We then go on to develop small space streaming algorithms for
-approximate Hamming distance which give worst case running time
guarantees per arriving symbol. (1) For binary input alphabets there is an
space and
time streaming -approximate Hamming distance algorithm. (2) For
general input alphabets there is an
space and time streaming
-approximate Hamming distance algorithm.Comment: Submitted to ICALP' 201
- …