15,219 research outputs found
Sketch-based Influence Maximization and Computation: Scaling up with Guarantees
Propagation of contagion through networks is a fundamental process. It is
used to model the spread of information, influence, or a viral infection.
Diffusion patterns can be specified by a probabilistic model, such as
Independent Cascade (IC), or captured by a set of representative traces.
Basic computational problems in the study of diffusion are influence queries
(determining the potency of a specified seed set of nodes) and Influence
Maximization (identifying the most influential seed set of a given size).
Answering each influence query involves many edge traversals, and does not
scale when there are many queries on very large graphs. The gold standard for
Influence Maximization is the greedy algorithm, which iteratively adds to the
seed set a node maximizing the marginal gain in influence. Greedy has a
guaranteed approximation ratio of at least (1-1/e) and actually produces a
sequence of nodes, with each prefix having approximation guarantee with respect
to the same-size optimum. Since Greedy does not scale well beyond a few million
edges, for larger inputs one must currently use either heuristics or
alternative algorithms designed for a pre-specified small seed set size.
We develop a novel sketch-based design for influence computation. Our greedy
Sketch-based Influence Maximization (SKIM) algorithm scales to graphs with
billions of edges, with one to two orders of magnitude speedup over the best
greedy methods. It still has a guaranteed approximation ratio, and in practice
its quality nearly matches that of exact greedy. We also present influence
oracles, which use linear-time preprocessing to generate a small sketch for
each node, allowing the influence of any seed set to be quickly answered from
the sketches of its nodes.Comment: 10 pages, 5 figures. Appeared at the 23rd Conference on Information
and Knowledge Management (CIKM 2014) in Shanghai, Chin
On the relationship between continuous- and discrete-time quantum walk
Quantum walk is one of the main tools for quantum algorithms. Defined by
analogy to classical random walk, a quantum walk is a time-homogeneous quantum
process on a graph. Both random and quantum walks can be defined either in
continuous or discrete time. But whereas a continuous-time random walk can be
obtained as the limit of a sequence of discrete-time random walks, the two
types of quantum walk appear fundamentally different, owing to the need for
extra degrees of freedom in the discrete-time case.
In this article, I describe a precise correspondence between continuous- and
discrete-time quantum walks on arbitrary graphs. Using this correspondence, I
show that continuous-time quantum walk can be obtained as an appropriate limit
of discrete-time quantum walks. The correspondence also leads to a new
technique for simulating Hamiltonian dynamics, giving efficient simulations
even in cases where the Hamiltonian is not sparse. The complexity of the
simulation is linear in the total evolution time, an improvement over
simulations based on high-order approximations of the Lie product formula. As
applications, I describe a continuous-time quantum walk algorithm for element
distinctness and show how to optimally simulate continuous-time query
algorithms of a certain form in the conventional quantum query model. Finally,
I discuss limitations of the method for simulating Hamiltonians with negative
matrix elements, and present two problems that motivate attempting to
circumvent these limitations.Comment: 22 pages. v2: improved presentation, new section on Hamiltonian
oracles; v3: published version, with improved analysis of phase estimatio
Span programs and quantum algorithms for st-connectivity and claw detection
We introduce a span program that decides st-connectivity, and generalize the
span program to develop quantum algorithms for several graph problems. First,
we give an algorithm for st-connectivity that uses O(n d^{1/2}) quantum queries
to the n x n adjacency matrix to decide if vertices s and t are connected,
under the promise that they either are connected by a path of length at most d,
or are disconnected. We also show that if T is a path, a star with two
subdivided legs, or a subdivision of a claw, its presence as a subgraph in the
input graph G can be detected with O(n) quantum queries to the adjacency
matrix. Under the promise that G either contains T as a subgraph or does not
contain T as a minor, we give O(n)-query quantum algorithms for detecting T
either a triangle or a subdivision of a star. All these algorithms can be
implemented time efficiently and, except for the triangle-detection algorithm,
in logarithmic space. One of the main techniques is to modify the
st-connectivity span program to drop along the way "breadcrumbs," which must be
retrieved before the path from s is allowed to enter t.Comment: 18 pages, 4 figure
Nearly Optimal Deterministic Algorithm for Sparse Walsh-Hadamard Transform
For every fixed constant , we design an algorithm for computing
the -sparse Walsh-Hadamard transform of an -dimensional vector in time . Specifically, the
algorithm is given query access to and computes a -sparse satisfying , for an absolute constant , where is the
transform of and is its best -sparse approximation. Our
algorithm is fully deterministic and only uses non-adaptive queries to
(i.e., all queries are determined and performed in parallel when the algorithm
starts).
An important technical tool that we use is a construction of nearly optimal
and linear lossless condensers which is a careful instantiation of the GUV
condenser (Guruswami, Umans, Vadhan, JACM 2009). Moreover, we design a
deterministic and non-adaptive compressed sensing scheme based
on general lossless condensers that is equipped with a fast reconstruction
algorithm running in time (for the GUV-based
condenser) and is of independent interest. Our scheme significantly simplifies
and improves an earlier expander-based construction due to Berinde, Gilbert,
Indyk, Karloff, Strauss (Allerton 2008).
Our methods use linear lossless condensers in a black box fashion; therefore,
any future improvement on explicit constructions of such condensers would
immediately translate to improved parameters in our framework (potentially
leading to reconstruction time with a reduced exponent in
the poly-logarithmic factor, and eliminating the extra parameter ).
Finally, by allowing the algorithm to use randomness, while still using
non-adaptive queries, the running time of the algorithm can be improved to
- …