775 research outputs found
Sketch-based Influence Maximization and Computation: Scaling up with Guarantees
Propagation of contagion through networks is a fundamental process. It is
used to model the spread of information, influence, or a viral infection.
Diffusion patterns can be specified by a probabilistic model, such as
Independent Cascade (IC), or captured by a set of representative traces.
Basic computational problems in the study of diffusion are influence queries
(determining the potency of a specified seed set of nodes) and Influence
Maximization (identifying the most influential seed set of a given size).
Answering each influence query involves many edge traversals, and does not
scale when there are many queries on very large graphs. The gold standard for
Influence Maximization is the greedy algorithm, which iteratively adds to the
seed set a node maximizing the marginal gain in influence. Greedy has a
guaranteed approximation ratio of at least (1-1/e) and actually produces a
sequence of nodes, with each prefix having approximation guarantee with respect
to the same-size optimum. Since Greedy does not scale well beyond a few million
edges, for larger inputs one must currently use either heuristics or
alternative algorithms designed for a pre-specified small seed set size.
We develop a novel sketch-based design for influence computation. Our greedy
Sketch-based Influence Maximization (SKIM) algorithm scales to graphs with
billions of edges, with one to two orders of magnitude speedup over the best
greedy methods. It still has a guaranteed approximation ratio, and in practice
its quality nearly matches that of exact greedy. We also present influence
oracles, which use linear-time preprocessing to generate a small sketch for
each node, allowing the influence of any seed set to be quickly answered from
the sketches of its nodes.Comment: 10 pages, 5 figures. Appeared at the 23rd Conference on Information
and Knowledge Management (CIKM 2014) in Shanghai, Chin
Submodular Maximization Meets Streaming: Matchings, Matroids, and More
We study the problem of finding a maximum matching in a graph given by an
input stream listing its edges in some arbitrary order, where the quantity to
be maximized is given by a monotone submodular function on subsets of edges.
This problem, which we call maximum submodular-function matching (MSM), is a
natural generalization of maximum weight matching (MWM), which is in turn a
generalization of maximum cardinality matching (MCM). We give two incomparable
algorithms for this problem with space usage falling in the semi-streaming
range---they store only edges, using working memory---that
achieve approximation ratios of in a single pass and in
passes respectively. The operations of these algorithms
mimic those of Zelke's and McGregor's respective algorithms for MWM; the
novelty lies in the analysis for the MSM setting. In fact we identify a general
framework for MWM algorithms that allows this kind of adaptation to the broader
setting of MSM.
In the sequel, we give generalizations of these results where the
maximization is over "independent sets" in a very general sense. This
generalization captures hypermatchings in hypergraphs as well as independence
in the intersection of multiple matroids.Comment: 18 page
Quantum Separability and Entanglement Detection via Entanglement-Witness Search and Global Optimization
We focus on determining the separability of an unknown bipartite quantum
state by invoking a sufficiently large subset of all possible
entanglement witnesses given the expected value of each element of a set of
mutually orthogonal observables. We review the concept of an entanglement
witness from the geometrical point of view and use this geometry to show that
the set of separable states is not a polytope and to characterize the class of
entanglement witnesses (observables) that detect entangled states on opposite
sides of the set of separable states. All this serves to motivate a classical
algorithm which, given the expected values of a subset of an orthogonal basis
of observables of an otherwise unknown quantum state, searches for an
entanglement witness in the span of the subset of observables. The idea of such
an algorithm, which is an efficient reduction of the quantum separability
problem to a global optimization problem, was introduced in PRA 70 060303(R),
where it was shown to be an improvement on the naive approach for the quantum
separability problem (exhaustive search for a decomposition of the given state
into a convex combination of separable states). The last section of the paper
discusses in more generality such algorithms, which, in our case, assume a
subroutine that computes the global maximum of a real function of several
variables. Despite this, we anticipate that such algorithms will perform
sufficiently well on small instances that they will render a feasible test for
separability in some cases of interest (e.g. in 3-by-3 dimensional systems)
Thresholded Covering Algorithms for Robust and Max-Min Optimization
The general problem of robust optimization is this: one of several possible
scenarios will appear tomorrow, but things are more expensive tomorrow than
they are today. What should you anticipatorily buy today, so that the
worst-case cost (summed over both days) is minimized? Feige et al. and
Khandekar et al. considered the k-robust model where the possible outcomes
tomorrow are given by all demand-subsets of size k, and gave algorithms for the
set cover problem, and the Steiner tree and facility location problems in this
model, respectively.
In this paper, we give the following simple and intuitive template for
k-robust problems: "having built some anticipatory solution, if there exists a
single demand whose augmentation cost is larger than some threshold, augment
the anticipatory solution to cover this demand as well, and repeat". In this
paper we show that this template gives us improved approximation algorithms for
k-robust Steiner tree and set cover, and the first approximation algorithms for
k-robust Steiner forest, minimum-cut and multicut. All our approximation ratios
(except for multicut) are almost best possible.
As a by-product of our techniques, we also get algorithms for max-min
problems of the form: "given a covering problem instance, which k of the
elements are costliest to cover?".Comment: 24 page
Mixed state discrimination using optimal control
We present theory and experiment for the task of discriminating two
nonorthogonal states, given multiple copies. We implement several local
measurement schemes, on both pure states and states mixed by depolarizing
noise. We find that schemes which are optimal (or have optimal scaling) without
noise perform worse with noise than simply repeating the optimal single-copy
measurement. Applying optimal control theory, we derive the globally optimal
local measurement strategy, which outperforms all other local schemes, and
experimentally implement it for various levels of noise.Comment: Corrected ref 1 date; 4 pages & 4 figures + 2 pages & 3 figures
supplementary materia
Complexity of Strong Implementability
We consider the question of implementability of a social choice function in a
classical setting where the preferences of finitely many selfish individuals
with private information have to be aggregated towards a social choice. This is
one of the central questions in mechanism design. If the concept of weak
implementation is considered, the Revelation Principle states that one can
restrict attention to truthful implementations and direct revelation
mechanisms, which implies that implementability of a social choice function is
easy to check. For the concept of strong implementation, however, the
Revelation Principle becomes invalid, and the complexity of deciding whether a
given social choice function is strongly implementable has been open so far. In
this paper, we show by using methods from polyhedral theory that strong
implementability of a social choice function can be decided in polynomial space
and that each of the payments needed for strong implementation can always be
chosen to be of polynomial encoding length. Moreover, we show that strong
implementability of a social choice function involving only a single selfish
individual can be decided in polynomial time via linear programming
ForestHash: Semantic Hashing With Shallow Random Forests and Tiny Convolutional Networks
Hash codes are efficient data representations for coping with the ever
growing amounts of data. In this paper, we introduce a random forest semantic
hashing scheme that embeds tiny convolutional neural networks (CNN) into
shallow random forests, with near-optimal information-theoretic code
aggregation among trees. We start with a simple hashing scheme, where random
trees in a forest act as hashing functions by setting `1' for the visited tree
leaf, and `0' for the rest. We show that traditional random forests fail to
generate hashes that preserve the underlying similarity between the trees,
rendering the random forests approach to hashing challenging. To address this,
we propose to first randomly group arriving classes at each tree split node
into two groups, obtaining a significantly simplified two-class classification
problem, which can be handled using a light-weight CNN weak learner. Such
random class grouping scheme enables code uniqueness by enforcing each class to
share its code with different classes in different trees. A non-conventional
low-rank loss is further adopted for the CNN weak learners to encourage code
consistency by minimizing intra-class variations and maximizing inter-class
distance for the two random class groups. Finally, we introduce an
information-theoretic approach for aggregating codes of individual trees into a
single hash code, producing a near-optimal unique hash for each class. The
proposed approach significantly outperforms state-of-the-art hashing methods
for image retrieval tasks on large-scale public datasets, while performing at
the level of other state-of-the-art image classification techniques while
utilizing a more compact and efficient scalable representation. This work
proposes a principled and robust procedure to train and deploy in parallel an
ensemble of light-weight CNNs, instead of simply going deeper.Comment: Accepted to ECCV 201
Streaming Algorithms for Submodular Function Maximization
We consider the problem of maximizing a nonnegative submodular set function
subject to a -matchoid
constraint in the single-pass streaming setting. Previous work in this context
has considered streaming algorithms for modular functions and monotone
submodular functions. The main result is for submodular functions that are {\em
non-monotone}. We describe deterministic and randomized algorithms that obtain
a -approximation using -space, where is
an upper bound on the cardinality of the desired set. The model assumes value
oracle access to and membership oracles for the matroids defining the
-matchoid constraint.Comment: 29 pages, 7 figures, extended abstract to appear in ICALP 201
Tight Kernel Bounds for Problems on Graphs with Small Degeneracy
In this paper we consider kernelization for problems on d-degenerate graphs,
i.e. graphs such that any subgraph contains a vertex of degree at most .
This graph class generalizes many classes of graphs for which effective
kernelization is known to exist, e.g. planar graphs, H-minor free graphs, and
H-topological-minor free graphs. We show that for several natural problems on
d-degenerate graphs the best known kernelization upper bounds are essentially
tight.Comment: Full version of ESA 201
Optimizing fire station locations for the Istanbul metropolitan municipality
Copyright @ 2013 INFORMSThe Istanbul Metropolitan Municipality (IMM) seeks to determine locations for additional fire stations to build in Istanbul; its objective is to make residences and historic sites reachable by emergency vehicles within five minutes of a fire stationâs receipt of a service request. In this paper, we discuss our development of a mathematical model to aid IMM in determining these locations by using data retrieved from its fire incident records. We use a geographic information system to implement the model on Istanbulâs road network, and solve two location modelsâset-covering and maximal-coveringâas what-if scenarios. We discuss 10 scenarios, including the situation that existed when we initiated the project and the scenario that IMM implemented. The scenario implemented increases the cityâs fire station coverage from 58.6 percent to 85.9 percent, based on a five-minute response time, with an implementation plan that spans three years
- âŠ