2,212 research outputs found
Group Testing with Probabilistic Tests: Theory, Design and Application
Identification of defective members of large populations has been widely
studied in the statistics community under the name of group testing. It
involves grouping subsets of items into different pools and detecting defective
members based on the set of test results obtained for each pool.
In a classical noiseless group testing setup, it is assumed that the sampling
procedure is fully known to the reconstruction algorithm, in the sense that the
existence of a defective member in a pool results in the test outcome of that
pool to be positive. However, this may not be always a valid assumption in some
cases of interest. In particular, we consider the case where the defective
items in a pool can become independently inactive with a certain probability.
Hence, one may obtain a negative test result in a pool despite containing some
defective items. As a result, any sampling and reconstruction method should be
able to cope with two different types of uncertainty, i.e., the unknown set of
defective items and the partially unknown, probabilistic testing procedure.
In this work, motivated by the application of detecting infected people in
viral epidemics, we design non-adaptive sampling procedures that allow
successful identification of the defective items through a set of probabilistic
tests. Our design requires only a small number of tests to single out the
defective items. In particular, for a population of size and at most
defective items with activation probability , our results show that tests is sufficient if the sampling procedure should
work for all possible sets of defective items, while
tests is enough to be successful for any single set of defective items.
Moreover, we show that the defective members can be recovered using a simple
reconstruction algorithm with complexity of .Comment: Full version of the conference paper "Compressed Sensing with
Probabilistic Measurements: A Group Testing Solution" appearing in
proceedings of the 47th Annual Allerton Conference on Communication, Control,
and Computing, 2009 (arXiv:0909.3508). To appear in IEEE Transactions on
Information Theor
Non-Markov stochastic dynamics of real epidemic process of respiratory infections
The study of social networks and especially of the stochastic dynamics of the
diseases spread in human population has recently attracted considerable
attention in statistical physics. In this work we present a new statistical
method of analyzing the spread of epidemic processes of grippe and acute
respiratory track infections (ARTI) by means of the theory of discrete
non-Markov stochastic processes. We use the results of our last theory (Phys.
Rev. E 65, 046107 (2002)) to study statistical memory effects, long - range
correlation and discreteness in real data series, describing the epidemic
dynamics of human ARTI infections and grippe. We have carried out the
comparative analysis of the data of the two infections (grippe and ARTI) in one
of the industrial districts of Kazan, one of the largest cities of Russia. The
experimental data are analyzed by the power spectra of the initial time
correlation function and the memory functions of junior orders, the phase
portraits of the four first dynamic variables, the three first points of the
statistical non-Markov parameter and the locally averaged kinetic and
relaxation parameters. The received results give an opportunity to provide
strict quantitative description of the regular and stochastic components in
epidemic dynamics of social networks taking into account their time
discreteness and effects of statistical memory. They also allow to reveal the
degree of randomness and predictability of the real epidemic process in the
specific social network.Comment: 18 pages, 8figs, 1 table
Graph Signal Processing: Overview, Challenges and Applications
Research in Graph Signal Processing (GSP) aims to develop tools for
processing data defined on irregular graph domains. In this paper we first
provide an overview of core ideas in GSP and their connection to conventional
digital signal processing. We then summarize recent developments in developing
basic GSP tools, including methods for sampling, filtering or graph learning.
Next, we review progress in several application areas using GSP, including
processing and analysis of sensor network data, biological data, and
applications to image processing and machine learning. We finish by providing a
brief historical perspective to highlight how concepts recently developed in
GSP build on top of prior research in other areas.Comment: To appear, Proceedings of the IEE
Controllability Metrics, Limitations and Algorithms for Complex Networks
This paper studies the problem of controlling complex networks, that is, the
joint problem of selecting a set of control nodes and of designing a control
input to steer a network to a target state. For this problem (i) we propose a
metric to quantify the difficulty of the control problem as a function of the
required control energy, (ii) we derive bounds based on the system dynamics
(network topology and weights) to characterize the tradeoff between the control
energy and the number of control nodes, and (iii) we propose an open-loop
control strategy with performance guarantees. In our strategy we select control
nodes by relying on network partitioning, and we design the control input by
leveraging optimal and distributed control techniques. Our findings show
several control limitations and properties. For instance, for Schur stable and
symmetric networks: (i) if the number of control nodes is constant, then the
control energy increases exponentially with the number of network nodes, (ii)
if the number of control nodes is a fixed fraction of the network nodes, then
certain networks can be controlled with constant energy independently of the
network dimension, and (iii) clustered networks may be easier to control
because, for sufficiently many control nodes, the control energy depends only
on the controllability properties of the clusters and on their coupling
strength. We validate our results with examples from power networks, social
networks, and epidemics spreading
Link Prediction by De-anonymization: How We Won the Kaggle Social Network Challenge
This paper describes the winning entry to the IJCNN 2011 Social Network
Challenge run by Kaggle.com. The goal of the contest was to promote research on
real-world link prediction, and the dataset was a graph obtained by crawling
the popular Flickr social photo sharing website, with user identities scrubbed.
By de-anonymizing much of the competition test set using our own Flickr crawl,
we were able to effectively game the competition. Our attack represents a new
application of de-anonymization to gaming machine learning contests, suggesting
changes in how future competitions should be run.
We introduce a new simulated annealing-based weighted graph matching
algorithm for the seeding step of de-anonymization. We also show how to combine
de-anonymization with link prediction---the latter is required to achieve good
performance on the portion of the test set not de-anonymized---for example by
training the predictor on the de-anonymized portion of the test set, and
combining probabilistic predictions from de-anonymization and link prediction.Comment: 11 pages, 13 figures; submitted to IJCNN'201
Spreading processes in Multilayer Networks
Several systems can be modeled as sets of interconnected networks or networks
with multiple types of connections, here generally called multilayer networks.
Spreading processes such as information propagation among users of an online
social networks, or the diffusion of pathogens among individuals through their
contact network, are fundamental phenomena occurring in these networks.
However, while information diffusion in single networks has received
considerable attention from various disciplines for over a decade, spreading
processes in multilayer networks is still a young research area presenting many
challenging research issues. In this paper we review the main models, results
and applications of multilayer spreading processes and discuss some promising
research directions.Comment: 21 pages, 3 figures, 4 table
An efficient curing policy for epidemics on graphs
We provide a dynamic policy for the rapid containment of a contagion process
modeled as an SIS epidemic on a bounded degree undirected graph with n nodes.
We show that if the budget of curing resources available at each time is
, where is the CutWidth of the graph, and also of order
, then the expected time until the extinction of the epidemic
is of order , which is within a constant factor from optimal, as well
as sublinear in the number of nodes. Furthermore, if the CutWidth increases
only sublinearly with n, a sublinear expected time to extinction is possible
with a sublinearly increasing budget
- âŠ