2,212 research outputs found

    Group Testing with Probabilistic Tests: Theory, Design and Application

    Get PDF
    Identification of defective members of large populations has been widely studied in the statistics community under the name of group testing. It involves grouping subsets of items into different pools and detecting defective members based on the set of test results obtained for each pool. In a classical noiseless group testing setup, it is assumed that the sampling procedure is fully known to the reconstruction algorithm, in the sense that the existence of a defective member in a pool results in the test outcome of that pool to be positive. However, this may not be always a valid assumption in some cases of interest. In particular, we consider the case where the defective items in a pool can become independently inactive with a certain probability. Hence, one may obtain a negative test result in a pool despite containing some defective items. As a result, any sampling and reconstruction method should be able to cope with two different types of uncertainty, i.e., the unknown set of defective items and the partially unknown, probabilistic testing procedure. In this work, motivated by the application of detecting infected people in viral epidemics, we design non-adaptive sampling procedures that allow successful identification of the defective items through a set of probabilistic tests. Our design requires only a small number of tests to single out the defective items. In particular, for a population of size NN and at most KK defective items with activation probability pp, our results show that M=O(K2log⁥(N/K)/p3)M = O(K^2\log{(N/K)}/p^3) tests is sufficient if the sampling procedure should work for all possible sets of defective items, while M=O(Klog⁥(N)/p3)M = O(K\log{(N)}/p^3) tests is enough to be successful for any single set of defective items. Moreover, we show that the defective members can be recovered using a simple reconstruction algorithm with complexity of O(MN)O(MN).Comment: Full version of the conference paper "Compressed Sensing with Probabilistic Measurements: A Group Testing Solution" appearing in proceedings of the 47th Annual Allerton Conference on Communication, Control, and Computing, 2009 (arXiv:0909.3508). To appear in IEEE Transactions on Information Theor

    Non-Markov stochastic dynamics of real epidemic process of respiratory infections

    Full text link
    The study of social networks and especially of the stochastic dynamics of the diseases spread in human population has recently attracted considerable attention in statistical physics. In this work we present a new statistical method of analyzing the spread of epidemic processes of grippe and acute respiratory track infections (ARTI) by means of the theory of discrete non-Markov stochastic processes. We use the results of our last theory (Phys. Rev. E 65, 046107 (2002)) to study statistical memory effects, long - range correlation and discreteness in real data series, describing the epidemic dynamics of human ARTI infections and grippe. We have carried out the comparative analysis of the data of the two infections (grippe and ARTI) in one of the industrial districts of Kazan, one of the largest cities of Russia. The experimental data are analyzed by the power spectra of the initial time correlation function and the memory functions of junior orders, the phase portraits of the four first dynamic variables, the three first points of the statistical non-Markov parameter and the locally averaged kinetic and relaxation parameters. The received results give an opportunity to provide strict quantitative description of the regular and stochastic components in epidemic dynamics of social networks taking into account their time discreteness and effects of statistical memory. They also allow to reveal the degree of randomness and predictability of the real epidemic process in the specific social network.Comment: 18 pages, 8figs, 1 table

    Graph Signal Processing: Overview, Challenges and Applications

    Full text link
    Research in Graph Signal Processing (GSP) aims to develop tools for processing data defined on irregular graph domains. In this paper we first provide an overview of core ideas in GSP and their connection to conventional digital signal processing. We then summarize recent developments in developing basic GSP tools, including methods for sampling, filtering or graph learning. Next, we review progress in several application areas using GSP, including processing and analysis of sensor network data, biological data, and applications to image processing and machine learning. We finish by providing a brief historical perspective to highlight how concepts recently developed in GSP build on top of prior research in other areas.Comment: To appear, Proceedings of the IEE

    Controllability Metrics, Limitations and Algorithms for Complex Networks

    Full text link
    This paper studies the problem of controlling complex networks, that is, the joint problem of selecting a set of control nodes and of designing a control input to steer a network to a target state. For this problem (i) we propose a metric to quantify the difficulty of the control problem as a function of the required control energy, (ii) we derive bounds based on the system dynamics (network topology and weights) to characterize the tradeoff between the control energy and the number of control nodes, and (iii) we propose an open-loop control strategy with performance guarantees. In our strategy we select control nodes by relying on network partitioning, and we design the control input by leveraging optimal and distributed control techniques. Our findings show several control limitations and properties. For instance, for Schur stable and symmetric networks: (i) if the number of control nodes is constant, then the control energy increases exponentially with the number of network nodes, (ii) if the number of control nodes is a fixed fraction of the network nodes, then certain networks can be controlled with constant energy independently of the network dimension, and (iii) clustered networks may be easier to control because, for sufficiently many control nodes, the control energy depends only on the controllability properties of the clusters and on their coupling strength. We validate our results with examples from power networks, social networks, and epidemics spreading

    Link Prediction by De-anonymization: How We Won the Kaggle Social Network Challenge

    Full text link
    This paper describes the winning entry to the IJCNN 2011 Social Network Challenge run by Kaggle.com. The goal of the contest was to promote research on real-world link prediction, and the dataset was a graph obtained by crawling the popular Flickr social photo sharing website, with user identities scrubbed. By de-anonymizing much of the competition test set using our own Flickr crawl, we were able to effectively game the competition. Our attack represents a new application of de-anonymization to gaming machine learning contests, suggesting changes in how future competitions should be run. We introduce a new simulated annealing-based weighted graph matching algorithm for the seeding step of de-anonymization. We also show how to combine de-anonymization with link prediction---the latter is required to achieve good performance on the portion of the test set not de-anonymized---for example by training the predictor on the de-anonymized portion of the test set, and combining probabilistic predictions from de-anonymization and link prediction.Comment: 11 pages, 13 figures; submitted to IJCNN'201

    Spreading processes in Multilayer Networks

    Get PDF
    Several systems can be modeled as sets of interconnected networks or networks with multiple types of connections, here generally called multilayer networks. Spreading processes such as information propagation among users of an online social networks, or the diffusion of pathogens among individuals through their contact network, are fundamental phenomena occurring in these networks. However, while information diffusion in single networks has received considerable attention from various disciplines for over a decade, spreading processes in multilayer networks is still a young research area presenting many challenging research issues. In this paper we review the main models, results and applications of multilayer spreading processes and discuss some promising research directions.Comment: 21 pages, 3 figures, 4 table

    An efficient curing policy for epidemics on graphs

    Full text link
    We provide a dynamic policy for the rapid containment of a contagion process modeled as an SIS epidemic on a bounded degree undirected graph with n nodes. We show that if the budget rr of curing resources available at each time is Ω(W){\Omega}(W), where WW is the CutWidth of the graph, and also of order Ω(log⁥n){\Omega}(\log n), then the expected time until the extinction of the epidemic is of order O(n/r)O(n/r), which is within a constant factor from optimal, as well as sublinear in the number of nodes. Furthermore, if the CutWidth increases only sublinearly with n, a sublinear expected time to extinction is possible with a sublinearly increasing budget rr
    • 

    corecore