31,707 research outputs found

    Recovery rate affects the effective epidemic threshold with synchronous updating

    Get PDF
    Accurate identification of effective epidemic threshold is essential for understanding epidemic dynamics on complex networks. In this paper, we systematically study how the recovery rate affects the susceptible-infected-removed spreading dynamics on complex networks, where synchronous and asynchronous updating processes are taken into account. We derive the theoretical effective epidemic threshold and final outbreak size based on the edge-based compartmental theory. To validate the proposed theoretical predictions, extensive numerical experiments are implemented by using asynchronous and synchronous updating methods. When asynchronous updating method is used in simulations, recovery rate does not affect the final state of spreading dynamics. But with synchronous updating, we find that the effective epidemic threshold decreases with recovery rate, and final outbreak size increases with recovery rate. A good agreement between the theoretical predictions and the numerical results are observed on both synthetic and real-world networks. Our results extend the existing theoretical studies and help us to understand the phase transition with arbitrary recovery rate

    Asynchronous iterative computations with Web information retrieval structures: The PageRank case

    Get PDF
    There are several ideas being used today for Web information retrieval, and specifically in Web search engines. The PageRank algorithm is one of those that introduce a content-neutral ranking function over Web pages. This ranking is applied to the set of pages returned by the Google search engine in response to posting a search query. PageRank is based in part on two simple common sense concepts: (i)A page is important if many important pages include links to it. (ii)A page containing many links has reduced impact on the importance of the pages it links to. In this paper we focus on asynchronous iterative schemes to compute PageRank over large sets of Web pages. The elimination of the synchronizing phases is expected to be advantageous on heterogeneous platforms. The motivation for a possible move to such large scale distributed platforms lies in the size of matrices representing Web structure. In orders of magnitude: 101010^{10} pages with 101110^{11} nonzero elements and 101210^{12} bytes just to store a small percentage of the Web (the already crawled); distributed memory machines are necessary for such computations. The present research is part of our general objective, to explore the potential of asynchronous computational models as an underlying framework for very large scale computations over the Grid. The area of ``internet algorithmics'' appears to offer many occasions for computations of unprecedent dimensionality that would be good candidates for this framework.Comment: 8 pages to appear at ParCo2005 Conference Proceeding
    • …
    corecore