263 research outputs found
Diffusive capture processes for information search
We show how effectively the diffusive capture processes (DCP) on complex
networks can be applied to information search in the networks. Numerical
simulations show that our method generates only 2% of traffic compared with the
most popular flooding-based query-packet-forwarding (FB) algorithm. We find
that the average searching time, , of the our model is more scalable than
another well known $n$-random walker model and comparable to the FB algorithm
both on real Gnutella network and scale-free networks with $\gamma =2.4$. We
also discuss the possible relationship between and , the second
moment of the degree distribution of the networks
Monitoring Challenges and Approaches for P2P File-Sharing Systems
Since the release of Napster in 1999, P2P file-sharing has enjoyed a dramatic rise in popularity. A 2000 study by Plonka on the University of Wisconsin campus network found that file-sharing accounted for a comparable volume of traffic to HTTP, while a 2002 study by Saroiu et al. on the University of Washington campus network found that file-sharing accounted for more than treble the volume of Web traffic observed, thus affirming the significance of P2P in the context of Internet traffic. Empirical studies of P2P traffic are essential for supporting the design of next-generation P2P systems, informing the provisioning of network infrastructure and underpinning the policing of P2P systems. The latter is of particular significance as P2P file-sharing systems have been implicated in supporting criminal behaviour including copyright infringement and the distribution of illegal pornograph
Centrality Measures for Networks with Community Structure
Understanding the network structure, and finding out the influential nodes is
a challenging issue in the large networks. Identifying the most influential
nodes in the network can be useful in many applications like immunization of
nodes in case of epidemic spreading, during intentional attacks on complex
networks. A lot of research is done to devise centrality measures which could
efficiently identify the most influential nodes in the network. There are two
major approaches to the problem: On one hand, deterministic strategies that
exploit knowledge about the overall network topology in order to find the
influential nodes, while on the other end, random strategies are completely
agnostic about the network structure. Centrality measures that can deal with a
limited knowledge of the network structure are required. Indeed, in practice,
information about the global structure of the overall network is rarely
available or hard to acquire. Even if available, the structure of the network
might be too large that it is too much computationally expensive to calculate
global centrality measures. To that end, a centrality measure is proposed that
requires information only at the community level to identify the influential
nodes in the network. Indeed, most of the real-world networks exhibit a
community structure that can be exploited efficiently to discover the
influential nodes. We performed a comparative evaluation of prominent global
deterministic strategies together with stochastic strategies with an available
and the proposed deterministic community-based strategy. Effectiveness of the
proposed method is evaluated by performing experiments on synthetic and
real-world networks with community structure in the case of immunization of
nodes for epidemic control.Comment: 30 pages, 4 figures. Accepted for publication in Physica A. arXiv
admin note: text overlap with arXiv:1411.627
Swarming Overlay Construction Strategies
Swarming peer-to-peer systems play an increasingly instrumental role in
Internet content distribution. It is therefore important to better understand
how these systems behave in practice. Recent research efforts have looked at
various protocol parameters and have measured how they affect system
performance and robustness. However, the importance of the strategy based on
which peers establish connections has been largely overlooked. This work
utilizes extensive simulations to examine the default overlay construction
strategy in BitTorrent systems. Based on the results, we identify a critical
parameter, the maximum allowable number of outgoing connections at each peer,
and evaluate its impact on the robustness of the generated overlay. We find
that there is no single optimal value for this parameter using the default
strategy. We then propose an alternative strategy that allows certain new peer
connection requests to replace existing connections. Further experiments with
the new strategy demonstrate that it outperforms the default one for all
considered metrics by creating an overlay more robust to churn. Additionally,
our proposed strategy exhibits optimal behavior for a well-defined value of the
maximum number of outgoing connections, thereby removing the need to set this
parameter in an ad-hoc manner
A framework for the dynamic management of Peer-to-Peer overlays
Peer-to-Peer (P2P) applications have been associated with inefficient operation, interference with other network services and large operational costs for network providers. This thesis presents a framework which can help ISPs address these issues by means of intelligent management of peer behaviour. The proposed approach involves limited control of P2P overlays without interfering with the fundamental characteristics of peer autonomy and decentralised operation.
At the core of the management framework lays the Active Virtual Peer (AVP). Essentially intelligent peers operated by the network providers, the AVPs interact with the overlay from within, minimising redundant or inefficient traffic, enhancing overlay stability and facilitating the efficient and balanced use of available peer and network resources. They offer an âinsiderâsâ view of the overlay and permit the management of P2P functions in a compatible and non-intrusive manner. AVPs can support multiple P2P protocols and coordinate to perform functions collectively.
To account for the multi-faceted nature of P2P applications and allow the incorporation of modern techniques and protocols as they appear, the framework is based on a modular architecture. Core modules for overlay control and transit traffic minimisation are presented. Towards the latter, a number of suitable P2P content caching strategies are proposed.
Using a purpose-built P2P network simulator and small-scale experiments, it is demonstrated that the introduction of AVPs inside the network can significantly reduce inter-AS traffic, minimise costly multi-hop flows, increase overlay stability and load-balancing and offer improved peer transfer performance
Residual-Based Measurement of Peer and Link Lifetimes in Gnutella Networks
Existing methods of measuring lifetimes in P2P systems usually rely on the so-called create-based method (CBM), which divides a given observation window into two halves and samples users created in the first half every Delta time units until they die or the observation period ends. Despite its frequent use, this approach has no rigorous accuracy or overhead analysis in the literature. To shed more light on its performance, we flrst derive a model for CBM and show that small window size or large Delta may lead to highly inaccurate lifetime distributions. We then show that create-based sampling exhibits an inherent tradeoff between overhead and accuracy, which does not allow any fundamental improvement to the method. Instead, we propose a completely different approach for sampling user dynamics that keeps track of only residual lifetimes of peers and uses a simple renewal-process model to recover the actual lifetimes from the observed residuals. Our analysis indicates that for reasonably large systems, the proposed method can reduce bandwidth consumption by several orders of magnitude compared to prior approaches while simultaneously achieving higher accuracy. We finish the paper by implementing a two-tier Gnutella network crawler equipped with the proposed sampling method and obtain the distribution of ultrapeer lifetimes in a network of 6.4 million users and 60 million links. Our experimental results show that ultrapeer lifetimes are Pareto with shape a alpha ap 1.1; however, link lifetimes exhibit much lighter tails with alpha ap 1.9
Residual-Based Estimation of Peer and Link Lifetimes in P2P Networks
Existing methods of measuring lifetimes in P2P systems usually rely on the so-called Create-BasedMethod (CBM), which divides a given observation window into two halves and samples users ldquocreatedrdquo in the first half every Delta time units until they die or the observation period ends. Despite its frequent use, this approach has no rigorous accuracy or overhead analysis in the literature. To shed more light on its performance, we first derive a model for CBM and show that small window size or large Delta may lead to highly inaccurate lifetime distributions. We then show that create-based sampling exhibits an inherent tradeoff between overhead and accuracy, which does not allow any fundamental improvement to the method. Instead, we propose a completely different approach for sampling user dynamics that keeps track of only residual lifetimes of peers and uses a simple renewal-process model to recover the actual lifetimes from the observed residuals. Our analysis indicates that for reasonably large systems, the proposed method can reduce bandwidth consumption by several orders of magnitude compared to prior approaches while simultaneously achieving higher accuracy. We finish the paper by implementing a two-tier Gnutella network crawler equipped with the proposed sampling method and obtain the distribution of ultrapeer lifetimes in a network of 6.4 million users and 60 million links. Our experimental results show that ultrapeer lifetimes are Pareto with shape alpha ap 1.1; however, link lifetimes exhibit much lighter tails with alpha ap 1.8
- âŠ