778 research outputs found
Understanding the Properties of the BitTorrent Overlay
In this paper, we conduct extensive simulations to understand the properties
of the overlay generated by BitTorrent. We start by analyzing how the overlay
properties impact the efficiency of BitTorrent. We focus on the average peer
set size (i.e., average number of neighbors), the time for a peer to reach its
maximum peer set size, and the diameter of the overlay. In particular, we show
that the later a peer arrives in a torrent, the longer it takes to reach its
maximum peer set size. Then, we evaluate the impact of the maximum peer set
size, the maximum number of outgoing connections per peer, and the number of
NATed peers on the overlay properties. We show that BitTorrent generates a
robust overlay, but that this overlay is not a random graph. In particular, the
connectivity of a peer to its neighbors depends on its arriving order in the
torrent. We also show that a large number of NATed peers significantly
compromise the robustness of the overlay to attacks. Finally, we evaluate the
impact of peer exchange on the overlay properties, and we show that it
generates a chain-like overlay with a large diameter, which will adversely
impact the efficiency of large torrents
Swarming Overlay Construction Strategies
Swarming peer-to-peer systems play an increasingly instrumental role in
Internet content distribution. It is therefore important to better understand
how these systems behave in practice. Recent research efforts have looked at
various protocol parameters and have measured how they affect system
performance and robustness. However, the importance of the strategy based on
which peers establish connections has been largely overlooked. This work
utilizes extensive simulations to examine the default overlay construction
strategy in BitTorrent systems. Based on the results, we identify a critical
parameter, the maximum allowable number of outgoing connections at each peer,
and evaluate its impact on the robustness of the generated overlay. We find
that there is no single optimal value for this parameter using the default
strategy. We then propose an alternative strategy that allows certain new peer
connection requests to replace existing connections. Further experiments with
the new strategy demonstrate that it outperforms the default one for all
considered metrics by creating an overlay more robust to churn. Additionally,
our proposed strategy exhibits optimal behavior for a well-defined value of the
maximum number of outgoing connections, thereby removing the need to set this
parameter in an ad-hoc manner
Taxonomy of P2P Applications
Peer-to-peer (p2p) networks have gained immense popularity in recent years and the number of services they provide continuously rises. Where p2p-networks were formerly known as file-sharing networks, p2p is now also used for services like VoIP and IPTV. With so many different p2p applications and services the need for a taxonomy framework rises. This paper describes the available p2p applications grouped by the services they provide. A taxonomy framework is proposed to classify old and recent p2p applications based on their characteristics
On Constructing Persistent Identifiers with Persistent Resolution Targets
Persistent Identifiers (PID) are the foundation referencing digital assets in
scientific publications, books, and digital repositories. In its realization,
PIDs contain metadata and resolving targets in form of URLs that point to data
sets located on the network. In contrast to PIDs, the target URLs are typically
changing over time; thus, PIDs need continuous maintenance -- an effort that is
increasing tremendously with the advancement of e-Science and the advent of the
Internet-of-Things (IoT). Nowadays, billions of sensors and data sets are
subject of PID assignment. This paper presents a new approach of embedding
location independent targets into PIDs that allows the creation of
maintenance-free PIDs using content-centric network technology and overlay
networks. For proving the validity of the presented approach, the Handle PID
System is used in conjunction with Magnet Link access information encoding,
state-of-the-art decentralized data distribution with BitTorrent, and Named
Data Networking (NDN) as location-independent data access technology for
networks. Contrasting existing approaches, no green-field implementation of PID
or major modifications of the Handle System is required to enable
location-independent data dissemination with maintenance-free PIDs.Comment: Published IEEE paper of the FedCSIS 2016 (SoFAST-WS'16) conference,
11.-14. September 2016, Gdansk, Poland. Also available online:
http://ieeexplore.ieee.org/document/7733372
Mesmerizer: A Effective Tool for a Complete Peer-to-Peer Software Development Life-cycle
In this paper we present what are, in our experience, the best
practices in Peer-To-Peer(P2P) application development and
how we combined them in a middleware platform called Mesmerizer. We explain how simulation is an integral part of
the development process and not just an assessment tool.
We then present our component-based event-driven framework for P2P application development, which can be used
to execute multiple instances of the same application in a
strictly controlled manner over an emulated network layer
for simulation/testing, or a single application in a concurrent
environment for deployment purpose. We highlight modeling aspects that are of critical importance for designing and
testing P2P applications, e.g. the emulation of Network Address Translation and bandwidth dynamics. We show how
our simulator scales when emulating low-level bandwidth
characteristics of thousands of concurrent peers while preserving a good degree of accuracy compared to a packet-level
simulator
Systematizing Decentralization and Privacy: Lessons from 15 Years of Research and Deployments
Decentralized systems are a subset of distributed systems where multiple
authorities control different components and no authority is fully trusted by
all. This implies that any component in a decentralized system is potentially
adversarial. We revise fifteen years of research on decentralization and
privacy, and provide an overview of key systems, as well as key insights for
designers of future systems. We show that decentralized designs can enhance
privacy, integrity, and availability but also require careful trade-offs in
terms of system complexity, properties provided, and degree of
decentralization. These trade-offs need to be understood and navigated by
designers. We argue that a combination of insights from cryptography,
distributed systems, and mechanism design, aligned with the development of
adequate incentives, are necessary to build scalable and successful
privacy-preserving decentralized systems
CLOSER: A Collaborative Locality-aware Overlay SERvice
Current Peer-to-Peer (P2P) file sharing systems make use of a considerable percentage of Internet Service Providers (ISPs) bandwidth. This paper presents the Collaborative Locality-aware Overlay SERvice (CLOSER), an architecture that aims at lessening the usage of expensive international links by exploiting traffic locality (i.e., a resource is downloaded from the inside of the ISP whenever possible). The paper proves the effectiveness of CLOSER by analysis and simulation, also comparing this architecture with existing solutions for traffic locality in P2P systems. While savings on international links can be attractive for ISPs, it is necessary to offer some features that can be of interest for users to favor a wide adoption of the application. For this reason, CLOSER also introduces a privacy module that may arouse the users' interest and encourage them to switch to the new architectur
- âŠ