1,627 research outputs found
Reducing the Download Time in Stochastic P2P Content Delivery Networks by Improving Peer Selection
Peer-to-peer (P2P) applications have become a popular method for obtaining digital content. Recent research has shown that the amount of time spent downloading from a poor performing peer effects the total download duration. Current peer selection strategies attempt to limit the amount of time spent downloading from a poor performing peer, but they do not use both advanced knowledge and service capacity after the connection has been made to aid in peer selection. Advanced knowledge has traditionally been obtained from methods that add additional overhead to the P2P network, such as polling peers for service capacity information, using round trip time techniques to calculate the distance between peers, and by using tracker peers. This work investigated the creation of a new download strategy that replaced the random selection of peers with a method that selects server peers based on historic service capacity and ISP in order to further reduce the amount of time needed to complete a download session. Peer-to-peer (P2P) applications have become a popular method for obtaining digital content. Recent research has shown that the amount of time spent downloading from a poor performing peer effects the total download duration. Current peer selection strategies attempt to limit the amount of time spent downloading from a poor performing peer, but they do not use both advanced knowledge and service capacity after the connection has been made to aid in peer selection. Advanced knowledge has traditionally been obtained from methods that add additional overhead to the P2P network, such as polling peers for service capacity information, using round trip time techniques to calculate the distance between peers, and by using tracker peers. This work investigated the creation of a new download strategy that replaced the random selection of peers with a method that selects server peers based on historic service capacity and ISP in order to further reduce the amount of time needed to complete a download session. The results of this new historic based peer selection strategy have shown that there are benefits in using advanced knowledge to select peers and only replacing the worst performing peers. This new approach showed an average download duration improvement of 16.6% in the single client simulation and an average cross ISP traffic reduction of 55.17% when ISPs were participating in cross ISP throttling. In the multiple clients simulation the new approach showed an average download duration improvement of 53.31% and an average cross ISP traffic reduction of 88.83% when ISPs were participating in cross ISP throttling. This new approach also significantly improved the consistency of the download duration between download sessions allowing for the more accurate prediction of download times
Reducing BitTorrent Download Time via Handshake-Based Switching
Peer-to-peer networking overcomes the single point of failure and bandwidth limitations inherent to the centralized server model of file-sharing. It is both a popular means of sharing digital content and a major consumer of internet traffic, with BitTorrent being the most-used protocol. As such, significant research has gone into improving peer-to-peer performance in order to reduce both download times and networking costs. One aspect that can affect performance is the client’s selection of peers to download from, as the time spent downloading from even a single poor-performing peer can impact the overall download duration.
A recent peer selection strategy explored having a client use historical knowledge acquired through third-party sources, as well as its own first-hand experience with previously visited peers, as a means of selecting likely good-performers, coupled with a peer switching strategy that replaced peers whose post-selection downloads exhibited poor performance contrary to what historical knowledge suggested in order to limit the time spent downloading from said poor-performers Though this tactic demonstrated reduced download times compared to various past works, it still suffered from poor peer selection due to its historical knowledge not necessarily reflecting the current state of the peers.
This work introduced and examined an enhancement to this hybrid peer selection and switching strategy by adding current intelligence regarding a peer’s available bandwidth, all the while avoiding the additional network costs associated with performing on-the-fly probing or querying techniques utilized by other peer selection strategies to benchmark prospective peers. With such on-the-fly knowledge about a peer’s current bandwidth availability, this new enhanced strategy quickly replaced poor performers without waiting for downloads to be performed and subsequently benchmarked, resulting in reduced overall peer-to-peer download times.
The results of adding this pre-download peer switching enhancement demonstrated improved download performance, particularly in early file transfer runs. However, as more runs occurred and the benefits of the original strategy’s historical knowledge became more pronounced, the time savings gained from this new enhancement diminished
Piracy on the internet: Accommodate it or fight it? A dynamic approach
This paper uses a dynamic stochastic model to solve for the optimal pricing policy of themusic recording companies in the presence of P2P file-sharing networks eroding their CDsales. We employ a policy iteration algorithm on a discretized state space to numericallycompute the optimal price policy. The realistically calibrated model reflects the real-worldfigures we observe and provides estimates of figures we can not observe, such as changesin total welfare. The results suggest that, thanks to the existence of P2P networks, totalwelfare in 2008 in the U.S. is about $25.6 billion more per annum than in 1999 before P2Pwas introduced. Moreover, the results predict that the current trend of decreasing CDsales will continue until around the year 2020 when it will stabilize at around 231.2 millioncopies per year, comparing to the industry all-time high of 938.9 million in 1999. Thecomparative static analysis shows that full enforcement of intellectual property rights,although helpful for the industrial profit, may have adverse effect on total welfare.microeconomics ;
Recommended from our members
QoS - Aware content oriented flow routing in optical computer network
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.In this thesis, one of the most important issues in the field of networks communication is tackled and addressed. This issue is represented by QoS, where the increasing demand on highquality
applications together with the fast increase in the rates of Internet users have led to
massive traffic being transmitted on the Internet. This thesis proposes new ideas to manage the flow of this huge traffic in a manner that contributes in improving the communication QoS. This can be achieved by replacing the conventional application-insensitive routing schemes by others
which take into account the type of applications when making the routing decision. As a first contribution, the effect on the potential development in the quality of experience on the loading of
Basra optical network has been investigated. Furthermore, the traffic due to each application was dealt with in different ways according to their delay and loss sensitivities. Load rate distributions
over the various links due to the different applications were deployed to investigate the places of possible congestions in the network and the dominant applications that cause such congestions. In addition, OpenFlow and Optica Burst Switching (OBS) techniques were used to provide a wider range of network controllability and management. A centralised routing protocol
that takes into account the available bandwidth, delay, and security as three important QoS parameters, when forwarding traffics of different types, was proposed and implemented using OMNeT++ networks simulator. As a novel idea, security has been incorporated in our QoS requirements by incorporating Oyster Optics Technology (OOT) to secure some of the optical links aiming to supply the network with some secure paths for those applications that have high
privacy requirements. A particular type of traffic is to be routed according to the importance of these three QoS parameters for such a traffic type. The link utilisation, end to end delays and securities due to the different applications were recorded to prove the feasibility of our proposed
system. In order to decrease the amount of traffic overhead, the same QoS constraints were implemented on a distributed Ant colony based routing. The traditional Ant routing protocol was improved by adopting the idea of Red-Green-Blue (RGB) pheromones routing to incorporate these QoS constraints. Improvements of 11% load balancing, and 9% security for private data was achieved compared to the conventional Ant routing techniques. In addition, this Ant based
routing was utilised to propose an improved solution for the routing and wavelength assignment problem in the WDM optical computer networks
Provider-Controlled Bandwidth Management for HTTP-based Video Delivery
Over the past few years, a revolution in video delivery technology has taken place as mobile viewers and over-the-top (OTT) distribution paradigms have significantly changed the landscape of video delivery services. For decades, high quality video was only available in the home via linear television or physical media. Though Web-based services brought video to desktop and laptop computers, the dominance of proprietary delivery protocols and codecs inhibited research efforts. The recent emergence of HTTP adaptive streaming protocols has prompted a re-evaluation of legacy video delivery paradigms and introduced new questions as to the scalability and manageability of OTT video delivery.
This dissertation addresses the question of how to enable for content and network service providers the ability to monitor and manage large numbers of HTTP adaptive streaming clients in an OTT environment. Our early work focused on demonstrating the viability of server-side pacing schemes to produce an HTTP-based streaming server. We also investigated the ability of client-side pacing schemes to work with both commodity HTTP servers and our HTTP streaming server. Continuing our client-side pacing research, we developed our own client-side data proxy architecture which was implemented on a variety of mobile devices and operating systems. We used the portable client architecture as a platform for investigating different rate adaptation schemes and algorithms. We then concentrated on evaluating the network impact of multiple adaptive bitrate clients competing for limited network resources, and developing schemes for enforcing fair access to network resources.
The main contribution of this dissertation is the definition of segment-level client and network techniques for enforcing class of service (CoS) differentiation between OTT HTTP adaptive streaming clients. We developed a segment-level network proxy architecture which works transparently with adaptive bitrate clients through the use of segment replacement. We also defined a segment-level rate adaptation algorithm which uses download aborts to enforce CoS differentiation across distributed independent clients. The segment-level abstraction more accurately models application-network interactions and highlights the difference between segment-level and packet-level time scales. Our segment-level CoS enforcement techniques provide a foundation for creating scalable managed OTT video delivery services
- …