31 research outputs found
Content-aware Traffic Engineering
Also appears as TU-Berlin technical report 2012-3, ISSN: 1436-9915Also appears as TU-Berlin technical report 2012-3, ISSN: 1436-9915Today, a large fraction of Internet traffic is originated by Content Providers (CPs) such as content distribution networks and hyper-giants. To cope with the increasing demand for content, CPs deploy massively distributed infrastructures. This poses new challenges for CPs as they have to dynamically map end-users to appropriate servers, without being fully aware of network conditions within an ISP as well as the end-users network locations. Furthermore, ISPs struggle to cope with rapid traffic shifts caused by the dynamic server selection process of CPs. In this paper, we argue that the challenges that CPs and ISPs face separately today can be turned into an opportunity. We show how they can jointly take advantage of the deployed distributed infrastructures to improve their operation and end-user performance. We propose Content-aware Traffic Engineering (CaTE), which dynamically adapts the traffic demand for content hosted on CPs by utilizing ISP network information and end-user location during the server selection process. As a result, CPs enhance their end-user to server mapping and improve end-user experience, thanks to the ability of network-informed server selection to circumvent network bottlenecks. In addition, ISPs gain the ability to partially influence the traffic demands in their networks. Our results with operational data show improvements in path length and delay between end-user and the assigned CP server, network wide traffic reduction of up to 15%, and a decrease in ISP link utilization of up to 40% when applying CaTE to traffic delivered by a small number of major CPs
Content Based Traffic Engineering in Software Defined Information Centric Networks
This paper describes a content centric network architecture which uses
software defined networking principles to implement efficient metadata driven
services by extracting content metadata at the network layer. The ability to
access content metadata transparently enables a number of new services in the
network. Specific examples discussed here include: a metadata driven traffic
engineering scheme which uses prior knowledge of content length to optimize
content delivery, a metadata driven content firewall which is more resilient
than traditional firewalls and differentiated treatment of content based on the
type of content being accessed. A detailed outline of an implementation of the
proposed architecture is presented along with some basic evaluation
On the Benefit of Information Centric Networks for Traffic Engineering
Current Internet performs traffic engineering (TE) by estimating traffic
matrices on a regular schedule, and allocating flows based upon weights
computed from these matrices. This means the allocation is based upon a guess
of the traffic in the network based on its history. Information-Centric
Networks on the other hand provide a finer-grained description of the traffic:
a content between a client and a server is uniquely identified by its name, and
the network can therefore learn the size of different content items, and
perform traffic engineering and resource allocation accordingly. We claim that
Information-Centric Networks can therefore provide a better handle to perform
traffic engineering, resulting in significant performance gain.
We present a mechanism to perform such resource allocation. We see that our
traffic engineering method only requires knowledge of the flow size (which, in
ICN, can be learned from previous data transfers) and outperforms a min-MLU
allocation in terms of response time. We also see that our method identifies
the traffic allocation patterns similar to that of min-MLU without having
access to the traffic matrix ahead of time. We show a very significant gain in
response time where min MLU is almost 50% slower than our ICN-based TE method
Temporal Locality in Today's Content Caching: Why it Matters and How to Model it
The dimensioning of caching systems represents a difficult task in the design
of infrastructures for content distribution in the current Internet. This paper
addresses the problem of defining a realistic arrival process for the content
requests generated by users, due its critical importance for both analytical
and simulative evaluations of the performance of caching systems. First, with
the aid of YouTube traces collected inside operational residential networks, we
identify the characteristics of real traffic that need to be considered or can
be safely neglected in order to accurately predict the performance of a cache.
Second, we propose a new parsimonious traffic model, named the Shot Noise Model
(SNM), that enables users to natively capture the dynamics of content
popularity, whilst still being sufficiently simple to be employed effectively
for both analytical and scalable simulative studies of caching systems.
Finally, our results show that the SNM presents a much better solution to
account for the temporal locality observed in real traffic compared to existing
approaches.Comment: 7 pages, 7 figures, Accepted for publication in ACM Computer
Communication Revie
End-to-end resource management for federated delivery of multimedia services
Recently, the Internet has become a popular platform for the delivery of multimedia content. Currently, multimedia services are either offered by Over-the-top (OTT) providers or by access ISPs over a managed IP network. As OTT providers offer their content across the best-effort Internet, they cannot offer any Quality of Service (QoS) guarantees to their users. On the other hand, users of managed multimedia services are limited to the relatively small selection of content offered by their own ISP. This article presents a framework that combines the advantages of both existing approaches, by dynamically setting up federations between the stakeholders involved in the content delivery process. Specifically, the framework provides an automated mechanism to set up end-to-end federations for QoS-aware delivery of multimedia content across the Internet. QoS contracts are automatically negotiated between the content provider, its customers, and the intermediary network domains. Additionally, a federated resource reservation algorithm is presented, which allows the framework to identify the optimal set of stakeholders and resources to include within a federation. Its goal is to minimize delivery costs for the content provider, while satisfying customer QoS requirements. Moreover, the presented framework allows intermediary storage sites to be included in these federations, supporting on-the-fly deployment of content caches along the delivery paths. The algorithm was thoroughly evaluated in order to validate our approach and assess the merits of including intermediary storage sites. The results clearly show the benefits of our method, with delivery cost reductions of up to 80 % in the evaluated scenario
Proactive multi-tenant cache management for virtualized ISP networks
The content delivery market has mainly been dominated by large Content Delivery Networks (CDNs) such as Akamai and Limelight. However, CDN traffic exerts a lot of pressure on Internet Service Provider (ISP) networks. Recently, ISPs have begun deploying so-called Telco CDNs, which have many advantages, such as reduced ISP network bandwidth utilization and improved Quality of Service (QoS) by bringing content closer to the end-user. Virtualization of storage and networking resources can enable the ISP to simultaneously lease its Telco CDN infrastructure to multiple third parties, opening up new business models and revenue streams. In this paper, we propose a proactive cache management system for ISP-operated multi-tenant Telco CDNs. The associated algorithm optimizes content placement and server selection across tenants and users, based on predicted content popularity and the geographical distribution of requests. Based on a Video-on-Demand (VoD) request trace of a leading European telecom operator, the presented algorithm is shown to reduce bandwidth usage by 17% compared to the traditional Least Recently Used (LRU) caching strategy, both inside the network and on the ingress links, while at the same time offering enhanced load balancing capabilities. Increasing the prediction accuracy is shown to have the potential to further improve bandwidth efficiency by up to 79%
Unravelling the Impact of Temporal and Geographical Locality in Content Caching Systems
To assess the performance of caching systems, the definition of a proper
process describing the content requests generated by users is required.
Starting from the analysis of traces of YouTube video requests collected inside
operational networks, we identify the characteristics of real traffic that need
to be represented and those that instead can be safely neglected. Based on our
observations, we introduce a simple, parsimonious traffic model, named Shot
Noise Model (SNM), that allows us to capture temporal and geographical locality
of content popularity. The SNM is sufficiently simple to be effectively
employed in both analytical and scalable simulative studies of caching systems.
We demonstrate this by analytically characterizing the performance of the LRU
caching policy under the SNM, for both a single cache and a network of caches.
With respect to the standard Independent Reference Model (IRM), some
paradigmatic shifts, concerning the impact of various traffic characteristics
on cache performance, clearly emerge from our results.Comment: 14 pages, 11 Figures, 2 Appendice
Towards multi-tenant cache management for ISP networks
The decreasing cost of storage and the advent of virtualization technology can allow Internet Service Providers (ISPs) to deploy multi-tenant caching infrastructures and lease them to content producers and Content Delivery Networks (CDNs). Serving content requests directly from the ISP network does not only reduce the delivery time, but also allows the ISP to optimize the network resources by controlling the placement and routing of content items. In this paper, we introduce a multi-tenant cache management approach that significantly reduces the bandwidth utilization of ISPs networks by pro-actively allocating caching space, leased by content producers and/or CDNs, and intelligently routing content to the end users. Using real content request traces, we show that the optimal solution to this problem can increase the cache hit ratio by 70.64% while reducing the bandwidth usage by 57.17% on average, compared to a commonly used reactive cache management scheme. These results provide a benchmark for the development of novel multi-tenant cache management strategies
Towards multi-tenant cache management for ISP networks
The decreasing cost of storage and the advent of virtualization technology can allow Internet Service Providers (ISPs) to deploy multi-tenant caching infrastructures and lease them to content producers and Content Delivery Networks (CDNs). Serving content requests directly from the ISP network does not only reduce the delivery time, but also allows the ISP to optimize the network resources by controlling the placement and routing of content items. In this paper, we introduce a multitenant cache management approach that significantly reduces the bandwidth utilization of ISPs networks by pro-actively allocating caching space, leased by content producers and/or CDNs, and intelligently routing content to the end users. Using real content request traces, we show that the optimal solution to this problem can increase the cache hit ratio by 70.64% while reducing the bandwidth usage by 57.17% on average, compared to a commonly used reactive cache management scheme. These results provide a benchmark for the development of novel multi-tenant cache management strategies
Service-centric networking
This chapter introduces a new paradigm for service centric networking. Building upon recent proposals in the area of information centric networking, a similar treatment of services – where networked software functions, rather than content, are dynamically deployed, replicated and invoked – is discussed. Service-centric networking provides the mechanisms required to deploy replicated service instances across highly distributed networked cloud infrastructures and to route client requests to the closest instance while providing more efficient network infrastructure usage, improved QoS and new business opportunities for application and service providers. </jats:p