1,003 research outputs found
Exploring the Memory-Bandwidth Tradeoff in an Information-Centric Network
An information-centric network should realize significant economies by
exploiting a favourable memory-bandwidth tradeoff: it is cheaper to store
copies of popular content close to users than to fetch them repeatedly over the
Internet. We evaluate this tradeoff for some simple cache network structures
under realistic assumptions concerning the size of the content catalogue and
its popularity distribution. Derived cost formulas reveal the relative impact
of various cost, traffic and capacity parameters, allowing an appraisal of
possible future network architectures. Our results suggest it probably makes
more sense to envisage the future Internet as a loosely interconnected set of
local data centers than a network like today's with routers augmented by
limited capacity content stores.Comment: Proceedings of ITC 25 (International Teletraffic Congress), Shanghai,
September, 201
Cost-aware caching: optimizing cache provisioning and object placement in ICN
Caching is frequently used by Internet Service Providers as a viable
technique to reduce the latency perceived by end users, while jointly
offloading network traffic. While the cache hit-ratio is generally considered
in the literature as the dominant performance metric for such type of systems,
in this paper we argue that a critical missing piece has so far been neglected.
Adopting a radically different perspective, in this paper we explicitly account
for the cost of content retrieval, i.e. the cost associated to the external
bandwidth needed by an ISP to retrieve the contents requested by its customers.
Interestingly, we discover that classical cache provisioning techniques that
maximize cache efficiency (i.e., the hit-ratio), lead to suboptimal solutions
with higher overall cost. To show this mismatch, we propose two optimization
models that either minimize the overall costs or maximize the hit-ratio,
jointly providing cache sizing, object placement and path selection. We
formulate a polynomial-time greedy algorithm to solve the two problems and
analytically prove its optimality. We provide numerical results and show that
significant cost savings are attainable via a cost-aware design
Proactive multi-tenant cache management for virtualized ISP networks
The content delivery market has mainly been dominated by large Content Delivery Networks (CDNs) such as Akamai and Limelight. However, CDN traffic exerts a lot of pressure on Internet Service Provider (ISP) networks. Recently, ISPs have begun deploying so-called Telco CDNs, which have many advantages, such as reduced ISP network bandwidth utilization and improved Quality of Service (QoS) by bringing content closer to the end-user. Virtualization of storage and networking resources can enable the ISP to simultaneously lease its Telco CDN infrastructure to multiple third parties, opening up new business models and revenue streams. In this paper, we propose a proactive cache management system for ISP-operated multi-tenant Telco CDNs. The associated algorithm optimizes content placement and server selection across tenants and users, based on predicted content popularity and the geographical distribution of requests. Based on a Video-on-Demand (VoD) request trace of a leading European telecom operator, the presented algorithm is shown to reduce bandwidth usage by 17% compared to the traditional Least Recently Used (LRU) caching strategy, both inside the network and on the ingress links, while at the same time offering enhanced load balancing capabilities. Increasing the prediction accuracy is shown to have the potential to further improve bandwidth efficiency by up to 79%
Efficient Proactive Caching for Supporting Seamless Mobility
We present a distributed proactive caching approach that exploits user
mobility information to decide where to proactively cache data to support
seamless mobility, while efficiently utilizing cache storage using a congestion
pricing scheme. The proposed approach is applicable to the case where objects
have different sizes and to a two-level cache hierarchy, for both of which the
proactive caching problem is hard. Additionally, our modeling framework
considers the case where the delay is independent of the requested data object
size and the case where the delay is a function of the object size. Our
evaluation results show how various system parameters influence the delay gains
of the proposed approach, which achieves robust and good performance relative
to an oracle and an optimal scheme for a flat cache structure.Comment: 10 pages, 9 figure
Optimal Data Placement on Networks With Constant Number of Clients
We introduce optimal algorithms for the problems of data placement (DP) and
page placement (PP) in networks with a constant number of clients each of which
has limited storage availability and issues requests for data objects. The
objective for both problems is to efficiently utilize each client's storage
(deciding where to place replicas of objects) so that the total incurred access
and installation cost over all clients is minimized. In the PP problem an extra
constraint on the maximum number of clients served by a single client must be
satisfied. Our algorithms solve both problems optimally when all objects have
uniform lengths. When objects lengths are non-uniform we also find the optimal
solution, albeit a small, asymptotically tight violation of each client's
storage size by lmax where lmax is the maximum length of the objects
and some arbitrarily small positive constant. We make no assumption
on the underlying topology of the network (metric, ultrametric etc.), thus
obtaining the first non-trivial results for non-metric data placement problems
- …