59 research outputs found
Asymptotic Miss Ratio of LRU Caching with Consistent Hashing
To efficiently scale data caching infrastructure to support emerging big data
applications, many caching systems rely on consistent hashing to group a large
number of servers to form a cooperative cluster. These servers are organized
together according to a random hash function. They jointly provide a unified
but distributed hash table to serve swift and voluminous data item requests.
Different from the single least-recently-used (LRU) server that has already
been extensively studied, theoretically characterizing a cluster that consists
of multiple LRU servers remains yet to be explored. These servers are not
simply added together; the random hashing complicates the behavior. To this
end, we derive the asymptotic miss ratio of data item requests on a LRU cluster
with consistent hashing. We show that these individual cache spaces on
different servers can be effectively viewed as if they could be pooled together
to form a single virtual LRU cache space parametrized by an appropriate cache
size. This equivalence can be established rigorously under the condition that
the cache sizes of the individual servers are large. For typical data caching
systems this condition is common. Our theoretical framework provides a
convenient abstraction that can directly apply the results from the simpler
single LRU cache to the more complex LRU cluster with consistent hashing.Comment: 11 pages, 4 figure
Implementation of Cache Attack on Real Information Centric Networking System
Network security is an ongoing major problem in today’s Internet world. Even though there have been simulation studies related to denial of service and cache attacks, studies of attacks on real networks are still lacking in the research. In this thesis, the effects of cache attacks in real information-centric networking systems were investigated. Cache attacks were implemented in real networks with different cache sizes and with Least Recently Used, Random and First In First Out algorithms to fill the caches in each node. The attacker hits the cache with unpopular content, making the user request that the results be fetched from web servers. The cache hit, time taken to get the result, and number of hops to serve the request were calculated with real network traffic. The results of the implementation are provided for different topologies and are compared with the simulation results
Recommended from our members
Resource Allocation in Distributed Service Networks
The past few years have witnessed significant growth in the use of distributed network analytics involving agile code, data and computational resources. In many such networked systems, for example, Internet of Things (IoT), a large number of smart devices, sensors, processing and storage resources are widely distributed in a geographic region. These devices and resources distributed over a physical space are collectively called a distributed service network. Efficient resource allocation in such high performance service networks remains one of the most critical problems. In this thesis, we model and optimize the allocation of resources in a distributed service network. This thesis contributes to two different types of service networks: caching, and spatial networks; and develops new techniques that optimize the overall performance of these services.
First, we propose a new method to compute an upper bound on hit probability for all non-anticipative caching policies in a distributed caching system. We find our bound to be tighter than state-of-the-art upper bounds for a variety of content request arrival processes. We then develop a utility based framework for content placement in a cache network for efficient and fair allocation of caching resources.
We develop provably optimal distributed algorithms that operate at each network cache to maximize the overall network utility. Next, we develop and evaluate assignment policies that allocate resources to users with a goal to minimize the expected distance traveled by a user request, where both resources and users are located on a line. Lastly, we design and evaluate resource proximity aware user-request allocation policies with a goal to reduce the implementation cost associated with moving a request/job to/from its allocated resource while balancing the number of requests allocated to a resource. Depending on the topology, our proposed policies achieve a 8% - 99% decrease in implementation cost as compared to the state-of-the-art
Database server workload characterization in an e-commerce environment
A typical E-commerce system that is deployed on the Internet has multiple layers that include Web users, Web servers, application servers, and a database server. As the system use and user request frequency increase, Web/application servers can be scaled up by replication. A load balancing proxy can be used to route user requests to individual machines that perform the same functionality. To address the increasing workload while avoiding replicating the database server, various dynamic caching policies have been proposed to reduce the database workload in E-commerce systems. However, the nature of the changes seen by the database server as a result of dynamic caching remains unknown. A good understanding of this change is fundamental for tuning a database server to get better performance. In this study, the TPC-W (a transactional Web E-commerce benchmark) workloads on a database server are characterized under two different dynamic caching mechanisms, which are
generalized and implemented as query-result cache and table cache. The characterization focuses on response time, CPU computation, buffer pool references, disk I/O references, and workload classification. This thesis combines a variety of analysis techniques: simulation, real time measurement and data mining. The experimental results in this thesis reveal some interesting effects that the dynamic caching has on the database server workload characteristics. The main observations include: (a) dynamic cache can considerably reduce the CPU usage of the database server and the number of database page references when it is heavily loaded; (b) dynamic cache can also reduce the database reference locality, but to a smaller
degree than that reported in file servers. The data classification results in this thesis show that with dynamic cache, the database server sees TPC-W profiles more like on-line transaction
processing workloads
Efficient caching algorithms for memory management in computer systems
As disk performance continues to lag behind that of memory systems and processors, fully utilizing memory to reduce disk accesses is a highly effective effort to improve the entire system performance. Furthermore, to serve the applications running on a computer in distributed systems, not only the local memory but also the memory on remote servers must be effectively managed to minimize I/O operations. The critical challenges in an effective memory cache management include: (1) Insightfully understanding and quantifying the locality inherent in the memory access requests; (2) Effectively utilizing the locality information in replacement algorithms; (3) Intelligently placing and replacing data in the multi-level caches of a distributed system; (4) Ensuring that the overheads of the proposed schemes are acceptable.;This dissertation provides solutions and makes unique and novel contributions in application locality quantification, general replacement algorithms, low-cost replacement policy, thrashing protection, as well as multi-level cache management in a distributed system. First, the dissertation proposes a new method to quantify locality strength, and accurately to identify the data with strong locality. It also provides a new replacement algorithm, which significantly outperforms existing algorithms. Second, considering the extremely low-cost requirements on replacement policies in virtual memory management, the dissertation proposes a policy meeting the requirements, and considerably exceeding the performance existing policies. Third, the dissertation provides an effective scheme to protect the system from thrashing for running memory-intensive applications. Finally, the dissertation provides a multi-level block placement and replacement protocol in a distributed client-server environment, exploiting non-uniform locality strengths in the I/O access requests.;The methodology used in this study include careful application behavior characterization, system requirement analysis, algorithm designs, trace-driven simulation, and system implementations. A main conclusion of the work is that there is still much room for innovation and significant performance improvement for the seemingly mature and stable policies that have been broadly used in the current operating system design
Characterizing Popularity Dynamics of User-generated Videos: A Category-based Study of YouTube
Understanding the growth pattern of content popularity has become a subject of immense interest to
Internet service providers, content makers and on-line advertisers. This understanding is also important for
the sustainable development of content distribution systems. As an approach to comprehend the characteristics of this growth pattern, a significant amount of research has been done in analyzing the popularity
growth patterns of YouTube videos. Unfortunately, no work has been done that intensively investigates the
popularity patterns of YouTube videos based on video object category. In this thesis, an in-depth analysis
of the popularity pattern of YouTube videos is performed, considering the categories of videos.
Metadata and request patterns were collected by employing category-specific YouTube crawlers. The
request patterns were observed for a period of five months. Results confirm that the time varying popularity
of di fferent YouTube categories are conspicuously diff erent, in spite of having sets of categories with very
similar viewing patterns. In particular, News and Sports exhibit similar growth curves, as do Music and
Film.
While for some categories views at early ages can be used to predict future popularity, for some others
predicting future popularity is a challenging task and require more sophisticated techniques, e.g., time-series clustering. The outcomes of these analyses are instrumental towards designing a reliable workload generator, which can be further used to evaluate diff erent caching policies for YouTube and similar sites. In this
thesis, workload generators for four of the YouTube categories are developed. Performance of these workload generators suggest that a complete category-specific workload generator can be developed using time-series clustering. Patterns of users' interaction with YouTube videos are also analyzed from a dataset collected in a local network. This shows the possible ways of improving the performance of Peer-to-Peer video distribution
technique along with a new video recommendation method
Pervasive Data Access in Wireless and Mobile Computing Environments
The rapid advance of wireless and portable computing technology has brought a lot of research interests and momentum to the area of mobile computing. One of the research focus is on pervasive data access. with wireless connections, users can access information at any place at any time. However, various constraints such as limited client capability, limited bandwidth, weak connectivity, and client mobility impose many challenging technical issues. In the past years, tremendous research efforts have been put forth to address the issues related to pervasive data access. A number of interesting research results were reported in the literature. This survey paper reviews important works in two important dimensions of pervasive data access: data broadcast and client caching. In addition, data access techniques aiming at various application requirements (such as time, location, semantics and reliability) are covered
Data Structures and Algorithms for Scalable NDN Forwarding
Named Data Networking (NDN) is a recently proposed general-purpose network architecture that aims to address the limitations of the Internet Protocol (IP), while maintaining its strengths. NDN takes an information-centric approach, focusing on named data rather than computer addresses. In NDN, the content is identified by its name, and each NDN packet has a name that specifies the content it is fetching or delivering. Since there are no source and destination addresses in an NDN packet, it is forwarded based on a lookup of its name in the forwarding plane, which consists of the Forwarding Information Base (FIB), Pending Interest Table (PIT), and Content Store (CS). In addition, as an in-network caching element, a scalable Repository (Repo) design is needed to provide large-scale long-term content storage in NDN networks.
Scalable NDN forwarding is a challenge. Compared to the well-understood approaches to IP forwarding, NDN forwarding performs lookups on packet names, which have variable and unbounded lengths, increasing the lookup complexity. The lookup tables are larger than in IP, requiring more memory space. Moreover, NDN forwarding has a read-write data plane, requiring per-packet updates at line rates. Designing and evaluating a scalable NDN forwarding node architecture is a major effort within the overall NDN research agenda.
The goal of this dissertation is to demonstrate that scalable NDN forwarding is feasible with the proposed data structures and algorithms. First, we propose a FIB lookup design based on the binary search of hash tables that provides a reliable longest name prefix lookup performance baseline for future NDN research. We have demonstrated 10 Gbps forwarding throughput with 256-byte packets and one billion synthetic forwarding rules, each containing up to seven name components. Second, we explore data structures and algorithms to optimize the FIB design based on the specific characteristics of real-world forwarding datasets. Third, we propose a fingerprint-only PIT design that reduces the memory requirements in the core routers. Lastly, we discuss the Content Store design issues and demonstrate that the NDN Repo implementation can leverage many of the existing databases and storage systems to improve performance
Optimization inWeb Caching: Cache Management, Capacity Planning, and Content Naming
Caching is fundamental to performance in distributed information retrieval systems
such as the World Wide Web. This thesis introduces novel techniques for optimizing performance
and cost-effectiveness in Web cache hierarchies.
When requests are served by nearby caches rather than distant servers, server loads and
network traffic decrease and transactions are faster. Cache system design and management,
however, face extraordinary challenges in loosely-organized environments like the Web,
where the many components involved in content creation, transport, and consumption are
owned and administered by different entities. Such environments call for decentralized
algorithms in which stakeholders act on local information and private preferences.
In this thesis I consider problems of optimally designing new Web cache hierarchies
and optimizing existing ones. The methods I introduce span the Web from point of content
creation to point of consumption: I quantify the impact of content-naming practices on
cache performance; present techniques for variable-quality-of-service cache management;
describe how a decentralized algorithm can compute economically-optimal cache sizes in
a branching two-level cache hierarchy; and introduce a new protocol extension that eliminates
redundant data transfers and allows “dynamic” content to be cached consistently.
To evaluate several of my new methods, I conducted trace-driven simulations on an
unprecedented scale. This in turn required novel workload measurement methods and efficient
new characterization and simulation techniques. The performance benefits of my proposed
protocol extension are evaluated using two extraordinarily large and detailed workload
traces collected in a traditional corporate network environment and an unconventional
thin-client system.
My empirical research follows a simple but powerful paradigm: measure on a large
scale an important production environment’s exogenous workload; identify performance
bounds inherent in the workload, independent of the system currently serving it; identify
gaps between actual and potential performance in the environment under study; and finally
devise ways to close these gaps through component modifications or through improved
inter-component integration. This approach may be applicable to a wide range of Web
services as they mature.Ph.D.Computer Science and EngineeringUniversity of Michiganhttp://deepblue.lib.umich.edu/bitstream/2027.42/90029/1/kelly-optimization_web_caching.pdfhttp://deepblue.lib.umich.edu/bitstream/2027.42/90029/2/kelly-optimization_web_caching.ps.bz
- …