702 research outputs found

    Lifecycle-Aware Online Video Caching

    Get PDF
    The current explosion of video traffic compels service providers to deploy caches at edge networks. Nowadays, most caching systems store data with a high programming voltage corresponding to the largest possible ‘expiry date’, typically on the order of years, which maximizes the cache damage. However, popular videos rarely exhibit lifecycles longer than a couple of months. Consequently, the programming voltage can instead be adapted to fit the lifecycle and mitigate the cache damage accordingly. In this paper, we propose LiA-cache, a Lifecycle-Aware caching policy for online videos. LiA-cache finds both near-optimal caching retention times and cache eviction policies by optimizing traffic delivery cost and cache damage cost conjointly. We first investigate temporal patterns of video access from a real-world dataset covering 10 million online videos collected by one of the largest mobile network operators in the world. We next cluster the videos based on their access lifecycles and integrate the clustering into a general model of the caching system. Specifically, LiA-cache analyzes videos and caches them depending on their cluster label. Compared to other popular policies in real-world scenarios, LiA-cache can reduce cache damage up to 90%, while keeping a cache hit ratio close to a policy purely relying on video popularity.Peer reviewe

    Audience-retention-rate-aware caching and coded video delivery with asynchronous demands

    Get PDF
    Most of the current literature on coded caching focus on a static scenario, in which a fixed number of users synchronously place their requests from a content library, and the performance is measured in terms of the latency in satisfying all of these requests. In practice, however, users start watching an online video content asynchronously over time, and often abort watching a video before it is completed. The latter behaviour is captured by the notion of audience retention rate, which measures the portion of a video content watched on average. In order to bring coded caching one step closer to practice, asynchronous user demands are considered in this paper, by allowing user demands to arrive randomly over time, and both the popularity of video files, and the audience retention rates are taken into account. A decentralized partial coded delivery (PCD) scheme is proposed, and two cache allocation schemes are employed; namely homogeneous cache allocation (HoCA) and heterogeneous cache allocation (HeCA), which allocate users’ caches among different chunks of the video files in the library. Numerical results validate that the proposed PCD scheme, either with HoCA or HeCA, outperforms conventional uncoded caching as well as the state-of-the-art decentralized caching schemes, which consider only the file popularities, and are designed for synchronous demand arrivals. An information-theoretical lower bound on the average delivery rate is also presented

    Centralized coded caching of correlated contents

    Get PDF
    Coded caching and delivery is studied taking into account the correlations among the contents in the library. Correlations are modeled as common parts shared by multiple contents; that is, each file in the database is composed of a group of subfiles, where each subfile is shared by a different subset of files. The number of files that include a certain subfile is defined as the level of commonness of this subfile. First, a correlation-aware uncoded caching scheme is proposed, and it is shown that the optimal placement for this scheme gives priority to the subfiles with the highest levels of commonness. Then a correlation- aware coded caching scheme is presented, and the cache capacity allocated to subfiles with different levels of commonness is optimized in order to minimize the delivery rate. The proposed correlation-aware coded caching scheme is shown to remarkably outperform state-of-the-art correlation-ignorant solutions, indicating the benefits of exploiting content correlations in coded caching and delivery in networks

    Distributed Deep Learning at the Edge: A Novel Proactive and Cooperative Caching Framework for Mobile Edge Networks

    Full text link
    © 2012 IEEE. We propose two novel proactive cooperative caching approaches using deep learning (DL) to predict users' content demand in a mobile edge caching network. In the first approach, a content server (CS) takes responsibilities to collect information from all mobile edge nodes (MENs) in the network and then performs the proposed DL algorithm to predict the content demand for the whole network. However, such a centralized approach may disclose the private information because MENs have to share their local users' data with the CS. Thus, in the second approach, we propose a novel distributed deep learning (DDL)-based framework. The DDL allows MENs in the network to collaborate and exchange information to reduce the error of content demand prediction without revealing the private information of mobile users. Through simulation results, we show that our proposed approaches can enhance the accuracy by reducing the root mean squared error (RMSE) up to 33.7% and reduce the service delay by 47.4% compared with other machine learning algorithms
    • …
    corecore