86 research outputs found

    Efficient Traffic Management Algorithms for the Core Network using Device-to-Device Communication and Edge Caching

    Get PDF
    Exponentially growing number of communicating devices and the need for faster, more reliable and secure communication are becoming major challenges for current mobile communication architecture. More number of connected devices means more bandwidth and a need for higher Quality of Service (QoS) requirements, which bring new challenges in terms of resource and traffic management. Traffic offload to the edge has been introduced to tackle this demand-explosion that let the core network offload some of the contents to the edge to reduce the traffic congestion. Device-to-Device (D2D) communication and edge caching, has been proposed as promising solutions for offloading data. D2D communication refers to the communication infrastructure where the users in proximity communicate with each other directly. D2D communication improves overall spectral efficiency, however, it introduces additional interference in the system. To enable D2D communication, efficient resource allocation must be introduced in order to minimize the interference in the system and this benefits the system in terms of bandwidth efficiency. In the first part of this thesis, low complexity resource allocation algorithm using stable matching is proposed to optimally assign appropriate uplink resources to the devices in order to minimize interference among D2D and cellular users. Edge caching has recently been introduced as a modification of the caching scheme in the core network, which enables a cellular Base Station (BS) to keep copies of the contents in order to better serve users and enhance Quality of Experience (QoE). However, enabling BSs to cache data on the edge of the network brings new challenges especially on deciding on which and how the contents should be cached. Since users in the same cell may share similar content-needs, we can exploit this temporal-spatial correlation in the favor of caching system which is referred to local content popularity. Content popularity is the most important factor in the caching scheme which helps the BSs to cache appropriate data in order to serve the users more efficiently. In the edge caching scheme, the BS does not know the users request-pattern in advance. To overcome this bottleneck, a content popularity prediction using Markov Decision Process (MDP) is proposed in the second part of this thesis to let the BS know which data should be cached in each time-slot. By using the proposed scheme, core network access request can be significantly reduced and it works better than caching based on historical data in both stable and unstable content popularity

    A Survey of Deep Learning for Data Caching in Edge Network

    Full text link
    The concept of edge caching provision in emerging 5G and beyond mobile networks is a promising method to deal both with the traffic congestion problem in the core network as well as reducing latency to access popular content. In that respect end user demand for popular content can be satisfied by proactively caching it at the network edge, i.e, at close proximity to the users. In addition to model based caching schemes learning-based edge caching optimizations has recently attracted significant attention and the aim hereafter is to capture these recent advances for both model based and data driven techniques in the area of proactive caching. This paper summarizes the utilization of deep learning for data caching in edge network. We first outline the typical research topics in content caching and formulate a taxonomy based on network hierarchical structure. Then, a number of key types of deep learning algorithms are presented, ranging from supervised learning to unsupervised learning as well as reinforcement learning. Furthermore, a comparison of state-of-the-art literature is provided from the aspects of caching topics and deep learning methods. Finally, we discuss research challenges and future directions of applying deep learning for cachin

    A review on green caching strategies for next generation communication networks

    Get PDF
    © 2020 IEEE. In recent years, the ever-increasing demand for networking resources and energy, fueled by the unprecedented upsurge in Internet traffic, has been a cause for concern for many service providers. Content caching, which serves user requests locally, is deemed to be an enabling technology in addressing the challenges offered by the phenomenal growth in Internet traffic. Conventionally, content caching is considered as a viable solution to alleviate the backhaul pressure. However, recently, many studies have reported energy cost reductions contributed by content caching in cache-equipped networks. The hypothesis is that caching shortens content delivery distance and eventually achieves significant reduction in transmission energy consumption. This has motivated us to conduct this study and in this article, a comprehensive survey of the state-of-the-art green caching techniques is provided. This review paper extensively discusses contributions of the existing studies on green caching. In addition, the study explores different cache-equipped network types, solution methods, and application scenarios. We categorically present that the optimal selection of the caching nodes, smart resource management, popular content selection, and renewable energy integration can substantially improve energy efficiency of the cache-equipped systems. In addition, based on the comprehensive analysis, we also highlight some potential research ideas relevant to green content caching
    corecore