721 research outputs found

    Wi-Fi Offload: Tragedy of the Commons or Land of Milk and Honey?

    Full text link
    Fueled by its recent success in provisioning on-site wireless Internet access, Wi-Fi is currently perceived as the best positioned technology for pervasive mobile macro network offloading. However, the broad transitions of multiple collocated operators towards this new paradigm may result in fierce competition for the common unlicensed spectrum at hand. In this light, our paper game-theoretically dissects market convergence scenarios by assessing the competition between providers in terms of network performance, capacity constraints, cost reductions, and revenue prospects. We will closely compare the prospects and strategic positioning of fixed line operators offering Wi-Fi services with respect to competing mobile network operators utilizing unlicensed spectrum. Our results highlight important dependencies upon inter-operator collaboration models, and more importantly, upon the ratio between backhaul and Wi-Fi access bit-rates. Furthermore, our investigation of medium- to long-term convergence scenarios indicates that a rethinking of control measures targeting the large-scale monetization of unlicensed spectrum may be required, as otherwise the used free bands may become subject to tragedy-of-commons type of problems.Comment: Workshop on Spectrum Sharing Strategies for Wireless Broadband Services, IEEE PIMRC'13, to appear 201

    A software-defined architecture for next-generation cellular networks

    Get PDF
    In the recent years, mobile cellular networks are undergoing fundamental changes and many established concepts are being revisited. New emerging paradigms, such as Software-Defined Networking (SDN), Mobile Cloud Computing (MCC), Network Function Virtualization (NFV), Internet of Things (IoT),and Mobile Social Networking (MSN), bring challenges in the design of cellular networks architectures. Current Long-Term Evolution (LTE) networks are not able to accommodate these new trends in a scalable and efficient way. In this paper, first we discuss the limitations of the current LTE architecture. Second, driven by the new communication needs and by the advances in aforementioned areas, we propose a new architecture for next generation cellular networks. Some of its characteristics include support for distributed content routing, Heterogeneous Networks(HetNets) and multiple Radio Access Technologies (RATs). Finally, we present simulation results which show that significant backhaul traffic savings can be achieved by implementing caching and routing functions at the network edge

    A Multi-Game Framework for Harmonized LTE-U and WiFi Coexistence over Unlicensed Bands

    Full text link
    The introduction of LTE over unlicensed bands (LTE-U) will enable LTE base stations (BSs) to boost their capacity and offload their traffic by exploiting the underused unlicensed bands. However, to reap the benefits of LTE-U, it is necessary to address various new challenges associated with LTE-U and WiFi coexistence. In particular, new resource management techniques must be developed to optimize the usage of the network resources while handling the interdependence between WiFi and LTE users and ensuring that WiFi users are not jeopardized. To this end, in this paper, a new game theoretic tool, dubbed as \emph{multi-game} framework is proposed as a promising approach for modeling resource allocation problems in LTE-U. In such a framework, multiple, co-existing and coupled games across heterogeneous channels can be formulated to capture the specific characteristics of LTE-U. Such games can be of different properties and types but their outcomes are largely interdependent. After introducing the basics of the multi-game framework, two classes of algorithms are outlined to achieve the new solution concepts of multi-games. Simulation results are then conducted to show how such a multi-game can effectively capture the specific properties of LTE-U and make of it a "friendly" neighbor to WiFi.Comment: Accepted for publication at IEEE Wireless Communications Magazine, Special Issue on LTE in Unlicensed Spectru

    Introducing mobile edge computing capabilities through distributed 5G Cloud Enabled Small Cells

    Get PDF
    Current trends in broadband mobile networks are addressed towards the placement of different capabilities at the edge of the mobile network in a centralised way. On one hand, the split of the eNB between baseband processing units and remote radio headers makes it possible to process some of the protocols in centralised premises, likely with virtualised resources. On the other hand, mobile edge computing makes use of processing and storage capabilities close to the air interface in order to deploy optimised services with minimum delay. The confluence of both trends is a hot topic in the definition of future 5G networks. The full centralisation of both technologies in cloud data centres imposes stringent requirements to the fronthaul connections in terms of throughput and latency. Therefore, all those cells with limited network access would not be able to offer these types of services. This paper proposes a solution for these cases, based on the placement of processing and storage capabilities close to the remote units, which is especially well suited for the deployment of clusters of small cells. The proposed cloud-enabled small cells include a highly efficient microserver with a limited set of virtualised resources offered to the cluster of small cells. As a result, a light data centre is created and commonly used for deploying centralised eNB and mobile edge computing functionalities. The paper covers the proposed architecture, with special focus on the integration of both aspects, and possible scenarios of application.Peer ReviewedPostprint (author's final draft

    Echo State Networks for Proactive Caching in Cloud-Based Radio Access Networks with Mobile Users

    Full text link
    In this paper, the problem of proactive caching is studied for cloud radio access networks (CRANs). In the studied model, the baseband units (BBUs) can predict the content request distribution and mobility pattern of each user, determine which content to cache at remote radio heads and BBUs. This problem is formulated as an optimization problem which jointly incorporates backhaul and fronthaul loads and content caching. To solve this problem, an algorithm that combines the machine learning framework of echo state networks with sublinear algorithms is proposed. Using echo state networks (ESNs), the BBUs can predict each user's content request distribution and mobility pattern while having only limited information on the network's and user's state. In order to predict each user's periodic mobility pattern with minimal complexity, the memory capacity of the corresponding ESN is derived for a periodic input. This memory capacity is shown to be able to record the maximum amount of user information for the proposed ESN model. Then, a sublinear algorithm is proposed to determine which content to cache while using limited content request distribution samples. Simulation results using real data from Youku and the Beijing University of Posts and Telecommunications show that the proposed approach yields significant gains, in terms of sum effective capacity, that reach up to 27.8% and 30.7%, respectively, compared to random caching with clustering and random caching without clustering algorithm.Comment: Accepted in the IEEE Transactions on Wireless Communication

    Improving Resilience of Communication in Information Dissemination for Time-Critical Applications

    Get PDF
    Severe weather impacts life and in this dire condition, people rely on communication, to organize relief and stay in touch with their loved ones. In such situations, cellular network infrastructure\footnote{We refer to cellular network infrastructure as infrastructure for the entirety of this document} might be affected due to power outage, link failures, etc. This urges us to look at Ad-hoc mode of communication, to offload major traffic partially or fully from the infrastructure, depending on the status of it. We look into threefold approach, ranging from the case where the infrastructure is completely unavailable, to where it has been replaced by make shift low capacity mobile cellular base station. First, we look into communication without infrastructure and timely, dissemination of weather alerts specific to geographical areas. We look into the specific case of floods as they affect significant number of people. Due to the nature of the problem we can utilize the properties of Information Centric Networking (ICN) in this context, namely: i) Flexibility and high failure resistance: Any node in the network that has the information can satisfy the query ii) Robust: Only sensor and car need to communicate iii) Fine grained geo-location specific information dissemination. We analyze how message forwarding using ICN on top of Ad hoc network, approach compares to the one based on infrastructure, that is less resilient in the case of disaster. In addition, we compare the performance of different message forwarding strategies in VANETs (Vehicular Adhoc Networks) using ICN. Our results show that ICN strategy outperforms the infrastructure-based approach as it is 100 times faster for 63\% of total messages delivered. Then we look into the case where we have the cellular network infrastructure, but it is being pressured due to rapid increase in volume of network traffic (as seen during a major event) or it has been replaced by low capacity mobile tower. In this case we look at offloading as much traffic as possible from the infrastructure to device-to-device communication. However, the host-oriented model of the TCP/IP-based Internet poses challenges to this communication pattern. A scheme that uses an ICN model to fetch content from nearby peers, increases the resiliency of the network in cases of outages and disasters. We collected content popularity statistics from social media to create a content request pattern and evaluate our approach through the simulation of realistic urban scenarios. Additionally, we analyze the scenario of large crowds in sports venues. Our simulation results show that we can offload traffic from the backhaul network by up to 51.7\%, suggesting an advantageous path to support the surge in traffic while keeping complexity and cost for the network operator at manageable levels. Finally, we look at adaptive bit-rate streaming (ABR) streaming, which has contributed significantly to the reduction of video playout stalling, mainly in highly variable bandwidth conditions. ABR clients continue to suffer from the variation of bit rate qualities over the duration of a streaming session. Similar to stalling, these variations in bit rate quality have a negative impact on the users’ Quality of Experience (QoE). We use a trace from a large-scale CDN to show that such quality changes occur in a significant amount of streaming sessions and investigate an ABR video segment retransmission approach to reduce the number of such quality changes. As the new HTTP/2 standard is becoming increasingly popular, we also see an increase in the usage of HTTP/2 as an alternative protocol for the transmission of web traffic including video streaming. Using various network conditions, we conduct a systematic comparison of existing transport layer approaches for HTTP/2 that is best suited for ABR segment retransmissions. Since it is well known that both protocols provide a series of improvements over HTTP/1.1, we perform experiments both in controlled environments and over transcontinental links in the Internet and find that these benefits also “trickle up” into the application layer when it comes to ABR video streaming where HTTP/2 retransmissions can significantly improve the average quality bitrate while simultaneously minimizing bit rate variations over the duration of a streaming session. Taking inspiration from the first two approaches, we take into account the resiliency of a multi-path approach and further look at a multi-path and multi-stream approach to ABR streaming and demonstrate that losses on one path have very little impact on the other from the same multi-path connection and this increases throughput and resiliency of communication
    • 

    corecore