57 research outputs found

    Proximity as a Service via Cellular Network-Assisted Mobile Device-to-Device

    Get PDF
    PhD ThesisThe research progress of communication has brought a lot of novel technologies to meet the multi-dimensional demands such as pervasive connection, low delay and high bandwidth. Device-to-Device (D2D) communication is a way to no longer treat the User Equipment (UEs) as a terminal, but rather as a part of the network for service provisioning. This thesis decouples UEs into service providers (helpers) and service requesters. By collaboration among proximal devices, with the coordination of cellular networks, some local tasks can be achieved, such as coverage extension, computation o oading, mobile crowdsourcing and mobile crowdsensing. This thesis proposes a generic framework Proximity as a Service (PaaS) for increasing the coverage with demands of service continuity. As one of the use cases, the optimal helper selection algorithm of PaaS for increasing the service coverage with demands of service continuity is called ContAct based Proximity (CAP). Mainly, fruitful contact information (e.g., contact duration, frequency, and interval) is captured, and is used to handle ubiquitous proximal services through the optimal selection of helpers. The nature of PaaS is evaluated under the Helsinki city scenario, with movement model of Points Of Interest (POI) and with critical factors in uencing the service demands (e.g., success ratio, disruption duration and frequency). Simulation results show the advantage of CAP, in both success ratio and continuity of the service (outputs). Based on this perspective, metrics such as service success ratio and continuity as a service evaluation of the PaaS are evaluated using the statistical theory of the Design Of Experiments (DOE). DOE is used as there are many dimensions to the state space (access tolerance, selected helper number, helper access limit, and transmit range) that can in uence the results. A key contribution of this work is that it brings rigorous statistical experiment design methods into the research into mobile computing. Results further reveal the influence of four factors (inputs), e.g., service tolerance, number of helpers allocated, the number of concurrent devices supported by each helper and transmit range. Based on this perspective, metrics such as service success ratio and continuity are evaluated using DOE. The results show that transmit range is the most dominant factor. The number of selected helpers is the second most dominant factor. Since di erent factors have di erent regression levels, a uni ed 4 level full factorial experiment and a cubic multiple regression analysis have been carried out. All the interactions and the corresponding coe cients have been found. This work is the rst one to evaluate LTE-Direct and WiFi-Direct in an opportunistic proximity service. The contribution of the results for industry is to guide how many users need to cooperate to enable mobile computing and for academia. This reveals the facts that: 1, in some cases, the improvement of spectrum e ciency brought by D2D is not important; 2, nodal density and the resources used in D2D air-interfaces are important in the eld of mobile computing. This work built a methodology to study the D2D networks with a di erent perspective (PaaS)

    Clustering algorithm for D2D communication in next generation cellular networks : thesis presented in partial fulfilment of the requirements for the degree of Doctor of Philosophy in Engineering, Massey University, Auckland, New Zealand

    Get PDF
    Next generation cellular networks will support many complex services for smartphones, vehicles, and other devices. To accommodate such services, cellular networks need to go beyond the capabilities of their previous generations. Device-to-Device communication (D2D) is a key technology that can help fulfil some of the requirements of future networks. The telecommunication industry expects a significant increase in the density of mobile devices which puts more pressure on centralized schemes and poses risk in terms of outages, poor spectral efficiencies, and low data rates. Recent studies have shown that a large part of the cellular traffic pertains to sharing popular contents. This highlights the need for decentralized and distributive approaches to managing multimedia traffic. Content-sharing via D2D clustered networks has emerged as a popular approach for alleviating the burden on the cellular network. Different studies have established that D2D communication in clusters can improve spectral and energy efficiency, achieve low latency while increasing the capacity of the network. To achieve effective content-sharing among users, appropriate clustering strategies are required. Therefore, the aim is to design and compare clustering approaches for D2D communication targeting content-sharing applications. Currently, most of researched and implemented clustering schemes are centralized or predominantly dependent on Evolved Node B (eNB). This thesis proposes a distributed architecture that supports clustering approaches to incorporate multimedia traffic. A content-sharing network is presented where some D2D User Equipment (DUE) function as content distributors for nearby devices. Two promising techniques are utilized, namely, Content-Centric Networking and Network Virtualization, to propose a distributed architecture, that supports efficient content delivery. We propose to use clustering at the user level for content-distribution. A weighted multi-factor clustering algorithm is proposed for grouping the DUEs sharing a common interest. Various performance parameters such as energy consumption, area spectral efficiency, and throughput have been considered for evaluating the proposed algorithm. The effect of number of clusters on the performance parameters is also discussed. The proposed algorithm has been further modified to allow for a trade-off between fairness and other performance parameters. A comprehensive simulation study is presented that demonstrates that the proposed clustering algorithm is more flexible and outperforms several well-known and state-of-the-art algorithms. The clustering process is subsequently evaluated from an individual user’s perspective for further performance improvement. We believe that some users, sharing common interests, are better off with the eNB rather than being in the clusters. We utilize machine learning algorithms namely, Deep Neural Network, Random Forest, and Support Vector Machine, to identify the users that are better served by the eNB and form clusters for the rest of the users. This proposed user segregation scheme can be used in conjunction with most clustering algorithms including the proposed multi-factor scheme. A comprehensive simulation study demonstrates that with such novel user segregation, the performance of individual users, as well as the whole network, can be significantly improved for throughput, energy consumption, and fairness

    Streaming and User Behaviour in Omnidirectional Videos

    Get PDF
    Omnidirectional videos (ODVs) have gone beyond the passive paradigm of traditional video, offering higher degrees of immersion and interaction. The revolutionary novelty of this technology is the possibility for users to interact with the surrounding environment, and to feel a sense of engagement and presence in a virtual space. Users are clearly the main driving force of immersive applications and consequentially the services need to be properly tailored to them. In this context, this chapter highlights the importance of the new role of users in ODV streaming applications, and thus the need for understanding their behaviour while navigating within ODVs. A comprehensive overview of the research efforts aimed at advancing ODV streaming systems is also presented. In particular, the state-of-the-art solutions under examination in this chapter are distinguished in terms of system-centric and user-centric streaming approaches: the former approach comes from a quite straightforward extension of well-established solutions for the 2D video pipeline while the latter one takes the benefit of understanding users’ behaviour and enable more personalised ODV streaming

    Gestion conjointe de ressources de communication et de calcul pour les réseaux sans fils à base de cloud

    Get PDF
    Mobile Edge Cloud brings the cloud closer to mobile users by moving the cloud computational efforts from the internet to the mobile edge. We adopt a local mobile edge cloud computing architecture, where small cells are empowered with computational and storage capacities. Mobile users’ offloaded computational tasks are executed at the cloud-enabled small cells. We propose the concept of small cells clustering for mobile edge computing, where small cells cooperate in order to execute offloaded computational tasks. A first contribution of this thesis is the design of a multi-parameter computation offloading decision algorithm, SM-POD. The proposed algorithm consists of a series of low complexity successive and nested classifications of computational tasks at the mobile side, leading to local computation, or offloading to the cloud. To reach the offloading decision, SM-POD jointly considers computational tasks, handsets, and communication channel parameters. In the second part of this thesis, we tackle the problem of small cell clusters set up for mobile edge cloud computing for both single-user and multi-user cases. The clustering problem is formulated as an optimization that jointly optimizes the computational and communication resource allocation, and the computational load distribution on the small cells participating in the computation cluster. We propose a cluster sparsification strategy, where we trade cluster latency for higher system energy efficiency. In the multi-user case, the optimization problem is not convex. In order to compute a clustering solution, we propose a convex reformulation of the problem, and we prove that both problems are equivalent. With the goal of finding a lower complexity clustering solution, we propose two heuristic small cells clustering algorithms. The first algorithm is based on resource allocation on the serving small cells where tasks are received, as a first step. Then, in a second step, unserved tasks are sent to a small cell managing unit (SCM) that sets up computational clusters for the execution of these tasks. The main idea of this algorithm is task scheduling at both serving small cells, and SCM sides for higher resource allocation efficiency. The second proposed heuristic is an iterative approach in which serving small cells compute their desired clusters, without considering the presence of other users, and send their cluster parameters to the SCM. SCM then checks for excess of resource allocation at any of the network small cells. SCM reports any load excess to serving small cells that re-distribute this load on less loaded small cells. In the final part of this thesis, we propose the concept of computation caching for edge cloud computing. With the aim of reducing the edge cloud computing latency and energy consumption, we propose caching popular computational tasks for preventing their re-execution. Our contribution here is two-fold: first, we propose a caching algorithm that is based on requests popularity, computation size, required computational capacity, and small cells connectivity. This algorithm identifies requests that, if cached and downloaded instead of being re-computed, will increase the computation caching energy and latency savings. Second, we propose a method for setting up a search small cells cluster for finding a cached copy of the requests computation. The clustering policy exploits the relationship between tasks popularity and their probability of being cached, in order to identify possible locations of the cached copy. The proposed method reduces the search cluster size while guaranteeing a minimum cache hit probability.Cette thèse porte sur le paradigme « Mobile Edge cloud» qui rapproche le cloud des utilisateurs mobiles et qui déploie une architecture de clouds locaux dans les terminaisons du réseau. Les utilisateurs mobiles peuvent désormais décharger leurs tâches de calcul pour qu’elles soient exécutées par les femto-cellules (FCs) dotées de capacités de calcul et de stockage. Nous proposons ainsi un concept de regroupement de FCs dans des clusters de calculs qui participeront aux calculs des tâches déchargées. A cet effet, nous proposons, dans un premier temps, un algorithme de décision de déportation de tâches vers le cloud, nommé SM-POD. Cet algorithme prend en compte les caractéristiques des tâches de calculs, des ressources de l’équipement mobile, et de la qualité des liens de transmission. SM-POD consiste en une série de classifications successives aboutissant à une décision de calcul local, ou de déportation de l’exécution dans le cloud.Dans un deuxième temps, nous abordons le problème de formation de clusters de calcul à mono-utilisateur et à utilisateurs multiples. Nous formulons le problème d’optimisation relatif qui considère l’allocation conjointe des ressources de calculs et de communication, et la distribution de la charge de calcul sur les FCs participant au cluster. Nous proposons également une stratégie d’éparpillement, dans laquelle l’efficacité énergétique du système est améliorée au prix de la latence de calcul. Dans le cas d’utilisateurs multiples, le problème d’optimisation d’allocation conjointe de ressources n’est pas convexe. Afin de le résoudre, nous proposons une reformulation convexe du problème équivalente à la première puis nous proposons deux algorithmes heuristiques dans le but d’avoir un algorithme de formation de cluster à complexité réduite. L’idée principale du premier est l’ordonnancement des tâches de calculs sur les FCs qui les reçoivent. Les ressources de calculs sont ainsi allouées localement au niveau de la FC. Les tâches ne pouvant pas être exécutées sont, quant à elles, envoyées à une unité de contrôle (SCM) responsable de la formation des clusters de calculs et de leur exécution. Le second algorithme proposé est itératif et consiste en une formation de cluster au niveau des FCs ne tenant pas compte de la présence d’autres demandes de calculs dans le réseau. Les propositions de cluster sont envoyées au SCM qui évalue la distribution des charges sur les différentes FCs. Le SCM signale tout abus de charges pour que les FCs redistribuent leur excès dans des cellules moins chargées.Dans la dernière partie de la thèse, nous proposons un nouveau concept de mise en cache des calculs dans l’Edge cloud. Afin de réduire la latence et la consommation énergétique des clusters de calculs, nous proposons la mise en cache de calculs populaires pour empêcher leur réexécution. Ici, notre contribution est double : d’abord, nous proposons un algorithme de mise en cache basé, non seulement sur la popularité des tâches de calculs, mais aussi sur les tailles et les capacités de calculs demandés, et la connectivité des FCs dans le réseau. L’algorithme proposé identifie les tâches aboutissant à des économies d’énergie et de temps plus importantes lorsqu’elles sont téléchargées d’un cache au lieu d’être recalculées. Nous proposons ensuite d’exploiter la relation entre la popularité des tâches et la probabilité de leur mise en cache, pour localiser les emplacements potentiels de leurs copies. La méthode proposée est basée sur ces emplacements, et permet de former des clusters de recherche de taille réduite tout en garantissant de retrouver une copie en cache

    Network and Content Intelligence for 360 Degree Video Streaming Optimization

    Get PDF
    In recent years, 360° videos, a.k.a. spherical frames, became popular among users creating an immersive streaming experience. Along with the advances in smart- phones and Head Mounted Devices (HMD) technology, many content providers have facilitated to host and stream 360° videos in both on-demand and live stream- ing modes. Therefore, many different applications have already arisen leveraging these immersive videos, especially to give viewers an impression of presence in a digital environment. For example, with 360° videos, now it is possible to connect people in a remote meeting in an interactive way which essentially increases the productivity of the meeting. Also, creating interactive learning materials using 360° videos for students will help deliver the learning outcomes effectively. However, streaming 360° videos is not an easy task due to several reasons. First, 360° video frames are 4–6 times larger than normal video frames to achieve the same quality as a normal video. Therefore, delivering these videos demands higher bandwidth in the network. Second, processing relatively larger frames requires more computational resources at the end devices, particularly for end user devices with limited resources. This will impact not only the delivery of 360° videos but also many other applications running on shared resources. Third, these videos need to be streamed with very low latency requirements due their interactive nature. Inability to satisfy these requirements can result in poor Quality of Experience (QoE) for the user. For example, insufficient bandwidth incurs frequent rebuffer- ing and poor video quality. Also, inadequate computational capacity can cause faster battery draining and unnecessary heating of the device, causing discomfort to the user. Motion or cyber–sickness to the user will be prevalent if there is an unnecessary delay in streaming. These circumstances will hinder providing im- mersive streaming experiences to the much-needed communities, especially those who do not have enough network resources. To address the above challenges, we believe that enhancements to the three main components in video streaming pipeline, server, network and client, are essential. Starting from network, it is beneficial for network providers to identify 360° video flows as early as possible and understand their behaviour in the network to effec- tively allocate sufficient resources for this video delivery without compromising the quality of other services. Content servers, at one end of this streaming pipeline, re- quire efficient 360° video frame processing mechanisms to support adaptive video streaming mechanisms such as ABR (Adaptive Bit Rate) based streaming, VP aware streaming, a streaming paradigm unique to 360° videos that select only part of the larger video frame that fall within the user-visible region, etc. On the other end, the client can be combined with edge-assisted streaming to deliver 360° video content with reduced latency and higher quality. Following the above optimization strategies, in this thesis, first, we propose a mech- anism named 360NorVic to extract 360° video flows from encrypted video traffic and analyze their traffic characteristics. We propose Machine Learning (ML) mod- els to classify 360° and normal videos under different scenarios such as offline, near real-time, VP-aware streaming and Mobile Network Operator (MNO) level stream- ing. Having extracted 360° video traffic traces both in packet and flow level data at higher accuracy, we analyze and understand the differences between 360° and normal video patterns in the encrypted traffic domain that is beneficial for effec- tive resource optimization for enhancing 360° video delivery. Second, we present a WGAN (Wesserstien Generative Adversarial Network) based data generation mechanism (namely VideoTrain++) to synthesize encrypted network video traffic, taking minimal data. Leveraging synthetic data, we show improved performance in 360° video traffic analysis, especially in ML-based classification in 360NorVic. Thirdly, we propose an effective 360° video frame partitioning mechanism (namely VASTile) at the server side to support VP-aware 360° video streaming with dy- namic tiles (or variable tiles) of different sizes and locations on the frame. VASTile takes a visual attention map on the video frames as the input and applies a com- putational geometric approach to generate a non-overlapping tile configuration to cover the video frames adaptive to the visual attention. We present VASTile as a scalable approach for video frame processing at the servers and a method to re- duce bandwidth consumption in network data transmission. Finally, by applying VASTile to the individual user VP at the client side and utilizing cache storage of Multi Access Edge Computing (MEC) servers, we propose OpCASH, a mech- anism to personalize the 360° video streaming with dynamic tiles with the edge assistance. While proposing an ILP based solution to effectively select cached variable tiles from MEC servers that might not be identical to the requested VP tiles by user, but still effectively cover the same VP region, OpCASH maximize the cache utilization and reduce the number of requests to the content servers in congested core network. With this approach, we demonstrate the gain in latency and bandwidth saving and video quality improvement in personalized 360° video streaming

    Compound popular content caching strategy to enhance the cache management performance in named data networking

    Get PDF
    Named Data Networking (NDN) is a leading research paradigm for the future Internet architecture. The NDN offers in-network cache which is the most beneficial feature to reduce the difficulties of the location-based Internet paradigm. The objective of cache is to achieve a scalable, effective, and consistent distribution of information. However, the main issue which NDN facing is the selection of appropriate router during the content’s transmission that can disrupt the overall network performance. The reason is that how each router takes a decision to the cache which content needs to cache at what location that can enhance the complete caching performance. Therefore, several cache management strategies have been developed. Still, it is not clear which caching strategy is the most ideal for each situation. This study proposes a new cache management strategy named as Compound Popular Content Caching Strategy (CPCCS) to minimize cache redundancy with enhanced diversity ratio and improving the accessibility of cached content by providing short stretch paths. The CPCCS was developed by combining two mechanisms named as Compound Popular Content Selection (CPCS) and Compound Popular Content Caching (CPCC) to differentiate the contents regarding their Interest frequencies using dynamic threshold and to find the best possible caching positions respectively. CPCCS is compared with other NDN-based caching strategies, such as Max-Gain In-network Caching, WAVE popularity-based caching strategy, Hop-based Probabilistic Caching, Leaf Popular Down, Most Popular Cache, and Cache Capacity Aware Caching in a simulation environment. The results show that the CPCCS performs better in which the diversity and cache hit ratio are increased by 34% and 14% respectively. In addition, the redundancy and path stretch are decreased by 44% and 46% respectively. The outcomes showed that the CPCCS have achieved enhanced caching performance with respect to different cache size (1GB to 10GB) and simulation parameters than other caching strategies. Thus, CPCCS can be applicable in future for the NDN-based emerging technologies such as Internet of Things, fog and edge computing

    Performance Analysis of Non-Orthogonal Multiple Access (NOMA) in C-RAN, H-CRAN and F-RAN for 5G Systems

    Get PDF
    The world of telecommunication is witnessing a swift transformation towards fifth generation (5G) cellular networks. The future networks present requisite needs in ubiquitous throughput, low latency, and high reliability. They are also envisioned to provide diversified services such as enhanced Mobile BroadBand (eMBB) and ultra-reliable low-latency communication (URLLC) as well as improved quality of user experience. More interestingly, a novel mobile network architecture allowing centralized processing and cloud computing has been proposed as one of the best candidates for fifth generation. It is denoted as Cloud Radio Access Network (CRAN) and Heterogeneous Cloud Radio Access Network (H-CRAN). Furthermore, the 5G architecture will be fog-like, namely fog radio access networks (F-RAN) enabling a functional split of network functionalities between cloud and edge nodes with caching and fog computing capabilities. Meanwhile non-orthogonal multiple access (NOMA) has been proposed as an promising multiple access (MA) technology for future radio access networks (RANs) to meet the heterogeneous demands for high throughput, low latency and massive connectivity. One of the main challenges of NOMA is that how well it is to be compatible with other emerging techniques for meeting the requirements of 5G. However, comprehensive performance analysis on NOMA and practical resource allocation designs in co-existence with other emerging networks have not been fully studied and investigated in the literature. This thesis focuses on potential performance enhancement brought by NOMA for the C-RAN, H-CRAN and F-RAN and is expected to address some of the aforementioned key challenges of 5G. The research work of this thesis can be divided into three parts. In the first part of our research, we focus on investigating the performance analysis of NOMA in a C-RAN. The problem of jointly optimizing user association, muting and power-bandwidth allocation is formulated for NOMA-enabled C-RANs. To solve the mixed integer programming problem, the joint problem is decomposed into two subproblems as 1) user association and muting 2) power-bandwidth allocation optimization. To deal with the first subproblem, we propose a centralized and heuristic algorithm to provide the optimal and suboptimal solutions to the remote radio head (RRH) muting problem for given bandwidth and transmit power, respectively. The second subproblem is then reformulated and we propose an optimal solution to bandwidth and power allocation subject to users data rate constraints. Moreover, for given user association and muting states, the optimal power allocation is derived in a closed-form. Simulation results show that the proposed NOMA-enabled C-RAN outperforms orthogonal multiple access (OMA)-based C-RANs in terms of total achievable rate, interference mitigation and can achieve significant fairness improvement. Our second work investigates the performance of NOMA in H-CRAN, where coordination of macro base station (MBS) and remote radio heads (RRHs) for H-CRAN with NOMA is introduced to improve network performance. We formulate the problem of jointly optimizing user association, coordinated scheduling and power allocation for NOMA-enabled H-CRANs. To efficiently solve this problem, we decompose the joint optimization problem into two subproblems as 1) user association and scheduling 2) power allocation optimization. Firstly the users are divided based on different interference they suffer. This interference-aware NOMA approach account for the inter-tier interference. Proportional fairness (PF) scheduling for NOMA is utilized to schedule users with a two-loop optimization method to enhance throughput and fairness. Based on the user scheduling scheme, optimal power allocation optimization is performed by the hierarchical decomposition approach. It is then followed by algorithm for joint scheduling and power allocation. Simulation results show that the proposed NOMA-enabled H-CRAN outperforms OMA-based H-CRANs in terms of total achievable rate and can achieve significant fairness improvement. In the third part of our research, we propose a NOMA-enabled fog-cloud structure in a novel density-aware F-RAN to tackle different aspects such as throughput and latency requirements of high and low user-density regions, in order to meet the heterogeneous requirements of eMBB and URLLC traffic. A framework of the multi-objective problem is formulated to cater the high throughput and low-latency requirements in a high and low user-density mode respectively. In the first problem, we study the joint caching placement and association strategy aiming at minimizing the average delay. To deal with the first problem, we apply McCormick envelopes and Lagrange partial relaxation method to transform it into three convex sub-problems, which is then solved by proposed distributed algorithm. The second problem is to jointly optimize transmission mode selection, subchannel assignment and power allocation to maximize the sum data rate of all fog user equipments (F-UEs) while satisfying fronthaul capacity and fog-computing access point (F-AP) power constraints. Moreover, for given transmission mode selection and subchannel assignment, the optimal power allocation is derived in a closed-form. Simulation results are provided for the proposed NOMA-enabled F-RAN framework and reveal that the ultra-low latency and high throughput can be achieved by properly utilizing the available resources
    • …
    corecore