10 research outputs found
Gestion flexible des ressources dans les réseaux de nouvelle génération avec SDN
Abstract : 5G and beyond-5G/6G are expected to shape the future economic growth of multiple vertical industries by providing the network infrastructure required to enable innovation and new business models. They have the potential to offer a wide spectrum of services, namely higher data rates, ultra-low latency, and high reliability. To achieve their promises, 5G and beyond-5G/6G rely on software-defined networking (SDN), edge computing, and radio access network (RAN) slicing technologies. In this thesis, we aim to use SDN as a key enabler to enhance resource management in next-generation networks. SDN allows programmable management of edge computing resources and dynamic orchestration of RAN slicing. However, achieving efficient performance based on SDN capabilities is a challenging task due to the permanent fluctuations of traffic in next-generation networks and the diversified quality of service requirements of emerging applications. Toward our objective, we address the load balancing problem in distributed SDN architectures, and we optimize the RAN slicing of communication and computation resources in the edge of the network. In the first part of this thesis, we present a proactive approach to balance the load in a distributed SDN control plane using the data plane component migration mechanism. First, we propose prediction models that forecast the load of SDN controllers in the long term. By using these models, we can preemptively detect whether the load will be unbalanced in the control plane and, thus, schedule migration operations in advance. Second, we improve the migration operation performance by optimizing the tradeoff between a load balancing factor and the cost of migration operations. This proactive load balancing approach not only avoids SDN controllers from being overloaded, but also allows a judicious selection of which data plane component should be migrated and where the migration should happen. In the second part of this thesis, we propose two RAN slicing schemes that efficiently allocate the communication and the computation resources in the edge of the network. The first RAN slicing scheme performs the allocation of radio resource blocks (RBs) to end-users in two time-scales, namely in a large time-scale and in a small time-scale. In the large time-scale, an SDN controller allocates to each base station a number of RBs from a shared radio RBs pool, according to its requirements in terms of delay and data rate. In the short time-scale, each base station assigns its available resources to its end-users and requests, if needed, additional resources from adjacent base stations. The second RAN slicing scheme jointly allocates the RBs and computation resources available in edge computing servers based on an open RAN architecture. We develop, for the proposed RAN slicing schemes, reinforcement learning and deep reinforcement learning algorithms to dynamically allocate RAN resources.La 5G et au-delà de la 5G/6G sont censées dessiner la future croissance économique de multiples industries verticales en fournissant l'infrastructure réseau nécessaire pour permettre l'innovation et la création de nouveaux modèles économiques. Elles permettent d'offrir un large spectre de services, à savoir des débits de données plus élevés, une latence ultra-faible et une fiabilité élevée. Pour tenir leurs promesses, la 5G et au-delà de la-5G/6G s'appuient sur le réseau défini par logiciel (SDN), l’informatique en périphérie et le découpage du réseau d'accès (RAN). Dans cette thèse, nous visons à utiliser le SDN en tant qu'outil clé pour améliorer la gestion des ressources dans les réseaux de nouvelle génération. Le SDN permet une gestion programmable des ressources informatiques en périphérie et une orchestration dynamique de découpage du RAN. Cependant, atteindre une performance efficace en se basant sur le SDN est une tâche difficile due aux fluctuations permanentes du trafic dans les réseaux de nouvelle génération et aux exigences de qualité de service diversifiées des applications émergentes. Pour atteindre notre objectif, nous abordons le problème de l'équilibrage de charge dans les architectures SDN distribuées, et nous optimisons le découpage du RAN des ressources de communication et de calcul à la périphérie du réseau. Dans la première partie de cette thèse, nous présentons une approche proactive pour équilibrer la charge dans un plan de contrôle SDN distribué en utilisant le mécanisme de migration des composants du plan de données. Tout d'abord, nous proposons des modèles pour prédire la charge des contrôleurs SDN à long terme. En utilisant ces modèles, nous pouvons détecter de manière préemptive si la charge sera déséquilibrée dans le plan de contrôle et, ainsi, programmer des opérations de migration à l'avance. Ensuite, nous améliorons les performances des opérations de migration en optimisant le compromis entre un facteur d'équilibrage de charge et le coût des opérations de migration. Cette approche proactive d'équilibrage de charge permet non seulement d'éviter la surcharge des contrôleurs SDN, mais aussi de choisir judicieusement le composant du plan de données à migrer et l'endroit où la migration devrait avoir lieu. Dans la deuxième partie de cette thèse, nous proposons deux mécanismes de découpage du RAN qui allouent efficacement les ressources de communication et de calcul à la périphérie des réseaux. Le premier mécanisme de découpage du RAN effectue l'allocation des blocs de ressources radio (RBs) aux utilisateurs finaux en deux échelles de temps, à savoir dans une échelle de temps large et dans une échelle de temps courte. Dans l’échelle de temps large, un contrôleur SDN attribue à chaque station de base un certain nombre de RB à partir d'un pool de RB radio partagé, en fonction de ses besoins en termes de délai et de débit. Dans l’échelle de temps courte, chaque station de base attribue ses ressources disponibles à ses utilisateurs finaux et demande, si nécessaire, des ressources supplémentaires aux stations de base adjacentes. Le deuxième mécanisme de découpage du RAN alloue conjointement les RB et les ressources de calcul disponibles dans les serveurs de l’informatique en périphérie en se basant sur une architecture RAN ouverte. Nous développons, pour les mécanismes de découpage du RAN proposés, des algorithmes d'apprentissage par renforcement et d'apprentissage par renforcement profond pour allouer dynamiquement les ressources du RAN
Multi-access edge computing: A survey
Multi-access Edge Computing (MEC) is a key solution that enables operators to open their networks to new services and IT ecosystems to leverage edge-cloud benefits in their networks and systems. Located in close proximity from the end users and connected devices, MEC provides extremely low latency and high bandwidth while always enabling applications to leverage cloud capabilities as necessary. In this article, we illustrate the integration of MEC into a current mobile networks' architecture as well as the transition mechanisms to migrate into a standard 5G network architecture.We also discuss SDN, NFV, SFC and network slicing as MEC enablers. Then, we provide a state-of-the-art study on the different approaches that optimize the MEC resources and its QoS parameters. In this regard, we classify these approaches based on the optimized resources and QoS parameters (i.e., processing, storage, memory, bandwidth, energy and latency). Finally, we propose an architectural framework for a MEC-NFV environment based on the standard SDN architecture
Communication and Computation O-RAN Resource Slicing for URLLC Services Using Deep Reinforcement Learning
The evolution of the future beyond-5G/6G networks towards a service-aware
network is based on network slicing technology. With network slicing,
communication service providers seek to meet all the requirements imposed by
the verticals, including ultra-reliable low-latency communication (URLLC)
services. In addition, the open radio access network (O-RAN) architecture paves
the way for flexible sharing of network resources by introducing more
programmability into the RAN. RAN slicing is an essential part of end-to-end
network slicing since it ensures efficient sharing of communication and
computation resources. However, due to the stringent requirements of URLLC
services and the dynamics of the RAN environment, RAN slicing is challenging.
In this article, we propose a two-level RAN slicing approach based on the O-RAN
architecture to allocate the communication and computation RAN resources among
URLLC end-devices. For each RAN slicing level, we model the resource slicing
problem as a single-agent Markov decision process and design a deep
reinforcement learning algorithm to solve it. Simulation results demonstrate
the efficiency of the proposed approach in meeting the desired quality of
service requirements
Dynamic SDN-based radio access network slicing with deep reinforcement learning for URLLC and eMBB services
Radio access network (RAN) slicing is a key technology that enables 5G
network to support heterogeneous requirements of generic services, namely
ultra-reliable low-latency communication (URLLC) and enhanced mobile broadband
(eMBB). In this paper, we propose a two time-scales RAN slicing mechanism to
optimize the performance of URLLC and eMBB services. In a large time-scale, an
SDN controller allocates radio resources to gNodeBs according to the
requirements of the eMBB and URLLC services. In a short time-scale, each gNodeB
allocates its available resources to its end-users and requests, if needed,
additional resources from adjacent gNodeBs. We formulate this problem as a
non-linear binary program and prove its NP-hardness. Next, for each time-scale,
we model the problem as a Markov decision process (MDP), where the large-time
scale is modeled as a single agent MDP whereas the shorter time-scale is
modeled as a multi-agent MDP. We leverage the exponential-weight algorithm for
exploration and exploitation (EXP3) to solve the single-agent MDP of the large
time-scale MDP and the multi-agent deep Q-learning (DQL) algorithm to solve the
multi-agent MDP of the short time-scale resource allocation. Extensive
simulations show that our approach is efficient under different network
parameters configuration and it outperforms recent benchmark solutions
A Deep Reinforcement Learning Approach for Service Migration in MEC-enabled Vehicular Networks
Multi-access edge computing (MEC) is a key enabler to reduce the latency of
vehicular network. Due to the vehicles mobility, their requested services
(e.g., infotainment services) should frequently be migrated across different
MEC servers to guarantee their stringent quality of service requirements. In
this paper, we study the problem of service migration in a MEC-enabled
vehicular network in order to minimize the total service latency and migration
cost. This problem is formulated as a nonlinear integer program and is
linearized to help obtaining the optimal solution using off-the-shelf solvers.
Then, to obtain an efficient solution, it is modeled as a multi-agent Markov
decision process and solved by leveraging deep Q learning (DQL) algorithm. The
proposed DQL scheme performs a proactive services migration while ensuring
their continuity under high mobility constraints. Finally, simulations results
show that the proposed DQL scheme achieves close-to-optimal performance