1,369 research outputs found
Thirty Years of Machine Learning: The Road to Pareto-Optimal Wireless Networks
Future wireless networks have a substantial potential in terms of supporting
a broad range of complex compelling applications both in military and civilian
fields, where the users are able to enjoy high-rate, low-latency, low-cost and
reliable information services. Achieving this ambitious goal requires new radio
techniques for adaptive learning and intelligent decision making because of the
complex heterogeneous nature of the network structures and wireless services.
Machine learning (ML) algorithms have great success in supporting big data
analytics, efficient parameter estimation and interactive decision making.
Hence, in this article, we review the thirty-year history of ML by elaborating
on supervised learning, unsupervised learning, reinforcement learning and deep
learning. Furthermore, we investigate their employment in the compelling
applications of wireless networks, including heterogeneous networks (HetNets),
cognitive radios (CR), Internet of things (IoT), machine to machine networks
(M2M), and so on. This article aims for assisting the readers in clarifying the
motivation and methodology of the various ML algorithms, so as to invoke them
for hitherto unexplored services as well as scenarios of future wireless
networks.Comment: 46 pages, 22 fig
Edge computing infrastructure for 5G networks: a placement optimization solution
This thesis focuses on how to optimize the placement of the Edge Computing infrastructure for upcoming 5G networks. To this aim, the core contributions of this research are twofold: 1) a novel heuristic called Hybrid Simulated Annealing to tackle the NP-hard nature of the problem and, 2) a framework called EdgeON providing a practical tool for real-life deployment optimization.
In more detail, Edge Computing has grown into a key solution to 5G latency, reliability and scalability requirements. By bringing computing, storage and networking resources to the edge of the network, delay-sensitive applications, location-aware systems and upcoming real-time services leverage the benefits of a reduced physical and logical path between the end-user and the data or service host.
Nevertheless, the edge node placement problem raises critical concerns regarding deployment and operational expenditures (i.e., mainly due to the number of nodes to be deployed), current backhaul network capabilities and non-technical placement limitations. Common approaches to the placement of edge nodes are based on: Mobile Edge Computing (MEC), where the processing capabilities are deployed at the Radio Access Network nodes and Facility Location Problem variations, where a simplistic cost function is used to determine where to optimally place the infrastructure. However, these methods typically lack the flexibility to be used for edge node placement under the strict technical requirements identified for 5G networks. They fail to place resources at the network edge for 5G ultra-dense networking environments in a network-aware manner.
This doctoral thesis focuses on rigorously defining the Edge Node Placement Problem (ENPP) for 5G use cases and proposes a novel framework called EdgeON aiming at reducing the overall expenses when deploying and operating an Edge Computing network, taking into account the usage and characteristics of the in-place backhaul network and the strict requirements of a 5G-EC ecosystem. The developed framework implements several placement and optimization strategies thoroughly assessing its suitability to solve the network-aware ENPP. The core of the framework is an in-house developed heuristic called Hybrid Simulated Annealing (HSA), seeking to address the high complexity of the ENPP while avoiding the non-convergent behavior of other traditional heuristics (i.e., when applied to similar problems).
The findings of this work validate our approach to solve the network-aware ENPP, the effectiveness of the heuristic proposed and the overall applicability of EdgeON. Thorough performance evaluations were conducted on the core placement solutions implemented revealing the superiority of HSA when compared to widely used heuristics and common edge placement approaches (i.e., a MEC-based strategy). Furthermore, the practicality of EdgeON was tested through two main case studies placing services and virtual network functions over the previously optimally placed edge nodes.
Overall, our proposal is an easy-to-use, effective and fully extensible tool that can be used by operators seeking to optimize the placement of computing, storage and networking infrastructure at the users’ vicinity. Therefore, our main contributions not only set strong foundations towards a cost-effective deployment and operation of an Edge Computing network, but directly impact the feasibility of upcoming 5G services/use cases and the extensive existing research regarding the placement of services and even network service chains at the edge
TinyML: Tools, Applications, Challenges, and Future Research Directions
In recent years, Artificial Intelligence (AI) and Machine learning (ML) have
gained significant interest from both, industry and academia. Notably,
conventional ML techniques require enormous amounts of power to meet the
desired accuracy, which has limited their use mainly to high-capability devices
such as network nodes. However, with many advancements in technologies such as
the Internet of Things (IoT) and edge computing, it is desirable to incorporate
ML techniques into resource-constrained embedded devices for distributed and
ubiquitous intelligence. This has motivated the emergence of the TinyML
paradigm which is an embedded ML technique that enables ML applications on
multiple cheap, resource- and power-constrained devices. However, during this
transition towards appropriate implementation of the TinyML technology,
multiple challenges such as processing capacity optimization, improved
reliability, and maintenance of learning models' accuracy require timely
solutions. In this article, various avenues available for TinyML implementation
are reviewed. Firstly, a background of TinyML is provided, followed by detailed
discussions on various tools supporting TinyML. Then, state-of-art applications
of TinyML using advanced technologies are detailed. Lastly, various research
challenges and future directions are identified.Comment: 12 pags, 3 tables, 4 figure
An Energy-driven Network Function Virtualization for Multi-domain Software Defined Networks
Network Functions Virtualization (NFV) in Software Defined Networks (SDN)
emerged as a new technology for creating virtual instances for smooth execution
of multiple applications. Their amalgamation provides flexible and programmable
platforms to utilize the network resources for providing Quality of Service
(QoS) to various applications. In SDN-enabled NFV setups, the underlying
network services can be viewed as a series of virtual network functions (VNFs)
and their optimal deployment on physical/virtual nodes is considered a
challenging task to perform. However, SDNs have evolved from single-domain to
multi-domain setups in the recent era. Thus, the complexity of the underlying
VNF deployment problem in multi-domain setups has increased manifold. Moreover,
the energy utilization aspect is relatively unexplored with respect to an
optimal mapping of VNFs across multiple SDN domains. Hence, in this work, the
VNF deployment problem in multi-domain SDN setup has been addressed with a
primary emphasis on reducing the overall energy consumption for deploying the
maximum number of VNFs with guaranteed QoS. The problem in hand is initially
formulated as a "Multi-objective Optimization Problem" based on Integer Linear
Programming (ILP) to obtain an optimal solution. However, the formulated ILP
becomes complex to solve with an increasing number of decision variables and
constraints with an increase in the size of the network. Thus, we leverage the
benefits of the popular evolutionary optimization algorithms to solve the
problem under consideration. In order to deduce the most appropriate
evolutionary optimization algorithm to solve the considered problem, it is
subjected to different variants of evolutionary algorithms on the widely used
MOEA framework (an open source java framework based on multi-objective
evolutionary algorithms).Comment: Accepted for publication in IEEE INFOCOM 2019 Workshop on Intelligent
Cloud Computing and Networking (ICCN 2019
Data semantic enrichment for complex event processing over IoT Data Streams
This thesis generalizes techniques for processing IoT data streams, semantically enrich data with contextual information, as well as complex event processing in IoT applications. A case study for ECG anomaly detection and signal classification was conducted to validate the knowledge foundation
- …