1,134 research outputs found
An Overview on Application of Machine Learning Techniques in Optical Networks
Today's telecommunication networks have become sources of enormous amounts of
widely heterogeneous data. This information can be retrieved from network
traffic traces, network alarms, signal quality indicators, users' behavioral
data, etc. Advanced mathematical tools are required to extract meaningful
information from these data and take decisions pertaining to the proper
functioning of the networks from the network-generated data. Among these
mathematical tools, Machine Learning (ML) is regarded as one of the most
promising methodological approaches to perform network-data analysis and enable
automated network self-configuration and fault management. The adoption of ML
techniques in the field of optical communication networks is motivated by the
unprecedented growth of network complexity faced by optical networks in the
last few years. Such complexity increase is due to the introduction of a huge
number of adjustable and interdependent system parameters (e.g., routing
configurations, modulation format, symbol rate, coding schemes, etc.) that are
enabled by the usage of coherent transmission/reception technologies, advanced
digital signal processing and compensation of nonlinear effects in optical
fiber propagation. In this paper we provide an overview of the application of
ML to optical communications and networking. We classify and survey relevant
literature dealing with the topic, and we also provide an introductory tutorial
on ML for researchers and practitioners interested in this field. Although a
good number of research papers have recently appeared, the application of ML to
optical networks is still in its infancy: to stimulate further work in this
area, we conclude the paper proposing new possible research directions
Thirty Years of Machine Learning: The Road to Pareto-Optimal Wireless Networks
Future wireless networks have a substantial potential in terms of supporting
a broad range of complex compelling applications both in military and civilian
fields, where the users are able to enjoy high-rate, low-latency, low-cost and
reliable information services. Achieving this ambitious goal requires new radio
techniques for adaptive learning and intelligent decision making because of the
complex heterogeneous nature of the network structures and wireless services.
Machine learning (ML) algorithms have great success in supporting big data
analytics, efficient parameter estimation and interactive decision making.
Hence, in this article, we review the thirty-year history of ML by elaborating
on supervised learning, unsupervised learning, reinforcement learning and deep
learning. Furthermore, we investigate their employment in the compelling
applications of wireless networks, including heterogeneous networks (HetNets),
cognitive radios (CR), Internet of things (IoT), machine to machine networks
(M2M), and so on. This article aims for assisting the readers in clarifying the
motivation and methodology of the various ML algorithms, so as to invoke them
for hitherto unexplored services as well as scenarios of future wireless
networks.Comment: 46 pages, 22 fig
Recommended from our members
Resource sharing in network slicing and human-machine interactions
In this thesis we explore two novel resource allocation models. The first addresses challenges associated with dynamic sharing of network resources by multiple tenants/services via network slicing. The second focuses on a data-driven approach to the optimization of resource allocation in interactive human-machine processes. In our first thrust we investigate how to allocate shared storage, computation, and/or connectivity resources distributed amongst multiple tenants/ virtual service providers which have dynamic loads. It is expected that next generation of wireless network will be shared by an increasing number of data-intensive mobile applications (e.g., autonomous cars, IoT, interactive 360° video streaming), and tenants/service providers. A key functional requirement for such infrastructure is enabling efficient sharing of heterogeneous resource among tenants/service providers supporting spatially varying and dynamic user demands, both from the point of view of enabling the deployment and performance management to diverse service providers and/or tenants, as well as means to increase utilization and reduce CAPEX/OPEX associated with deploying possible new infrastructures. To that end, we propose a novel dynamic resource sharing policy, namely, Share Constrained Proportional Fair (SCPF), which allocates a predefined ‘share’ of a pool of (distributed) resources to each slice. We provide a characterization of the achievable performance gains over General Processor Sharing (GPS), and Static Slicing (SS), i.e., fixed allocation of resources to slices. We also characterize the associated share dimensioning problem, asking when a particular set of load profiles and QoS requirements are feasible, as well as what should be an appropriate pricing strategy. We further consider possible slice-based admission control scheme where slices engage in an underlying game to maximize their carried loads subject to performance requirements. In order to accommodate settings where one would wish to provision different types of resources which are coupled through user demands, we generalize SCPF to a more general resource allocation criterion, namely, Share Constrained Slicing (SCS), which extends traditional α—fairness criterion, by striking a balance among inter- and intra-slice fairness vs. overall efficiency. We show that SCS has several desirable properties including slice-level protection, envyfreeness, and load-driven elasticity. In practice, mobile users' dynamics could make the cost of implementing SCS high, so we also study the feasibility of using a dynamically weighted max-min fair policy as a surrogate resource allocation scheme. For a setting with stochastic loads and elastic user requirements, we model the user dynamics under SCS as a queuing network and establish the stability condition. Finally, and perhaps surprisingly, we show via extensive simulation that while SCS (and/or the surrogate weighted max-min allocation) provides inter-slice protection, they can also achieve improved job delay and/or perceived throughput, as compared to other weighted max-min based allocation schemes whose intra-slice weight allocation is not share-constrained, e.g., traditional max-min and/or discriminatory processor sharing. In our second thrust we study how to optimize resource allocation in the context of human-machine interactions. Examples of such processes could include systems aimed at assisting humans in interactive learning, workload allocation, or web-search advertising. We devise an innovative framework to enable the optimization of a reward over an interactive process in a data-driven manner. This is a challenging problem for several reasons: (1) humans' behavior is not easily modeled and may reflect biases, memory and be sensitive to sequencing, all of which should/could be inferred from data; (2) because these interactions are typically sequential and transient, inferring such complex models for human behavior is difficult; (3) furthermore, in order to collect data on human-machine interactions one must choose a machine policy which in turn may bias inferences on human behavior. In this thesis we approach the problem of jointly estimating human behavior and optimizing machine policies via Alternating Entropy-Reward Ascent (AREA) algorithm. We characterize AREA in terms of its space and time complexity and convergence. We also provide an initial validation based on synthetic data generated by an established noisy nonlinear model for human decision-makingElectrical and Computer Engineerin
Artificial intelligence (AI) methods in optical networks: A comprehensive survey
Producción CientíficaArtificial intelligence (AI) is an extensive scientific discipline which enables computer systems to solve problems by emulating complex biological processes such as learning, reasoning and self-correction. This paper presents a comprehensive review of the application of AI techniques for improving performance of optical communication systems and networks. The use of AI-based techniques is first studied in applications related to optical transmission, ranging from the characterization and operation of network components to performance monitoring, mitigation of nonlinearities, and quality of transmission estimation. Then, applications related to optical network control and management are also reviewed, including topics like optical network planning and operation in both transport and access networks. Finally, the paper also presents a summary of opportunities and challenges in optical networking where AI is expected to play a key role in the near future.Ministerio de Economía, Industria y Competitividad (Project EC2014-53071-C3-2-P, TEC2015-71932-REDT
Resource management in QoS-aware wireless cellular networks
2011 Summer.Includes bibliographical references.Emerging broadband wireless networks that support high speed packet data with heterogeneous quality of service (QoS) requirements demand more flexible and efficient use of the scarce spectral resource. Opportunistic scheduling exploits the time-varying, location-dependent channel conditions to achieve multiuser diversity. In this work, we study two types of resource allocation problems in QoS-aware wireless cellular networks. First, we develop a rigorous framework to study opportunistic scheduling in multiuser OFDM systems. We derive optimal opportunistic scheduling policies under three common QoS/fairness constraints for multiuser OFDM systems--temporal fairness, utilitarian fairness, and minimum-performance guarantees. To implement these optimal policies efficiently, we provide a modified Hungarian algorithm and a simple suboptimal algorithm. We then propose a generalized opportunistic scheduling framework that incorporates multiple mixed QoS/fairness constraints, including providing both lower and upper bound constraints. Next, taking input queues and channel memory into consideration, we reformulate the transmission scheduling problem as a new class of Markov decision processes (MDPs) with fairness constraints. We investigate the throughput maximization and the delay minimization problems in this context. We study two categories of fairness constraints, namely temporal fairness and utilitarian fairness. We consider two criteria: infinite horizon expected total discounted reward and expected average reward. We derive and prove explicit dynamic programming equations for the above constrained MDPs, and characterize optimal scheduling policies based on those equations. An attractive feature of our proposed schemes is that they can easily be extended to fit different objective functions and other fairness measures. Although we only focus on uplink scheduling, the scheme is equally applicable to the downlink case. Furthermore, we develop an efficient approximation method--temporal fair rollout--to reduce the computational cost
A machine learning approach to 5G infrastructure market optimization
It is now commonly agreed that future 5G Networks will build upon the network slicing concept. The ability to provide virtual, logically independent "slices" of the network will also have an impact on the models that will sustain the business ecosystem. Network slicing will open the door to new players: the infrastructure provider, which is the owner of the infrastructure, and the tenants, which may acquire a network slice from the infrastructure provider to deliver a specific service to their customers. In this new context, how to correctly handle resource allocation among tenants and how to maximize the monetization of the infrastructure become fundamental problems that need to be solved. In this paper, we address this issue by designing a network slice admission control algorithm that (i) autonomously learns the best acceptance policy while (ii) it ensures that the service guarantees provided to tenants are always satisfied. The contributions of this paper include: (i) an analytical model for the admissibility region of a network slicing-capable 5G Network, (ii) the analysis of the system (modeled as a Semi-Markov Decision Process) and the optimization of the infrastructure providers revenue, and (iii) the design of a machine learning algorithm that can be deployed in practical settings and achieves close to optimal performance.The work of University Carlos III of Madrid was supported by the H2020 5G-MoNArch project (Grant Agreement No. 761445) and the 5GCity project of the Spanish Ministry of Economy and Competitiveness (TEC2016-76795-C6-3-R). The work of NEC Laboratories Europe was supported by the 5G-Transformer project (Grant Agreement No. 761536)
- …