14,544 research outputs found
Multiuser MIMO-OFDM for Next-Generation Wireless Systems
This overview portrays the 40-year evolution of orthogonal frequency division multiplexing (OFDM) research. The amelioration of powerful multicarrier OFDM arrangements with multiple-input multiple-output (MIMO) systems has numerous benefits, which are detailed in this treatise. We continue by highlighting the limitations of conventional detection and channel estimation techniques designed for multiuser MIMO OFDM systems in the so-called rank-deficient scenarios, where the number of users supported or the number of transmit antennas employed exceeds the number of receiver antennas. This is often encountered in practice, unless we limit the number of users granted access in the base station’s or radio port’s coverage area. Following a historical perspective on the associated design problems and their state-of-the-art solutions, the second half of this treatise details a range of classic multiuser detectors (MUDs) designed for MIMO-OFDM systems and characterizes their achievable performance. A further section aims for identifying novel cutting-edge genetic algorithm (GA)-aided detector solutions, which have found numerous applications in wireless communications in recent years. In an effort to stimulate the cross pollination of ideas across the machine learning, optimization, signal processing, and wireless communications research communities, we will review the broadly applicable principles of various GA-assisted optimization techniques, which were recently proposed also for employment inmultiuser MIMO OFDM. In order to stimulate new research, we demonstrate that the family of GA-aided MUDs is capable of achieving a near-optimum performance at the cost of a significantly lower computational complexity than that imposed by their optimum maximum-likelihood (ML) MUD aided counterparts. The paper is concluded by outlining a range of future research options that may find their way into next-generation wireless systems
Massive MIMO for Internet of Things (IoT) Connectivity
Massive MIMO is considered to be one of the key technologies in the emerging
5G systems, but also a concept applicable to other wireless systems. Exploiting
the large number of degrees of freedom (DoFs) of massive MIMO essential for
achieving high spectral efficiency, high data rates and extreme spatial
multiplexing of densely distributed users. On the one hand, the benefits of
applying massive MIMO for broadband communication are well known and there has
been a large body of research on designing communication schemes to support
high rates. On the other hand, using massive MIMO for Internet-of-Things (IoT)
is still a developing topic, as IoT connectivity has requirements and
constraints that are significantly different from the broadband connections. In
this paper we investigate the applicability of massive MIMO to IoT
connectivity. Specifically, we treat the two generic types of IoT connections
envisioned in 5G: massive machine-type communication (mMTC) and ultra-reliable
low-latency communication (URLLC). This paper fills this important gap by
identifying the opportunities and challenges in exploiting massive MIMO for IoT
connectivity. We provide insights into the trade-offs that emerge when massive
MIMO is applied to mMTC or URLLC and present a number of suitable communication
schemes. The discussion continues to the questions of network slicing of the
wireless resources and the use of massive MIMO to simultaneously support IoT
connections with very heterogeneous requirements. The main conclusion is that
massive MIMO can bring benefits to the scenarios with IoT connectivity, but it
requires tight integration of the physical-layer techniques with the protocol
design.Comment: Submitted for publicatio
A cross-layer approach to enhance QoS for multimedia applications over satellite
The need for on-demand QoS support for communications over satellite is of primary importance for distributed multimedia applications. This is particularly true for the return link which is often a bottleneck due to the large set of end-users accessing a very limited uplink resource. Facing this need, Demand Assignment Multiple Access (DAMA) is a classical technique that allows satellite operators to offer various types of services, while managing the resources of the satellite system efficiently. Tackling the quality degradation and delay accumulation issues that can result from the use of these techniques, this paper proposes an instantiation of the Application Layer Framing (ALF) approach, using a cross-layer interpreter(xQoS-Interpreter). The information provided by this interpreter is used to manage the resource provided to a terminal by the satellite system in order to improve the quality of multimedia presentations from the end users point of view. Several experiments are carried out for different loads on the return link. Their impact on QoS is measured through different application as well as network level metrics
Enabling Disaster Resilient 4G Mobile Communication Networks
The 4G Long Term Evolution (LTE) is the cellular technology expected to
outperform the previous generations and to some extent revolutionize the
experience of the users by taking advantage of the most advanced radio access
techniques (i.e. OFDMA, SC-FDMA, MIMO). However, the strong dependencies
between user equipments (UEs), base stations (eNBs) and the Evolved Packet Core
(EPC) limit the flexibility, manageability and resiliency in such networks. In
case the communication links between UEs-eNB or eNB-EPC are disrupted, UEs are
in fact unable to communicate. In this article, we reshape the 4G mobile
network to move towards more virtual and distributed architectures for
improving disaster resilience, drastically reducing the dependency between UEs,
eNBs and EPC. The contribution of this work is twofold. We firstly present the
Flexible Management Entity (FME), a distributed entity which leverages on
virtualized EPC functionalities in 4G cellular systems. Second, we introduce a
simple and novel device-todevice (D2D) communication scheme allowing the UEs in
physical proximity to communicate directly without resorting to the
coordination with an eNB.Comment: Submitted to IEEE Communications Magazin
Massive MIMO is a Reality -- What is Next? Five Promising Research Directions for Antenna Arrays
Massive MIMO (multiple-input multiple-output) is no longer a "wild" or
"promising" concept for future cellular networks - in 2018 it became a reality.
Base stations (BSs) with 64 fully digital transceiver chains were commercially
deployed in several countries, the key ingredients of Massive MIMO have made it
into the 5G standard, the signal processing methods required to achieve
unprecedented spectral efficiency have been developed, and the limitation due
to pilot contamination has been resolved. Even the development of fully digital
Massive MIMO arrays for mmWave frequencies - once viewed prohibitively
complicated and costly - is well underway. In a few years, Massive MIMO with
fully digital transceivers will be a mainstream feature at both sub-6 GHz and
mmWave frequencies. In this paper, we explain how the first chapter of the
Massive MIMO research saga has come to an end, while the story has just begun.
The coming wide-scale deployment of BSs with massive antenna arrays opens the
door to a brand new world where spatial processing capabilities are
omnipresent. In addition to mobile broadband services, the antennas can be used
for other communication applications, such as low-power machine-type or
ultra-reliable communications, as well as non-communication applications such
as radar, sensing and positioning. We outline five new Massive MIMO related
research directions: Extremely large aperture arrays, Holographic Massive MIMO,
Six-dimensional positioning, Large-scale MIMO radar, and Intelligent Massive
MIMO.Comment: 20 pages, 9 figures, submitted to Digital Signal Processin
Optimal Control of Wireless Computing Networks
Augmented information (AgI) services allow users to consume information that
results from the execution of a chain of service functions that process source
information to create real-time augmented value. Applications include real-time
analysis of remote sensing data, real-time computer vision, personalized video
streaming, and augmented reality, among others. We consider the problem of
optimal distribution of AgI services over a wireless computing network, in
which nodes are equipped with both communication and computing resources. We
characterize the wireless computing network capacity region and design a joint
flow scheduling and resource allocation algorithm that stabilizes the
underlying queuing system while achieving a network cost arbitrarily close to
the minimum, with a tradeoff in network delay. Our solution captures the unique
chaining and flow scaling aspects of AgI services, while exploiting the use of
the broadcast approach coding scheme over the wireless channel.Comment: 30 pages, journa
- …