1,462 research outputs found
End-to-End Simulation of 5G mmWave Networks
Due to its potential for multi-gigabit and low latency wireless links,
millimeter wave (mmWave) technology is expected to play a central role in 5th
generation cellular systems. While there has been considerable progress in
understanding the mmWave physical layer, innovations will be required at all
layers of the protocol stack, in both the access and the core network.
Discrete-event network simulation is essential for end-to-end, cross-layer
research and development. This paper provides a tutorial on a recently
developed full-stack mmWave module integrated into the widely used open-source
ns--3 simulator. The module includes a number of detailed statistical channel
models as well as the ability to incorporate real measurements or ray-tracing
data. The Physical (PHY) and Medium Access Control (MAC) layers are modular and
highly customizable, making it easy to integrate algorithms or compare
Orthogonal Frequency Division Multiplexing (OFDM) numerologies, for example.
The module is interfaced with the core network of the ns--3 Long Term Evolution
(LTE) module for full-stack simulations of end-to-end connectivity, and
advanced architectural features, such as dual-connectivity, are also available.
To facilitate the understanding of the module, and verify its correct
functioning, we provide several examples that show the performance of the
custom mmWave stack as well as custom congestion control algorithms designed
specifically for efficient utilization of the mmWave channel.Comment: 25 pages, 16 figures, submitted to IEEE Communications Surveys and
Tutorials (revised Jan. 2018
X-TCP: A Cross Layer Approach for TCP Uplink Flows in mmWave Networks
Millimeter wave frequencies will likely be part of the fifth generation of
mobile networks and of the 3GPP New Radio (NR) standard. MmWave communication
indeed provides a very large bandwidth, thus an increased cell throughput, but
how to exploit these resources at the higher layers is still an open research
question. A very relevant issue is the high variability of the channel, caused
by the blockage from obstacles and the human body. This affects the design of
congestion control mechanisms at the transport layer, and state-of-the-art TCP
schemes such as TCP CUBIC present suboptimal performance. In this paper, we
present a cross layer approach for uplink flows that adjusts the congestion
window of TCP at the mobile equipment side using an estimation of the available
data rate at the mmWave physical layer, based on the actual resource allocation
and on the Signal to Interference plus Noise Ratio. We show that this approach
reduces the latency, avoiding to fill the buffers in the cellular stack, and
has a quicker recovery time after RTO events than several other TCP congestion
control algorithms.Comment: 6 pages, 5 figures, accepted for presentation at the 2017 16th Annual
Mediterranean Ad Hoc Networking Workshop (MED-HOC-NET
Will TCP work in mmWave 5G Cellular Networks?
The vast available spectrum in the millimeter wave (mmWave) bands offers the
possibility of multi-Gbps data rates for fifth generation (5G) cellular
networks. However, mmWave capacity can be highly intermittent due to the
vulnerability of mmWave signals to blockages and delays in directional
searching. Such highly variable links present unique challenges for adaptive
control mechanisms in transport layer protocols and end-to-end applications.
This paper considers the fundamental question of whether TCP - the most widely
used transport protocol - will work in mmWave cellular systems. The paper
provides a comprehensive simulation study of TCP considering various factors
such as the congestion control algorithm, including the recently proposed TCP
BBR, edge vs. remote servers, handover and multi- connectivity, TCP packet size
and 3GPP-stack parameters. We show that the performance of TCP on mmWave links
is highly dependent on different combinations of these parameters, and identify
the open challenges in this area.Comment: 7 pages, 4 figures, 2 tables. To be published in the IEEE
Communication Magazin
A Survey on Wireless Security: Technical Challenges, Recent Advances and Future Trends
This paper examines the security vulnerabilities and threats imposed by the
inherent open nature of wireless communications and to devise efficient defense
mechanisms for improving the wireless network security. We first summarize the
security requirements of wireless networks, including their authenticity,
confidentiality, integrity and availability issues. Next, a comprehensive
overview of security attacks encountered in wireless networks is presented in
view of the network protocol architecture, where the potential security threats
are discussed at each protocol layer. We also provide a survey of the existing
security protocols and algorithms that are adopted in the existing wireless
network standards, such as the Bluetooth, Wi-Fi, WiMAX, and the long-term
evolution (LTE) systems. Then, we discuss the state-of-the-art in
physical-layer security, which is an emerging technique of securing the open
communications environment against eavesdropping attacks at the physical layer.
We also introduce the family of various jamming attacks and their
counter-measures, including the constant jammer, intermittent jammer, reactive
jammer, adaptive jammer and intelligent jammer. Additionally, we discuss the
integration of physical-layer security into existing authentication and
cryptography mechanisms for further securing wireless networks. Finally, some
technical challenges which remain unresolved at the time of writing are
summarized and the future trends in wireless security are discussed.Comment: 36 pages. Accepted to Appear in Proceedings of the IEEE, 201
Scheduling Policies in Time and Frequency Domains for LTE Downlink Channel: A Performance Comparison
A key feature of the Long-Term Evolution (LTE) system is that the packet scheduler can make use of the channel quality information (CQI), which is periodically reported by user equipment either in an aggregate form for the whole downlink channel or distinguished for each available subchannel. This mechanism allows for wide discretion in resource allocation, thus promoting the flourishing of several scheduling algorithms, with different purposes. It is therefore of great interest to compare the performance of such algorithms under different scenarios. Here, we carry out a thorough performance analysis of different scheduling algorithms for saturated User Datagram Protocol (UDP) and Transmission Control Protocol (TCP) traffic sources, as well as consider both the time- and frequency-domain versions of the schedulers and for both flat and frequency-selective channels. The analysis makes it possible to appreciate the difference among the scheduling algorithms and to assess the performance gain, in terms of cell capacity, users' fairness, and packet service time, obtained by exploiting the richer, but heavier, information carried by subchannel CQI. An important part of this analysis is a throughput guarantee scheduler, which we propose in this paper. The analysis reveals that the proposed scheduler provides a good tradeoff between cell capacity and fairness both for TCP and UDP traffic sources
LTE Optimization and Resource Management in Wireless Heterogeneous Networks
Mobile communication technology is evolving with a great pace. The development of the Long Term Evolution (LTE) mobile system by 3GPP is one of the milestones in this direction. This work highlights a few areas in the LTE radio access network where the proposed innovative mechanisms can substantially improve overall LTE system performance. In order to further extend the capacity of LTE networks, an integration with the non-3GPP networks (e.g., WLAN, WiMAX etc.) is also proposed in this work. Moreover, it is discussed how bandwidth resources should be managed in such heterogeneous networks. The work has purposed a comprehensive system architecture as an overlay of the 3GPP defined SAE architecture, effective resource management mechanisms as well as a Linear Programming based analytical solution for the optimal network resource allocation problem. In addition, alternative computationally efficient heuristic based algorithms have also been designed to achieve near-optimal performance
Performance Comparison of Dual Connectivity and Hard Handover for LTE-5G Tight Integration in mmWave Cellular Networks
MmWave communications are expected to play a major role in the Fifth
generation of mobile networks. They offer a potential multi-gigabit throughput
and an ultra-low radio latency, but at the same time suffer from high isotropic
pathloss, and a coverage area much smaller than the one of LTE macrocells. In
order to address these issues, highly directional beamforming and a very
high-density deployment of mmWave base stations were proposed. This Thesis aims
to improve the reliability and performance of the 5G network by studying its
tight and seamless integration with the current LTE cellular network. In
particular, the LTE base stations can provide a coverage layer for 5G mobile
terminals, because they operate on microWave frequencies, which are less
sensitive to blockage and have a lower pathloss. This document is a copy of the
Master's Thesis carried out by Mr. Michele Polese under the supervision of Dr.
Marco Mezzavilla and Prof. Michele Zorzi. It will propose an LTE-5G tight
integration architecture, based on mobile terminals' dual connectivity to LTE
and 5G radio access networks, and will evaluate which are the new network
procedures that will be needed to support it. Moreover, this new architecture
will be implemented in the ns-3 simulator, and a thorough simulation campaign
will be conducted in order to evaluate its performance, with respect to the
baseline of handover between LTE and 5G.Comment: Master's Thesis carried out by Mr. Michele Polese under the
supervision of Dr. Marco Mezzavilla and Prof. Michele Zorz
THROUGHPUT OPTIMIZATION AND ENERGY EFFICIENCY OF THE DOWNLINK IN THE LTE SYSTEM
Nowadays, the usage of smart phones is very popular. More and more people access the Internet with their smart phones. This demands higher data rates from the mobile network operators. Every year the number of users and the amount of information is increasing dramatically. The wireless technology should ensure high data rates to be able to compete with the wire-based technology. The main advantage of the wireless system is the ability for user to be mobile. The 4G LTE system made it possible to gain very high peak data rates. The purpose of this thesis was to investigate the improvement of the system performance for the downlink based on different antenna configurations and different scheduling algorithms. Moreover, the fairness between the users using different schedulers has been analyzed and evaluated. Furthermore, the energy efficiency of the scheduling algorithms in the downlink of LTE systems has been considered. Some important parts of the LTE system are described in the theoretical part of this thesis.fi=Opinnäytetyö kokotekstinä PDF-muodossa.|en=Thesis fulltext in PDF format.|sv=Lärdomsprov tillgängligt som fulltext i PDF-format
- …