123 research outputs found

    Artificial neural networks for location estimation and co-cannel interference suppression in cellular networks

    Get PDF
    This thesis reports on the application of artificial neural networks to two important problems encountered in cellular communications, namely, location estimation and co-channel interference suppression. The prediction of a mobile location using propagation path loss (signal strength) is a very difficult and complex task. Several techniques have been proposed recently mostly based on linearized, geometrical and maximum likelihood methods. An alternative approach based on artificial neural networks is proposed in this thesis which offers the advantages of increased flexibility to adapt to different environments and high speed parallel processing. Location estimation provides users of cellular telephones with information about their location. Some of the existing location estimation techniques such as those used in GPS satellite navigation systems require non-standard features, either from the cellular phone or the cellular network. However, it is possible to use the existing GSM technology for location estimation by taking advantage of the signals transmitted between the phone and the network. This thesis proposes the application of neural networks to predict the location coordinates from signal strength data. New multi-layered perceptron and radial basis function based neural networks are employed for the prediction of mobile locations using signal strength measurements in a simulated COST-231 metropolitan environment. In addition, initial preliminary results using limited available real signal-strength measurements in a metropolitan environment are also reported comparing the performance of the neural predictors with a conventional linear technique. The results indicate that the neural predictors can be trained to provide a near perfect mapping using signal strength measurements from two or more base stations. The second application of neural networks addressed in this thesis, is concerned with adaptive equalization, which is known to be an important technique for combating distortion and Inter-Symbol Interference (ISI) in digital communication channels. However, many communication systems are also impaired by what is known as co-channel interference (CCI). Many digital communications systems such as digital cellular radio (DCR) and dual polarized micro-wave radio, for example, employ frequency re-usage and often exhibit performance limitation due to co-channel interference. The degradation in performance due to CCI is more severe than due to ISI. Therefore, simple and effective interference suppression techniques are required to mitigate the interference for a high-quality signal reception. The current work briefly reviews the application of neural network based non-linear adaptive equalizers to the problem of combating co-channel interference, without a priori knowledge of the channel or co-channel orders. A realistic co-channel system is used as a case study to demonstrate the superior equalization capability of the functional-link neural network based Decision Feedback Equalizer (DFE) compared to other conventional linear and neural network based non-linear adaptive equalizers.This project was funded by Solectron (Scotland) Ltd

    A Comprehensive Investigation of Beam Management Through Conventional and Deep Learning Approach

    Get PDF
    5G spectrum uses cutting-edge technology which delivers high data rates, low latency, increased capacity, and high spectrum utilization. To cater to these requirements various technologies are available such as Multiple Access Technology (MAT), Multiple Input Multiple Output technology (MIMO), Millimetre (mm) wave technology, Non-Orthogonal Multiple Access Technology (NOMA), Simultaneous Wireless Information and Power Transfer (SWIPT). Of all available technologies, mmWave is prominent as it provides favorable opportunities for 5G. Millimeter-wave is capable of providing a high data rate i.e., 10 Gbit/sec. Also, a tremendous amount of raw bandwidth is available i.e., around 250 GHz, which is an attractive characteristic of the mmWave band to relieve mobile data traffic congestion in the low frequency band. It has a high frequency i.e., 30 – 300 GHz, giving very high speed. It has a very short wavelength i.e., 1-10mm, because of this it provides the compact size of the component. It will provide a throughput of up to 20 Gbps. It has narrow beams and will increase security and reduce interference. When the main beam of the transmitter and receiver are not aligned properly there is a problem in ideal communication. To solve this problem beam management is one of the solutions to form a strong communication link between transmitter and receiver. This paper aims to address challenges in beam management and proposes a framework for realization. Towards the same, the paper initially introduces various challenges in beam management. Towards building an effective beam management system when a user is moving, various steps are present like beam selection, beam tracking, beam alignment, and beam forming. Hence the subsequent sections of the paper illustrate various beam management procedures in mmWave using conventional methods as well as using deep learning techniques. The paper also presents a case study on the framework's implementation using the above-mentioned techniques in mmWave communication. Also glimpses on future research directions are detailed in the final sections. Such beam management techniques when used for mmWave technology will enable build fast, efficient, and capable 5G networks

    Single-Frequency Network Terrestrial Broadcasting with 5GNR Numerology

    Get PDF
    L'abstract è presente nell'allegato / the abstract is in the attachmen

    Contribution to reliable end-to-end communication over 5G networks using advanced techniques

    Get PDF
    5G cellular communication, especially with its hugely available bandwidth provided by millimeter-wave, is a promising technology to fulfill the coming high demand for vast data rates. These networks can support new use cases such as Vehicle to Vehicle and augmented reality due to its novel features such as network slicing along with the mmWave multi-gigabit-persecond data rate. Nevertheless, 5G cellular networks suffer from some shortcomings, especially in high frequencies because of the intermittent nature of channels when the frequency rises. Non-line of sight state is one of the significant issues that the new generation encounters. This drawback is because of the intense susceptibility of higher frequencies to blockage caused by obstacles and misalignment. This unique characteristic can impair the performance of the reliable transport layer widely deployed protocol, TCP, in attaining high throughput and low latency throughout a fair network. As a result, the protocol needs to adjust the congestion window size based on the current situation of the network. However, TCP cannot adjust its congestion window efficiently, which leads to throughput degradation of the protocol. This thesis presents a comprehensive analysis of reliable end-to-end communications in 5G networks and analyzes TCP’s behavior in one of the 3GPP’s well-known s cenarios called urban deployment. Furtherm ore, two novel TCPs bas ed on artificial intelligence have been proposed to deal with this issue. The first protocol uses Fuzzy logic, a subset of artificial intelligence, and the second one is based on deep learning. The extensively conducted simulations showed that the newly proposed protocols could attain higher performance than common TCPs, such as BBR, HighSpeed, Cubic, and NewReno in terms of throughput, RTT, and sending rate adjustment in the urban scenario. The new protocols' superiority is achieved by employing smartness in the conges tions control mechanism of TCP, which is a powerful enabler in fos tering TCP’s functionality. To s um up, the 5G network is a promising telecommunication infrastructure that will revolute various aspects of communication. However, different parts of the Internet, such as its regulations and protocol stack, will face new challenges, which need to be solved in order to exploit 5G capacity, and without intelligent rules and protocols, the high bandwidth of 5G, especially 5G mmWave will be wasted. Two novel schemes to solve the issues have been proposed based on an Artificial Intelligence subset technique called fuzzy and a machine learning-based approach called Deep learning to enhance the performance of 5G mmWave by improving the functionality of the transport layer. The obtained results indicated that the new schemes could improve the functionality of TCP by giving intelligence to the protocol. As the protocol works more smartly, it can make sufficient decisions on different conditions.La comunicació cel·lular 5G, especialment amb l’amplada de banda molt disponible que proporciona l’ona mil·limètrica, és una tecnologia prometedora per satisfer l’elevada demanda de grans velocitats de dades. Aquestes xarxes poden admetre casos d’ús nous, com ara Vehicle to Vehicle i realitat augmentada, a causa de les seves novetats, com ara el tall de xarxa juntament amb la velocitat de dades mWave de multi-gigabit per segon. Tot i això, les xarxes cel·lulars 5G pateixen algunes deficiències, sobretot en freqüències altes a causa de la naturalesa intermitent dels canals quan augmenta la freqüència. L’estat de no visió és un dels problemes significatius que troba la nova generació. Aquest inconvenient es deu a la intensa susceptibilitat de freqüències més altes al bloqueig causat per obstacles i desalineació. Aquesta característica única pot perjudicar el rendiment del protocol TCP, àmpliament desplegat, de capa de transport fiable en aconseguir un alt rendiment i una latència baixa en tota una xarxa justa. Com a resultat, el protocol ha d’ajustar la mida de la finestra de congestió en funció de la situació actual de la xarxa. Tot i això, TCP no pot ajustar la seva finestra de congestió de manera eficient, cosa que provoca una degradació del rendiment del protocol. Aquesta tesi presenta una anàlisi completa de comunicacions extrem a extrem en xarxes 5G i analitza el comportament de TCP en un dels escenaris coneguts del 3GPP anomenat desplegament urbà. A més, s'han proposat dos TCP nous basats en intel·ligència artificial per tractar aquest tema. El primer protocol utilitza la lògica Fuzzy, un subconjunt d’intel·ligència artificial, i el segon es basa en l’aprenentatge profund. Les simulacions àmpliament realitzades van mostrar que els protocols proposats recentment podrien assolir un rendiment superior als TCP habituals, com ara BBR, HighSpeed, Cubic i NewReno, en termes de rendiment, RTT i ajust d’índex d’enviament en l’escenari urbà. La superioritat dels nous protocols s’aconsegueix utilitzant la intel·ligència en el mecanisme de control de congestions de TCP, que és un poderós facilitador per fomentar la funcionalitat de TCP. En resum, la xarxa 5G és una prometedora infraestructura de telecomunicacions que revolucionarà diversos aspectes de la comunicació. No obstant això, diferents parts d’Internet, com ara les seves regulacions i la seva pila de protocols, s’enfrontaran a nous reptes, que cal resoldre per explotar la capacitat 5G, i sens regles i protocols intel·ligents, l’amplada de banda elevada de 5G, especialment 5G mmWave, pot ser desaprofitat. S'han proposat dos nous es quemes per resoldre els problemes basats en una tècnica de subconjunt d'Intel·ligència Artificial anomenada “difusa” i un enfocament basat en l'aprenentatge automàtic anomenat “Aprenentatge profund” per millorar el rendiment de 5G mmWave, millorant la funcionalitat de la capa de transport. Els resultats obtinguts van indicar que els nous esquemes podrien millorar la funcionalitat de TCP donant intel·ligència al protocol. Com que el protocol funciona de manera més intel·ligent, pot prendre decisions suficients en diferents condicionsPostprint (published version

    Cooperative Radio Communications for Green Smart Environments

    Get PDF
    The demand for mobile connectivity is continuously increasing, and by 2020 Mobile and Wireless Communications will serve not only very dense populations of mobile phones and nomadic computers, but also the expected multiplicity of devices and sensors located in machines, vehicles, health systems and city infrastructures. Future Mobile Networks are then faced with many new scenarios and use cases, which will load the networks with different data traffic patterns, in new or shared spectrum bands, creating new specific requirements. This book addresses both the techniques to model, analyse and optimise the radio links and transmission systems in such scenarios, together with the most advanced radio access, resource management and mobile networking technologies. This text summarises the work performed by more than 500 researchers from more than 120 institutions in Europe, America and Asia, from both academia and industries, within the framework of the COST IC1004 Action on "Cooperative Radio Communications for Green and Smart Environments". The book will have appeal to graduates and researchers in the Radio Communications area, and also to engineers working in the Wireless industry. Topics discussed in this book include: • Radio waves propagation phenomena in diverse urban, indoor, vehicular and body environments• Measurements, characterization, and modelling of radio channels beyond 4G networks• Key issues in Vehicle (V2X) communication• Wireless Body Area Networks, including specific Radio Channel Models for WBANs• Energy efficiency and resource management enhancements in Radio Access Networks• Definitions and models for the virtualised and cloud RAN architectures• Advances on feasible indoor localization and tracking techniques• Recent findings and innovations in antenna systems for communications• Physical Layer Network Coding for next generation wireless systems• Methods and techniques for MIMO Over the Air (OTA) testin

    Reliable indoor optical wireless communication in the presence of fixed and random blockers

    Get PDF
    The advanced innovation of smartphones has led to the exponential growth of internet users which is expected to reach 71% of the global population by the end of 2027. This in turn has given rise to the demand for wireless data and internet devices that is capable of providing energy-efficient, reliable data transmission and high-speed wireless data services. Light-fidelity (LiFi), known as one of the optical wireless communication (OWC) technology is envisioned as a promising solution to accommodate these demands. However, the indoor LiFi channel is highly environment-dependent which can be influenced by several crucial factors (e.g., presence of people, furniture, random users' device orientation and the limited field of view (FOV) of optical receivers) which may contribute to the blockage of the line-of-sight (LOS) link. In this thesis, it is investigated whether deep learning (DL) techniques can effectively learn the distinct features of the indoor LiFi environment in order to provide superior performance compared to the conventional channel estimation techniques (e.g., minimum mean square error (MMSE) and least squares (LS)). This performance can be seen particularly when access to real-time channel state information (CSI) is restricted and is achieved with the cost of collecting large and meaningful data to train the DL neural networks and the training time which was conducted offline. Two DL-based schemes are designed for signal detection and resource allocation where it is shown that the proposed methods were able to offer close performance to the optimal conventional schemes and demonstrate substantial gain in terms of bit-error ratio (BER) and throughput especially in a more realistic or complex indoor environment. Performance analysis of LiFi networks under the influence of fixed and random blockers is essential and efficient solutions capable of diminishing the blockage effect is required. In this thesis, a CSI acquisition technique for a reconfigurable intelligent surface (RIS)-aided LiFi network is proposed to significantly reduce the dimension of the decision variables required for RIS beamforming. Furthermore, it is shown that several RIS attributes such as shape, size, height and distribution play important roles in increasing the network performance. Finally, the performance analysis for an RIS-aided realistic indoor LiFi network are presented. The proposed RIS configuration shows outstanding performances in reducing the network outage probability under the effect of blockages, random device orientation, limited receiver's FOV, furniture and user behavior. Establishing a LOS link that achieves uninterrupted wireless connectivity in a realistic indoor environment can be challenging. In this thesis, an analysis of link blockage is presented for an indoor LiFi system considering fixed and random blockers. In particular, novel analytical framework of the coverage probability for a single source and multi-source are derived. Using the proposed analytical framework, link blockages of the indoor LiFi network are carefully investigated and it is shown that the incorporation of multiple sources and RIS can significantly reduce the LOS coverage blockage probability in indoor LiFi systems
    • …
    corecore