2 research outputs found

    Channel estimation and beam training with machine learning applications for millimetre-wave communication systems

    Get PDF
    The fifth generation (5G) wireless system will extend the capabilities of the fourth generation (4G) standards to serve more users and provide timely communication. To this end, the carriers of 5G systems will be able to operate at higher frequency bands, such as the millimetre-wave (mmWave) bands that span from 30 GHz to 300 GHz, to obtain greater bandwidths and higher data rates. As a result, the deployment of 5G networks is required to accommodate more antennas and offer pervasive coverage with controlled power consumption. The complexity of 5G systems introduces new challenges to traditional signal processing techniques. To address these challenges, a major step is to integrate machine learning (ML) algorithms into wireless communication systems. ML can learn patterns from datasets to achieve control and optimisation of complex radio frequency (RF) networks. This PhD thesis focuses on developing efficient channel estimation methods and beam training strategies with the application of ML algorithms for mmWave wireless systems. Firstly, the channel estimation and signal detection problem is investigated for orthogonal frequency-division multiplexing (OFDM) systems that operate at mmWave bands. A deep neural network (DNN)-based joint channel estimation and signal detection approach is proposed to achieve multi-user detection in a one-shot process for non-orthogonal multiple access (NOMA) systems. The DNN acts as the receiver, which can recover the transmitted data by learning the channel implicitly from suitable training. The proposed approach can be adapted to work for both single-input and single-output (SISO) systems and multiple-output and multipleoutput (MIMO) systems. This DNN-based approach is shown to provide good performance for OFDM systems that suffer from severe inter-symbol interference or where small numbers of pilot symbols are used. Secondly, the beam training and tracking problem is studied for mmWave channels with receiver mobility. To reduce the signalling overhead caused by frequent beam training, a lowcomplexity beam training strategy is proposed for mobile mmWave channels, which searches a set of selected beams obtained based on the recent beam search results. By searching only the adjacent beams to the one recently used, the proposed beam training strategy can reduce the beam training delay significantly while maintaining high transmission rates. The proposed strategy works effectively for channel datasets generated using either the stochastic or the raytracing channel model. This strategy is shown to approach the performance for an exhaustive beam search while saving up to 92% on the required beam training overhead. Thirdly, the proposed low-complexity beam training strategy is enhanced with the use of deep reinforcement learning (DRL) for mobile mmWave channels. A DRL-based beam training algorithm is proposed, which can intelligently switch between different beam training methods such that the average beam training overhead is minimised while achieving good spectral efficiency or energy efficiency performance. Given the desired performance requirement in the reward function for the DRL model, the spectral efficiency or energy efficiency can be maximised for the current channel condition by controlling the number of activated RF chains. The DRL-based approach can adjust the amount of beam training overhead required according to the dynamics of the environment. This approach can provide a good overhead-performance trade-off and achieve higher data rates in channels with significant levels of signal blockage
    corecore