53 research outputs found

    Security Concerns on Machine Learning Solutions for 6G Networks in mmWave Beam Prediction

    Get PDF
    6G – sixth generation – is the latest cellular technology currently under development for wireless communication systems. In recent years, machine learning (ML) algorithms have been applied widely in various fields, such as healthcare, transportation, energy, autonomous cars, and many more. Those algorithms have also been used in communication technologies to improve the system performance in terms of frequency spectrum usage, latency, and security. With the rapid developments of ML techniques, especially deep learning (DL), it is critical to consider the security concern when applying the algorithms. While ML algorithms offer significant advantages for 6G networks, security concerns on artificial intelligence (AI) models are typically ignored by the scientific community so far. However, security is also a vital part of AI algorithms because attackers can poison the AI model itself. This paper proposes a mitigation method for adversarial attacks against proposed 6G ML models for the millimeter-wave (mmWave) beam prediction using adversarial training. The main idea behind generating adversarial attacks against ML models is to produce faulty results by manipulating trained DL models for 6G applications for mmWave beam prediction. We also present a proposed adversarial learning mitigation method’s performance for 6G security in mmWave beam prediction application a fast gradient sign method attack. The results show that the defended model under attack’s mean square errors (i.e., the prediction accuracy) are very close to the undefended model without attack

    DL-based CSI Feedback and Cooperative Recovery in Massive MIMO

    Full text link
    In this paper, we exploit the correlation between nearby user equipment (UE) and develop a deep learning-based channel state information (CSI) feedback and cooperative recovery framework, CoCsiNet, to reduce the feedback overhead. The CSI information can be divided into two parts: shared by nearby UE and owned by individual UE. The key idea of exploiting the correlation is to reduce the overhead used to repeatedly feedback shared information. Unlike in the general autoencoder framework, an extra decoder and a combination network are added at the base station to recover the shared information from the feedback CSI of two nearby UE and combine the shared and individual information, respectively, but no modification is performed at the UEs. For a UE with multiple antennas, we also introduce a baseline neural network architecture with long short-term memory modules to extract the correlation of nearby antennas. Given that the CSI phase is not sparse, we propose two magnitude-dependent phase feedback strategies that introduce statistical and instant CSI magnitude information to the phase feedback process, respectively. Simulation results on two different channel datasets show the effectiveness of the proposed CoCsiNet.Comment: This work has been submitted to the IEEE for possible publication. Copyright may be transferred without notice, after which this version may no longer be accessibl

    Defending Adversarial Attacks on Deep Learning Based Power Allocation in Massive MIMO Using Denoising Autoencoders

    Full text link
    Recent work has advocated for the use of deep learning to perform power allocation in the downlink of massive MIMO (maMIMO) networks. Yet, such deep learning models are vulnerable to adversarial attacks. In the context of maMIMO power allocation, adversarial attacks refer to the injection of subtle perturbations into the deep learning model's input, during inference (i.e., the adversarial perturbation is injected into inputs during deployment after the model has been trained) that are specifically crafted to force the trained regression model to output an infeasible power allocation solution. In this work, we develop an autoencoder-based mitigation technique, which allows deep learning-based power allocation models to operate in the presence of adversaries without requiring retraining. Specifically, we develop a denoising autoencoder (DAE), which learns a mapping between potentially perturbed data and its corresponding unperturbed input. We test our defense across multiple attacks and in multiple threat models and demonstrate its ability to (i) mitigate the effects of adversarial attacks on power allocation networks using two common precoding schemes, (ii) outperform previously proposed benchmarks for mitigating regression-based adversarial attacks on maMIMO networks, (iii) retain accurate performance in the absence of an attack, and (iv) operate with low computational overhead.Comment: This work is currently under review for publicatio
    • …
    corecore