59 research outputs found
Benchmarking End-to-end Learning of MIMO Physical-Layer Communication
End-to-end data-driven machine learning (ML) of multiple-input
multiple-output (MIMO) systems has been shown to have the potential of
exceeding the performance of engineered MIMO transceivers, without any a priori
knowledge of communication-theoretic principles. In this work, we aim to
understand to what extent and for which scenarios this claim holds true when
comparing with fair benchmarks. We study closed-loop MIMO, open-loop MIMO, and
multi-user MIMO and show that the gains of ML-based communication in the former
two cases can be to a large extent ascribed to implicitly learned geometric
shaping and bit and power allocation, not to learning new spatial encoders. For
MU-MIMO, we demonstrate the feasibility of a novel method with centralized
learning and decentralized executing, outperforming conventional zero-forcing.
For each scenario, we provide explicit descriptions as well as open-source
implementations of the selected neural-network architectures.Comment: 6 pages, 8 figures, conference pape
Truncated Polynomial Expansion-Based Detection in Massive MIMO: A Model-Driven Deep Learning Approach
In this paper, we propose a deep learning (DL)-based approach for efficiently
computing the inverse of Hermitian matrices using truncated polynomial
expansion (TPE). Our model-driven approach involves optimizing the coefficients
of the TPE during an offline training procedure for a given number of TPE
terms. We apply this method to signal detection in uplink massive
multiple-input multiple-output (MIMO) systems, where the matrix inverse
operation required by linear detectors, such as zero-forcing (ZF) and minimum
mean square error (MMSE), is approximated using TPE. Our simulation results
demonstrate that the proposed learned TPE-based method outperforms the
conventional TPE method with optimal coefficients in terms of asymptotic
convergence speed and reduces the computational complexity of the online
detection stage, albeit at the expense of the offline training stage. However,
the limited number of trainable parameters leads to a swift offline training
process.Comment: 5 pages, 2 figures, 2 table
- …