835 research outputs found
Bilinear Gaussian Belief Propagation for Massive MIMO Detection with Non-Orthogonal Pilots
Ito K., Takahashi T., Ibi S., et al. Bilinear Gaussian Belief Propagation for Massive MIMO Detection with Non-Orthogonal Pilots. IEEE Transactions on Communications , (2023); https://doi.org/10.1109/TCOMM.2023.3325479.We propose a novel joint channel and data estimation (JCDE) algorithm via bilinear Gaussian belief propagation (BiGaBP) for massive multi-user MIMO (MU-MIMO) systems with non-orthogonal pilot sequences. The contribution aims to reduce significantly the communication overhead required for channel acquisition by enabling the use of short non-orthogonal pilots, while maintaining multi-user detection (MUD) capability. Bilinear generalized approximate message passing (BiGAMP), which is systematically derived by extending approximate message passing (AMP) to the bilinear inference problem (BIP), provides computationally efficient approximate implementations of large-scale JCDE via sum-product algorithm (SPA); however, as the pilot length decreases, the estimation accuracy is severely degraded. To tackle this issue, the proposed BiGaBP algorithm generalizes BiGAMP by relaxing its dependence on the large-system limit approximation and leveraging the belief propagation (BP) concept. In addition, a novel belief scaling method complying with the data detection accuracy for each iteration step is designed to avoid the divergence behavior of iterative estimation in the early iterations due to the use of non-orthogonal pilots, especially in insufficient large-system conditions. Simulation results show that the proposed method outperforms the state-of-the-art schemes and approaches the performance of idealized (genie-aided) scheme in terms of mean square error (MSE) and bit error rate (BER) performances
Integrated Sensing and Communications for 3D Object Imaging via Bilinear Inference
We consider an uplink integrated sensing and communications (ISAC) scenario
where the detection of data symbols from multiple user equipment (UEs) occurs
simultaneously with a three-dimensional (3D) estimation of the environment,
extracted from the scattering features present in the channel state information
(CSI) and utilizing the same physical layer communications air interface, as
opposed to radar technologies. By exploiting a discrete (voxelated)
representation of the environment, two novel ISAC schemes are derived with
purpose-built message passing (MP) rules for the joint estimation of data
symbols and status (filled/empty) of the discretized environment. The first
relies on a modular feedback structure in which the data symbols and the
environment are estimated alternately, whereas the second leverages a bilinear
inference framework to estimate both variables concurrently. Both contributed
methods are shown via simulations to outperform the state-of-the-art (SotA) in
accurately recovering the transmitted data as well as the 3D image of the
environment. An analysis of the computational complexities of the proposed
methods reveals distinct advantages of each scheme, namely, that the bilinear
solution exhibits a superior robustness to short pilots and channel blockages,
while the alternating solution offers lower complexity with large number of UEs
and superior performance in ideal conditions
Speech emotion recognition based on bi-directional acoustic–articulatory conversion
Acoustic and articulatory signals are naturally coupled and complementary. The challenge of acquiring articulatory data and the nonlinear ill-posedness of acoustic–articulatory conversions have resulted in previous studies on speech emotion recognition (SER) primarily relying on unidirectional acoustic–articulatory conversions. However, these studies have ignored the potential benefits of bi-directional acoustic–articulatory conversion. Addressing the problem of nonlinear ill-posedness and effectively extracting and utilizing these two modal features in SER remain open research questions. To bridge this gap, this study proposes a Bi-A2CEmo framework that simultaneously addresses the bi-directional acoustic-articulatory conversion for SER. This framework comprises three components: a Bi-MGAN that addresses the nonlinear ill-posedness problem, KCLNet that enhances the emotional attributes of the mapped features, and ResTCN-FDA that fully exploits the emotional attributes of the features. Another challenge is the absence of a parallel acoustic-articulatory emotion database. To overcome this issue, this study utilizes electromagnetic articulography (EMA) to create a multi-modal acoustic-articulatory emotion database for Mandarin Chinese called STEM-EVA. A comparative analysis is then conducted between the proposed method and state-of-the-art models to evaluate the effectiveness of the framework. Bi-A2CEmo achieves an accuracy of 89.04\% in SER, which is an improvement of 5.27\% compared with the actual acoustic and articulatory features recorded by the EMA. The results for the STEM-EVA dataset show that Bi-MGAN achieves a higher accuracy in mapping and inversion than conventional conversion networks. Visualization of the mapped features before and after enhancement reveals that KCLNet reduces the intra-class spacing while increasing the inter-class spacing of the features. ResTCN-FDA demonstrates high recognition accuracy on three publicly available datasets. The experimental results show that the proposed bi-directional acoustic-articulatory conversion framework can significantly improve the SER performance
- …