11,065 research outputs found
Deep Joint Source-Channel Coding for Wireless Image Transmission
We propose a joint source and channel coding (JSCC) technique for wireless image transmission that does not rely on explicit codes for either compression or error correction; instead, it directly maps the image pixel values to the complex-valued channel input symbols. We parameterize the encoder and decoder functions by two convolutional neural networks (CNNs), which are trained jointly, and can be considered as an autoencoder with a non-trainable layer in the middle that represents the noisy communication channel. Our results show that the proposed deep JSCC scheme outperforms digital transmission concatenating JPEG or JPEG2000 compression with a capacity achieving channel code at low signal-to-noise ratio (SNR) and channel bandwidth values in the presence of additive white Gaussian noise (AWGN). More strikingly, deep JSCC does not suffer from the “cliff effect,” and it provides a graceful performance degradation as the channel SNR varies with respect to the SNR value assumed during training. In the case of a slow Rayleigh fading channel, deep JSCC learns noise resilient coded representations and significantly outperforms separation-based digital communication at all SNR and channel bandwidth values
Deep Joint Source-Channel Coding for Wireless Image Transmission
We propose a novel joint source and channel coding (JSCC)
scheme for wireless image transmission that departs from
the conventional use of explicit source and channel codes for
compression and error correction, and directly maps the image pixel values to the complex-valued channel input signal.
Our encoder-decoder pair form an autoencoder with a nontrainable layer in the middle, which represents the noisy communication channel. Our results show that the proposed deep
JSCC scheme outperforms separation-based digital transmission at low signal-to-noise ratio (SNR) and low channel bandwidth regimes in the presence of additive white Gaussian
noise (AWGN). More strikingly, deep JSCC does not suffer
from the “cliff effect” as the channel SNR varies with respect
to the SNR value assumed during training. In the case of a
slow Rayleigh fading channel, deep JSCC can learn to communicate without explicit pilot signals or channel estimation,
and significantly outperforms separation-based digital communication at all SNR and channel bandwidth values
Channel Adaptive DL based Joint Source-Channel Coding without A Prior Knowledge
Significant progress has been made in wireless Joint Source-Channel Coding
(JSCC) using deep learning techniques. The latest DL-based image JSCC methods
have demonstrated exceptional performance across various signal-to-noise ratio
(SNR) levels during transmission, while also avoiding cliff effects. However,
current channel adaptive JSCC methods rely heavily on channel prior knowledge,
which can lead to performance degradation in practical applications due to
channel mismatch effects. This paper proposes a novel approach for image
transmission, called Channel Blind Joint Source-Channel Coding (CBJSCC). CBJSCC
utilizes Deep Learning techniques to achieve exceptional performance across
various signal-to-noise ratio (SNR) levels during transmission, without relying
on channel prior information. We have designed an Inverted Residual Attention
Bottleneck (IRAB) module for the model, which can effectively reduce the number
of parameters while expanding the receptive field. In addition, we have
incorporated a convolution and self-attention mixed encoding module to
establish long-range dependency relationships between channel symbols. Our
experiments have shown that CBJSCC outperforms existing channel adaptive
DL-based JSCC methods that rely on feedback information. Furthermore, we found
that channel estimation does not significantly benefit CBJSCC, which provides
insights for the future design of DL-based JSCC methods. The reliability of the
proposed method is further demonstrated through an analysis of the model
bottleneck and its adaptability to different domains, as shown by our
experiments
Deep Joint Source-Channel Coding for Adaptive Image Transmission over MIMO Channels
This paper introduces a vision transformer (ViT)-based deep joint source and
channel coding (DeepJSCC) scheme for wireless image transmission over
multiple-input multiple-output (MIMO) channels, denoted as DeepJSCC-MIMO. We
consider DeepJSCC-MIMO for adaptive image transmission in both open-loop and
closed-loop MIMO systems. The novel DeepJSCC-MIMO architecture surpasses the
classical separation-based benchmarks with robustness to channel estimation
errors and showcases remarkable flexibility in adapting to diverse channel
conditions and antenna numbers without requiring retraining. Specifically, by
harnessing the self-attention mechanism of ViT, DeepJSCC-MIMO intelligently
learns feature mapping and power allocation strategies tailored to the unique
characteristics of the source image and prevailing channel conditions.
Extensive numerical experiments validate the significant improvements in
transmission quality achieved by DeepJSCC-MIMO for both open-loop and
closed-loop MIMO systems across a wide range of scenarios. Moreover,
DeepJSCC-MIMO exhibits robustness to varying channel conditions, channel
estimation errors, and different antenna numbers, making it an appealing
solution for emerging semantic communication systems.Comment: arXiv admin note: text overlap with arXiv:2210.1534
Deep Joint Source-Channel Coding for Wireless Image Transmission with Entropy-Aware Adaptive Rate Control
Adaptive rate control for deep joint source and channel coding (JSCC) is
considered as an effective approach to transmit sufficient information in
scenarios with limited communication resources. We propose a deep JSCC scheme
for wireless image transmission with entropy-aware adaptive rate control, using
a single deep neural network to support multiple rates and automatically adjust
the rate based on the feature maps of the input image and their entropy, as
well as the channel conditions. In particular, we maximize the entropy of the
feature maps to increase the average information carried by each transmitted
symbol during the training. We further decide which feature maps should be
activated based on their entropy, which improves the efficiency of the
transmitted symbols. We also propose a pruning module to remove less important
pixels in the activated feature maps in order to further improve transmission
efficiency. The experimental results demonstrate that our proposed scheme
learns an effective rate control strategy that reduces the required channel
bandwidth while preserving the quality of the reconstructed images
- …