439 research outputs found
Design and Analysis of LT Codes with Decreasing Ripple Size
In this paper we propose a new design of LT codes, which decreases the amount
of necessary overhead in comparison to existing designs. The design focuses on
a parameter of the LT decoding process called the ripple size. This parameter
was also a key element in the design proposed in the original work by Luby.
Specifically, Luby argued that an LT code should provide a constant ripple size
during decoding. In this work we show that the ripple size should decrease
during decoding, in order to reduce the necessary overhead. Initially we
motivate this claim by analytical results related to the redundancy within an
LT code. We then propose a new design procedure, which can provide any desired
achievable decreasing ripple size. The new design procedure is evaluated and
compared to the current state of the art through simulations. This reveals a
significant increase in performance with respect to both average overhead and
error probability at any fixed overhead
Information Loss in the Human Auditory System
From the eardrum to the auditory cortex, where acoustic stimuli are decoded,
there are several stages of auditory processing and transmission where
information may potentially get lost. In this paper, we aim at quantifying the
information loss in the human auditory system by using information theoretic
tools.
To do so, we consider a speech communication model, where words are uttered
and sent through a noisy channel, and then received and processed by a human
listener.
We define a notion of information loss that is related to the human word
recognition rate. To assess the word recognition rate of humans, we conduct a
closed-vocabulary intelligibility test. We derive upper and lower bounds on the
information loss. Simulations reveal that the bounds are tight and we observe
that the information loss in the human auditory system increases as the signal
to noise ratio (SNR) decreases. Our framework also allows us to study whether
humans are optimal in terms of speech perception in a noisy environment.
Towards that end, we derive optimal classifiers and compare the human and
machine performance in terms of information loss and word recognition rate. We
observe a higher information loss and lower word recognition rate for humans
compared to the optimal classifiers. In fact, depending on the SNR, the machine
classifier may outperform humans by as much as 8 dB. This implies that for the
speech-in-stationary-noise setup considered here, the human auditory system is
sub-optimal for recognizing noisy words
Real-Time Perceptual Moving-Horizon Multiple-Description Audio Coding
A novel scheme for perceptual coding of audio for robust and real-time communication is designed and analyzed. As an alternative to PCM, DPCM, and more general noise-shaping converters, we propose to use psychoacoustically optimized noise-shaping quantizers based on the moving-horizon principle. In moving-horizon quantization, a few samples look-ahead is allowed at the encoder, which makes it possible to better shape the quantization noise and thereby reduce the resulting distortion over what is possible with conventional noise-shaping techniques. It is first shown that significant gains over linear PCM can be obtained without introducing a delay and without requiring postprocessing at the decoder, i.e., the encoded samples can be stored as, e.g., 16-bit linear PCM on CD-ROMs, and played out on standards-compliant CD players. We then show that multiple-description coding can be combined with moving-horizon quantization in order to combat possible erasures on the wireless link without introducing additional delays
- …