61 research outputs found
Improved Modeling of the Correlation Between Continuous-Valued Sources in LDPC-Based DSC
Accurate modeling of the correlation between the sources plays a crucial role
in the efficiency of distributed source coding (DSC) systems. This correlation
is commonly modeled in the binary domain by using a single binary symmetric
channel (BSC), both for binary and continuous-valued sources. We show that
"one" BSC cannot accurately capture the correlation between continuous-valued
sources; a more accurate model requires "multiple" BSCs, as many as the number
of bits used to represent each sample. We incorporate this new model into the
DSC system that uses low-density parity-check (LDPC) codes for compression. The
standard Slepian-Wolf LDPC decoder requires a slight modification so that the
parameters of all BSCs are integrated in the log-likelihood ratios (LLRs).
Further, using an interleaver the data belonging to different bit-planes are
shuffled to introduce randomness in the binary domain. The new system has the
same complexity and delay as the standard one. Simulation results prove the
effectiveness of the proposed model and system.Comment: 5 Pages, 4 figures; presented at the Asilomar Conference on Signals,
Systems, and Computers, Pacific Grove, CA, November 201
A New Reduction Scheme for Gaussian Sum Filters
In many signal processing applications it is required to estimate the
unobservable state of a dynamic system from its noisy measurements. For linear
dynamic systems with Gaussian Mixture (GM) noise distributions, Gaussian Sum
Filters (GSF) provide the MMSE state estimate by tracking the GM posterior.
However, since the number of the clusters of the GM posterior grows
exponentially over time, suitable reduction schemes need to be used to maintain
the size of the bank in GSF. In this work we propose a low computational
complexity reduction scheme which uses an initial state estimation to find the
active noise clusters and removes all the others. Since the performance of our
proposed method relies on the accuracy of the initial state estimation, we also
propose five methods for finding this estimation. We provide simulation results
showing that with suitable choice of the initial state estimation (based on the
shape of the noise models), our proposed reduction scheme provides better state
estimations both in terms of accuracy and precision when compared with other
reduction methods
Non-Adaptive Distributed Compression in Networks
In this paper, we discuss non-adaptive distributed compression of inter-node
correlated real-valued messages. To do so, we discuss the performance of
conventional packet forwarding via routing, in terms of the total network load
versus the resulting quality of service (distortion level). As a better
alternative for packet forwarding, we briefly describe our previously proposed
one-step Quantized Network Coding (QNC), and make motivating arguments on its
advantage when the appropriate marginal rates for distributed source coding are
not available at the encoder source nodes. We also derive analytic guarantees
on the resulting distortion of our one-step QNC scenario. Finally, we conclude
the paper by providing a mathematical comparison between the total network
loads of one-step QNC and conventional packet forwarding, showing a significant
reduction in the case of one-step QNC.Comment: Submitted for 2013 IEEE International Symposium on Information Theor
Medically Relevant Criteria used in EEG Compression for Improved Post-Compression Seizure Detection
Biomedical signals aid in the diagnosis of different disorders and
abnormalities. When targeting lossy compression of such signals, the medically
relevant information that lies within the data should maintain its accuracy and
thus its reliability. In fact, signal models that are inspired by the
bio-physical properties of the signals at hand allow for a compression that
preserves more naturally the clinically significant features of these signals.
In this paper, we illustrate this through the example of EEG signals; more
specifically, we analyze three specific lossy EEG compression schemes. These
schemes are based on signal models that have different degrees of reliance on
signal production and physiological characteristics of EEG. The resilience of
these schemes is illustrated through the performance of seizure detection post
compression.Comment: This work has been submitted to the IEEE for possible publication.
Copyright may be transferred without notice, after which this version may no
longer be accessibl
- …