29,058 research outputs found
Filtering Random Graph Processes Over Random Time-Varying Graphs
Graph filters play a key role in processing the graph spectra of signals
supported on the vertices of a graph. However, despite their widespread use,
graph filters have been analyzed only in the deterministic setting, ignoring
the impact of stochastic- ity in both the graph topology as well as the signal
itself. To bridge this gap, we examine the statistical behavior of the two key
filter types, finite impulse response (FIR) and autoregressive moving average
(ARMA) graph filters, when operating on random time- varying graph signals (or
random graph processes) over random time-varying graphs. Our analysis shows
that (i) in expectation, the filters behave as the same deterministic filters
operating on a deterministic graph, being the expected graph, having as input
signal a deterministic signal, being the expected signal, and (ii) there are
meaningful upper bounds for the variance of the filter output. We conclude the
paper by proposing two novel ways of exploiting randomness to improve (joint
graph-time) noise cancellation, as well as to reduce the computational
complexity of graph filtering. As demonstrated by numerical results, these
methods outperform the disjoint average and denoise algorithm, and yield a (up
to) four times complexity redution, with very little difference from the
optimal solution
Introduction to Random Signals and Noise
Random signals and noise are present in many engineering systems and networks. Signal processing techniques allow engineers to distinguish between useful signals in audio, video or communication equipment, and interference, which disturbs the desired signal. With a strong mathematical grounding, this text provides a clear introduction to the fundamentals of stochastic processes and their practical applications to random signals and noise. With worked examples, problems, and detailed appendices, Introduction to Random Signals and Noise gives the reader the knowledge to design optimum systems for effectively coping with unwanted signals.\ud
\ud
Key features:\ud
âą Considers a wide range of signals and noise, including analogue, discrete-time and bandpass signals in both time and frequency domains.\ud
âą Analyses the basics of digital signal detection using matched filtering, signal space representation and correlation receiver.\ud
âą Examines optimal filtering methods and their consequences.\ud
âą Presents a detailed discussion of the topic of Poisson processed and shot noise.\u
One- and two-level filter-bank convolvers
In a recent paper, it was shown in detail that in the case of orthonormal and biorthogonal filter banks we can convolve two signals by directly convolving the subband signals and combining the results. In this paper, we further generalize the result. We also derive the statistical coding gain for the generalized subband convolver. As an application, we derive a novel low sensitivity structure for FIR filters from the convolution theorem. We define and derive a deterministic coding gain of the subband convolver over direct convolution for a fixed wordlength implementation. This gain serves as a figure of merit for the low sensitivity structure. Several numerical examples are included to demonstrate the usefulness of these ideas. By using the generalized polyphase representation, we show that the subband convolvers, linear periodically time varying systems, and digital block filtering can be viewed in a unified manner. Furthermore, the scheme called IFIR filtering is shown to be a special case of the convolver
Streaming an image through the eye: The retina seen as a dithered scalable image coder
We propose the design of an original scalable image coder/decoder that is
inspired from the mammalians retina. Our coder accounts for the time-dependent
and also nondeterministic behavior of the actual retina. The present work
brings two main contributions: As a first step, (i) we design a deterministic
image coder mimicking most of the retinal processing stages and then (ii) we
introduce a retinal noise in the coding process, that we model here as a dither
signal, to gain interesting perceptual features. Regarding our first
contribution, our main source of inspiration will be the biologically plausible
model of the retina called Virtual Retina. The main novelty of this coder is to
show that the time-dependent behavior of the retina cells could ensure, in an
implicit way, scalability and bit allocation. Regarding our second
contribution, we reconsider the inner layers of the retina. We emit a possible
interpretation for the non-determinism observed by neurophysiologists in their
output. For this sake, we model the retinal noise that occurs in these layers
by a dither signal. The dithering process that we propose adds several
interesting features to our image coder. The dither noise whitens the
reconstruction error and decorrelates it from the input stimuli. Furthermore,
integrating the dither noise in our coder allows a faster recognition of the
fine details of the image during the decoding process. Our present paper goal
is twofold. First, we aim at mimicking as closely as possible the retina for
the design of a novel image coder while keeping encouraging performances.
Second, we bring a new insight concerning the non-deterministic behavior of the
retina.Comment: arXiv admin note: substantial text overlap with arXiv:1104.155
Mutual Information and Minimum Mean-square Error in Gaussian Channels
This paper deals with arbitrarily distributed finite-power input signals
observed through an additive Gaussian noise channel. It shows a new formula
that connects the input-output mutual information and the minimum mean-square
error (MMSE) achievable by optimal estimation of the input given the output.
That is, the derivative of the mutual information (nats) with respect to the
signal-to-noise ratio (SNR) is equal to half the MMSE, regardless of the input
statistics. This relationship holds for both scalar and vector signals, as well
as for discrete-time and continuous-time noncausal MMSE estimation. This
fundamental information-theoretic result has an unexpected consequence in
continuous-time nonlinear estimation: For any input signal with finite power,
the causal filtering MMSE achieved at SNR is equal to the average value of the
noncausal smoothing MMSE achieved with a channel whose signal-to-noise ratio is
chosen uniformly distributed between 0 and SNR
- âŠ