20 research outputs found
Sub-Nyquist Sampling: Bridging Theory and Practice
Sampling theory encompasses all aspects related to the conversion of
continuous-time signals to discrete streams of numbers. The famous
Shannon-Nyquist theorem has become a landmark in the development of digital
signal processing. In modern applications, an increasingly number of functions
is being pushed forward to sophisticated software algorithms, leaving only
those delicate finely-tuned tasks for the circuit level.
In this paper, we review sampling strategies which target reduction of the
ADC rate below Nyquist. Our survey covers classic works from the early 50's of
the previous century through recent publications from the past several years.
The prime focus is bridging theory and practice, that is to pinpoint the
potential of sub-Nyquist strategies to emerge from the math to the hardware. In
that spirit, we integrate contemporary theoretical viewpoints, which study
signal modeling in a union of subspaces, together with a taste of practical
aspects, namely how the avant-garde modalities boil down to concrete signal
processing systems. Our hope is that this presentation style will attract the
interest of both researchers and engineers in the hope of promoting the
sub-Nyquist premise into practical applications, and encouraging further
research into this exciting new frontier.Comment: 48 pages, 18 figures, to appear in IEEE Signal Processing Magazin
From Theory to Practice: Sub-Nyquist Sampling of Sparse Wideband Analog Signals
Conventional sub-Nyquist sampling methods for analog signals exploit prior
information about the spectral support. In this paper, we consider the
challenging problem of blind sub-Nyquist sampling of multiband signals, whose
unknown frequency support occupies only a small portion of a wide spectrum. Our
primary design goals are efficient hardware implementation and low
computational load on the supporting digital processing. We propose a system,
named the modulated wideband converter, which first multiplies the analog
signal by a bank of periodic waveforms. The product is then lowpass filtered
and sampled uniformly at a low rate, which is orders of magnitude smaller than
Nyquist. Perfect recovery from the proposed samples is achieved under certain
necessary and sufficient conditions. We also develop a digital architecture,
which allows either reconstruction of the analog input, or processing of any
band of interest at a low rate, that is, without interpolating to the high
Nyquist rate. Numerical simulations demonstrate many engineering aspects:
robustness to noise and mismodeling, potential hardware simplifications,
realtime performance for signals with time-varying support and stability to
quantization effects. We compare our system with two previous approaches:
periodic nonuniform sampling, which is bandwidth limited by existing hardware
devices, and the random demodulator, which is restricted to discrete multitone
signals and has a high computational load. In the broader context of Nyquist
sampling, our scheme has the potential to break through the bandwidth barrier
of state-of-the-art analog conversion technologies such as interleaved
converters.Comment: 17 pages, 12 figures, to appear in IEEE Journal of Selected Topics in
Signal Processing, the special issue on Compressed Sensin
Estimation and Calibration Algorithms for Distributed Sampling Systems
Thesis Supervisor: Gregory W. Wornell
Title: Professor of Electrical Engineering and Computer ScienceTraditionally, the sampling of a signal is performed using a single component such as an
analog-to-digital converter. However, many new technologies are motivating the use of
multiple sampling components to capture a signal. In some cases such as sensor networks,
multiple components are naturally found in the physical layout; while in other cases like
time-interleaved analog-to-digital converters, additional components are added to increase
the sampling rate. Although distributing the sampling load across multiple channels can
provide large benefits in terms of speed, power, and resolution, a variety mismatch errors
arise that require calibration in order to prevent a degradation in system performance.
In this thesis, we develop low-complexity, blind algorithms for the calibration of distributed
sampling systems. In particular, we focus on recovery from timing skews that
cause deviations from uniform timing. Methods for bandlimited input reconstruction from
nonuniform recurrent samples are presented for both the small-mismatch and the low-SNR
domains. Alternate iterative reconstruction methods are developed to give insight into the
geometry of the problem.
From these reconstruction methods, we develop time-skew estimation algorithms that
have high performance and low complexity even for large numbers of components. We also
extend these algorithms to compensate for gain mismatch between sampling components.
To understand the feasibility of implementation, analysis is also presented for a sequential
implementation of the estimation algorithm.
In distributed sampling systems, the minimum input reconstruction error is dependent
upon the number of sampling components as well as the sample times of the components. We
develop bounds on the expected reconstruction error when the time-skews are distributed
uniformly. Performance is compared to systems where input measurements are made via
projections onto random bases, an alternative to the sinc basis of time-domain sampling.
From these results, we provide a framework on which to compare the effectiveness of any
calibration algorithm.
Finally, we address the topic of extreme oversampling, which pertains to systems with
large amounts of oversampling due to redundant sampling components. Calibration algorithms
are developed for ordering the components and for estimating the input from ordered
components. The algorithms exploit the extra samples in the system to increase estimation
performance and decrease computational complexity
Applications of nonuniform sampling in wideband multichannel communication systems
This research is an investigation into utilising randomised sampling in communication systems to ease the sampling rate requirements of digitally processing narrowband signals residing within a wide range of overseen frequencies. By harnessing the aliasing suppression capabilities of such sampling schemes, it is shown that certain processing tasks, namely spectrum sensing, can be performed at significantly low sampling rates compared to those demanded by uniform-sampling-based digital signal processing. The latter imposes sampling frequencies of at least twice the monitored bandwidth regardless of the spectral activity within. Aliasing can otherwise result in irresolvable processing problems, as the spectral support of the present signal is a priori unknown. Lower sampling rates exploit the processing module(s) resources (such as power) more efficiently and avoid the possible need for premium specialised high-cost DSP, especially if the handled bandwidth is considerably wide.
A number of randomised sampling schemes are examined and appropriate spectral analysis tools are used to furnish their salient features. The adopted periodogram-type estimators are tailored to each of the schemes and their statistical characteristics are assessed for stationary, and cyclostationary signals. Their ability to alleviate the bandwidth limitation of uniform sampling is demonstrated and the smeared-aliasing defect that accompanies randomised sampling is also quantified.
In employing the aforementioned analysis tools a novel wideband spectrum sensing approach is introduced. It permits the simultaneous sensing of a number of nonoverlapping spectral subbands constituting a wide range of monitored frequencies. The
operational sampling rates of the sensing procedure are not limited or dictated by the overseen bandwidth antithetical to uniform-sampling-based techniques. Prescriptive guidelines are developed to ensure that the proposed technique satisfies certain detection probabilities predefined by the user. These recommendations address the trade-off between the required sampling rate and the length of the signal observation window (sensing time) in a given scenario. Various aspects of the introduced multiband spectrum sensing approach are investigated and its applicability highlighted
Digitally-Assisted Mixed-Signal Wideband Compressive Sensing
Digitizing wideband signals requires very demanding analog-to-digital conversion (ADC) speed and resolution specifications. In this dissertation, a mixed-signal parallel compressive sensing system is proposed to realize the sensing of wideband sparse signals at sub-Nqyuist rate by exploiting the signal sparsity. The mixed-signal compressive sensing is realized with a parallel segmented compressive sensing (PSCS) front-end, which not only can filter out the harmonic spurs that leak from the local random generator, but also provides a tradeoff between the sampling rate and the system complexity such that a practical hardware implementation is possible. Moreover, the signal randomization in the
system is able to spread the spurious energy due to ADC nonlinearity along the signal bandwidth rather than concentrate on a few frequencies as it is the case for a conventional ADC. This important new property relaxes the ADC SFDR requirement when sensing frequency-domain
sparse signals.
The mixed-signal compressive sensing system performance is greatly impacted by the accuracy of analog circuit components, especially with the scaling of CMOS technology. In this dissertation, the effect of the circuit imperfection in the mixed-signal compressive
sensing system based on the PSCS front-end is investigated in detail, such as the finite settling
time, the timing uncertainty and so on. An iterative background calibration algorithm based on LMS (Least Mean Square) is proposed, which is shown to be able to effectively calibrate the error due to the circuit nonideal factors.
A low-speed prototype built with off-the-shelf components is presented. The prototype is able to sense sparse analog signals with up to 4 percent sparsity at 32 percent of the Nqyuist rate. Many practical constraints that arose during building the prototype such as circuit nonidealities are addressed in detail, which provides good insights for a future high-frequency integrated
circuit implementation. Based on that, a high-frequency sub-Nyquist rate receiver exploiting the parallel compressive sensing is designed and fabricated with IBM90nm CMOS technology, and measurement results are presented to show the capability of wideband
compressive sensing at sub-Nyquist rate. To the best of our knowledge, this prototype is the first reported integrated chip for wideband mixed-signal compressive sensing. The proposed prototype achieves 7 bits ENOB and 3 GS/s equivalent sampling rate in simulation assuming a 0.5 ps state-of-art jitter variance, whose FOM beats the FOM of the high speed state-of-the-art Nyquist ADCs by 2-3 times.
The proposed mixed-signal compressive sensing system can be applied in various fields. In particular, its applications for wideband spectrum sensing for cognitive radios and spectrum analysis in RF tests are discussed in this work
Novel Digital Alias-Free Signal Processing Approaches to FIR Filtering Estimation
This thesis aims at developing a new methodology of filtering continuous-time bandlimited signals and piecewise-continuous signals from their discrete-time samples. Unlike the existing state-of-the-art filters, my filters are not adversely affected by aliasing, allowing the designers to flexibly select the sampling rates of the processed signal to reach the required accuracy of signal filtering rather than meeting stiff and often demanding constraints imposed by the classical theory of digital signal processing (DSP). The impact of this thesis is cost reduction of alias-free sampling, filtering and other digital processing blocks, particularly when the processed signals have sparse and unknown spectral support.
Novel approaches are proposed which can mitigate the negative effects of aliasing, thanks to the use of nonuniform random/pseudorandom sampling and processing algorithms. As such, the proposed approaches belong to the family of digital alias-free signal processing (DASP). Namely, three main approaches are considered: total random (ToRa), stratified (StSa) and antithetical stratified (AnSt) random sampling techniques.
First, I introduce a finite impulse response (FIR) filter estimator for each of the three considered techniques. In addition, a generalised estimator that encompasses the three filter estimators is also proposed. Then, statistical properties of all estimators are investigated to assess their quality. Properties such as expected value, bias, variance, convergence rate, and consistency are all inspected and unveiled. Moreover, closed-form mathematical expression is devised for the variance of each single estimator.
Furthermore, quality assessment of the proposed estimators is examined in two main cases related to the smoothness status of the filter convolution’s integrand function, \u1d454(\u1d461,\u1d70f)∶=\u1d465(\u1d70f)ℎ(\u1d461−\u1d70f), and its first two derivatives. The first main case is continuous and differentiable functions \u1d454(\u1d461,\u1d70f), \u1d454′(\u1d461,\u1d70f), and \u1d454′′(\u1d461,\u1d70f). Whereas in the second main case, I cover all possible instances where some/all of such functions are piecewise-continuous and involving a finite number of bounded discontinuities.
Primarily obtained results prove that all considered filter estimators are unbiassed and consistent. Hence, variances of the estimators converge to zero after certain number of sample points. However, the convergence rate depends on the selected estimator and which case of smoothness is being considered.
In the first case (i.e. continuous \u1d454(\u1d461,\u1d70f) and its derivatives), ToRa, StSa and AnSt filter estimators converge uniformly at rates of \u1d441−1, \u1d441−3, and \u1d441−5 respectively, where 2\u1d441 is the total number of sample points. More interestingly, in the second main case, the convergence rates of StSa and AnSt estimators are maintained even if there are some discontinuities in the first-order derivative (FOD) with respect to \u1d70f of \u1d454(\u1d461,\u1d70f) (for StSa estimator) or in the second-order derivative (SOD) with respect to \u1d70f of \u1d454(\u1d461,\u1d70f) (for AnSt). Whereas these rates drop to \u1d441−2 and \u1d441−4 (for StSa and AnSt, respectively) if the zero-order derivative (ZOD) (for StSa) and FOD (for AnSt) are piecewise-continuous. Finally, if the ZOD of \u1d454(\u1d461,\u1d70f) is piecewise-continuous, then the uniform convergence rate of the AnSt estimator further drops to \u1d441−2.
For practical reasons, I also introduce the utilisation of the three estimators in a special situation where the input signal is pseudorandomly sampled from otherwise uniform and dense grid. An FIR filter model with an oversampled finite-duration impulse response, timely aligned with the grid, is proposed and meant to be stored in a lookup table of the implemented filter’s memory to save processing time. Then, a synchronised convolution sum operation is conducted to estimate the filter output.
Finally, a new unequally spaced Lagrange interpolation-based rule is proposed. The so-called composite 3-nonuniform-sample (C3NS) rule is employed to estimate area under the curve (AUC) of an integrand function rather than the simple Rectangular rule. I then carry out comparisons for the convergence rates of different estimators based on the two interpolation rules. The proposed C3NS estimator outperforms other Rectangular rule estimators on the expense of higher computational complexity. Of course, this extra cost could only be justifiable for some specific applications where more accurate estimation is required
Estimation of Time-Limited Channel Spectra From Nonuniform Samples
This paper deals with the estimation of a time-invariant channel spectrum from its own nonuniform samples, assuming there is a bound on the channel’s delay spread. Except for this last assumption, this is the basic estimation problem in systems providing channel spectral samples. However, as shown in the paper, the delay spread bound leads us to view the spectrum as a band-limited signal, rather than the Fourier transform of a tapped delay line (TDL). Using this alternative model, a linear estimator is presented that approximately minimizes the expected root-mean-square (RMS) error for a deterministic channel. Its main advantage over the TDL is that it takes into account the spectrum’s smoothness (time width), thus providing a performance improvement. The proposed estimator is compared numerically with the maximum likelihood (ML) estimator based on a TDL model in pilot-assisted channel estimation (PACE) for OFDM.This work was supported by the Spanish Ministry of Economy and Competitiveness (MINECO) under Project TEC2011-28201-C02-02
Recommended from our members
A multi-sensor sub-Nyquist power spectrum blind sampling approach for low-power wireless sensors in operational modal analysis applications
A novel multi-sensor power spectrum blind sampling (PSBS) approach is proposed supporting low-power wireless sensor networks (WSN) for Operational Modal Analysis (OMA) applications. The developed approach relies on arrays of wireless sensors, employing deterministic non-uniform in time multi-coset sampling to acquire structural response acceleration signals at sub-Nyquist sampling rates, treated as realizations of stationary random processes without making any assumption about the average signal frequency content and spectral support. The acquired compressed measurements are transmitted to a central server and collectively processed via a PSBS technique, herein extended to the multi-sensor case, to estimate the power spectral density matrix of an underlying spatially correlated stationary response acceleration random process directly from the compressed measurements. Structural modal properties are then extracted through standard frequency domain decomposition (FDD). The efficacy of the proposed approach to resolve closely-spaced modes is numerically tested for various data compression levels using noisy response acceleration signals of a white-noise excited finite element model of a space truss as well as field-recorded acceleration time-histories of an instrumented bridge under operational loading. It is shown that accurate mode shapes based on the modal assurance criterion can be obtained from as low as 89% less measurements compared to conventional non-compressive FDD at Nyquist sampling rate. Further, significant gains in energy consumption and battery lifetime prolongation of the order of years are estimated, assuming wireless sensors operating on multi-coset sampling at different data compression levels. It is, therefore, concluded that the proposed PSBS approach could provide long-term structural health monitoring systems with low-maintenance cost once wireless sensors with multi-coset sampling capabilities become commercially available