40 research outputs found
An efficient JPEG-2000 based multimodal compression scheme
In this paper, a wavelet-based multimodal compression method is proposed. The method jointly compresses a medical image and an ECG signal within a single codec, i.e., JPEG-2000 in an effective and simple way. The multimodal scheme operates in two main stages: the first stage, consists of the encoder and involves a mixing function, aiming at inserting the samples of the signal in the image according to a predefined insertion pattern in the wavelet domain. The second stage represented by a separation function, consists of the extraction process of the ECG signal from the image after performing the decoding stage. Both the cubic spline and the median edge detection (MED) predictor have been adopted to conduct the interpolation process for estimating image pixels
Scalable and perceptual audio compression
This thesis deals with scalable perceptual audio compression. Two scalable perceptual solutions as well as a scalable to lossless solution are proposed and investigated. One of the scalable perceptual solutions is built around sinusoidal modelling of the audio signal whilst the other is built on a transform coding paradigm. The scalable coders are shown to scale both in a waveform matching manner as well as a psychoacoustic manner. In order to measure the psychoacoustic scalability of the systems investigated in this thesis, the similarity between the original signal\u27s psychoacoustic parameters and that of the synthesized signal are compared. The psychoacoustic parameters used are loudness, sharpness, tonahty and roughness. This analysis technique is a novel method used in this thesis and it allows an insight into the perceptual distortion that has been introduced by any coder analyzed in this manner
Progressive contour coding in the wavelet domain
This paper presents a new wavelet-based image contour coding technique, suitable for representing either shapes or generic contour maps. Starting from a contour map (e.g. a segmentation map or the result of a contour extraction operator), this is first converted in a one-dimensional signal. Coordinate jumps among different contour extremities are converted, if under a suitable threshold, into signal discontinuities which can be compactly represented in the wavelet domain. Otherwise, the exceeding discontinuities are coded as side information. This side information is minimized by an optimized contour segment sequencing. The obtained 1D signal is decomposed and coded in the wavelet domain by using a 1D version of an improved implementation of the SPIHT algorithm. This technique can efficiently code every kind of 2D contour map, from one to many unconnected contour segments. It guarantees a fully embedded progressive coding, state-of-art coding performance, good approximation capabilities for both open and closed contours, and visually graceful degradation at low bit-rates
Discrete Wavelet Transforms
The discrete wavelet transform (DWT) algorithms have a firm position in processing of signals in several areas of research and industry. As DWT provides both octave-scale frequency and spatial timing of the analyzed signal, it is constantly used to solve and treat more and more advanced problems. The present book: Discrete Wavelet Transforms: Algorithms and Applications reviews the recent progress in discrete wavelet transform algorithms and applications. The book covers a wide range of methods (e.g. lifting, shift invariance, multi-scale analysis) for constructing DWTs. The book chapters are organized into four major parts. Part I describes the progress in hardware implementations of the DWT algorithms. Applications include multitone modulation for ADSL and equalization techniques, a scalable architecture for FPGA-implementation, lifting based algorithm for VLSI implementation, comparison between DWT and FFT based OFDM and modified SPIHT codec. Part II addresses image processing algorithms such as multiresolution approach for edge detection, low bit rate image compression, low complexity implementation of CQF wavelets and compression of multi-component images. Part III focuses watermaking DWT algorithms. Finally, Part IV describes shift invariant DWTs, DC lossless property, DWT based analysis and estimation of colored noise and an application of the wavelet Galerkin method. The chapters of the present book consist of both tutorial and highly advanced material. Therefore, the book is intended to be a reference text for graduate students and researchers to obtain state-of-the-art knowledge on specific applications
Progressive contour coding in the wavelet domain
This paper presents a new wavelet-based image contour coding technique, suitable for representing either shapes or generic contour maps. Starting from a contour map (e.g. a segmentation map or the result of an edge detector process), a unique one-dimensional signal is generated from the set of contour points. Coordinate jumps between contour extremities when under a tolerance threshold represent signal discontinuities but they can still be compactly coded in the wavelet domain. Exceeding threshold discontinuities are coded as side information. This side information and the amount of remaining discontinuity are minimized by an optimized contour segment sequencing. The obtained 1D signal is decomposed and coded in the wavelet domain by using a 1D extension of the SPIHT algorithm. The described technique can efficiently code any kind of 2D contour map, from one to many unconnected contour segments. It guarantees a fully embedded progressive coding, state-of-art coding performance, good approximation capabilities for both open and closed contours, and graceful visual degradation at low bit-rates
Centralized and distributed semi-parametric compression of piecewise smooth functions
This thesis introduces novel wavelet-based semi-parametric centralized and distributed
compression methods for a class of piecewise smooth functions. Our proposed compression schemes are based on a non-conventional transform coding structure with simple
independent encoders and a complex joint decoder.
Current centralized state-of-the-art compression schemes are based on the conventional structure where an encoder is relatively complex and nonlinear. In addition, the
setting usually allows the encoder to observe the entire source. Recently, there has been
an increasing need for compression schemes where the encoder is lower in complexity
and, instead, the decoder has to handle more computationally intensive tasks. Furthermore, the setup may involve multiple encoders, where each one can only partially
observe the source. Such scenario is often referred to as distributed source coding.
In the first part, we focus on the dual situation of the centralized compression where
the encoder is linear and the decoder is nonlinear. Our analysis is centered around a
class of 1-D piecewise smooth functions. We show that, by incorporating parametric
estimation into the decoding procedure, it is possible to achieve the same distortion-
rate performance as that of a conventional wavelet-based compression scheme. We also
present a new constructive approach to parametric estimation based on the sampling
results of signals with finite rate of innovation.
The second part of the thesis focuses on the distributed compression scenario, where
each independent encoder partially observes the 1-D piecewise smooth function. We
propose a new wavelet-based distributed compression scheme that uses parametric estimation to perform joint decoding. Our distortion-rate analysis shows that it is possible
for the proposed scheme to achieve that same compression performance as that of a
joint encoding scheme.
Lastly, we apply the proposed theoretical framework in the context of distributed
image and video compression. We start by considering a simplified model of the video
signal and show that we can achieve distortion-rate performance close to that of a joint
encoding scheme. We then present practical compression schemes for real world signals.
Our simulations confirm the improvement in performance over classical schemes, both
in terms of the PSNR and the visual quality
Progressively communicating rich telemetry from autonomous underwater vehicles via relays
Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy at the Massachusetts Institute of Technology and the Woods Hole Oceanographic Institution June 2012As analysis of imagery and environmental data plays a greater role in mission construction
and execution, there is an increasing need for autonomous marine vehicles
to transmit this data to the surface. Without access to the data acquired by a
vehicle, surface operators cannot fully understand the state of the mission. Communicating
imagery and high-resolution sensor readings to surface observers remains
a significant challenge – as a result, current telemetry from free-roaming
autonomous marine vehicles remains limited to ‘heartbeat’ status messages, with
minimal scientific data available until after recovery. Increasing the challenge, longdistance
communication may require relaying data across multiple acoustic hops
between vehicles, yet fixed infrastructure is not always appropriate or possible.
In this thesis I present an analysis of the unique considerations facing telemetry
systems for free-roaming Autonomous Underwater Vehicles (AUVs) used in exploration.
These considerations include high-cost vehicle nodes with persistent storage
and significant computation capabilities, combined with human surface operators
monitoring each node. I then propose mechanisms for interactive, progressive
communication of data across multiple acoustic hops. These mechanisms include
wavelet-based embedded coding methods, and a novel image compression scheme
based on texture classification and synthesis. The specific characteristics of underwater
communication channels, including high latency, intermittent communication,
the lack of instantaneous end-to-end connectivity, and a broadcast medium,
inform these proposals. Human feedback is incorporated by allowing operators to
identify segments of data thatwarrant higher quality refinement, ensuring efficient
use of limited throughput. I then analyze the performance of these mechanisms
relative to current practices.
Finally, I present CAPTURE, a telemetry architecture that builds on this analysis.
CAPTURE draws on advances in compression and delay tolerant networking to
enable progressive transmission of scientific data, including imagery, across multiple acoustic hops. In concert with a physical layer, CAPTURE provides an endto-
end networking solution for communicating science data from autonomous marine
vehicles. Automatically selected imagery, sonar, and time-series sensor data
are progressively transmitted across multiple hops to surface operators. Human
operators can request arbitrarily high-quality refinement of any resource, up to an
error-free reconstruction. The components of this system are then demonstrated
through three field trials in diverse environments on SeaBED, OceanServer and
Bluefin AUVs, each in different software architectures.Thanks to the National Science Foundation, and the
National Oceanic and Atmospheric Administration for
their funding of my education and this work
Selection of Wavelet Basis Function for Image Compression : a Review
Wavelets are being suggested as a platform for various tasks in image processing. The advantage of wavelets lie in its time frequency resolution. The use of different basis functions in the form of different wavelets made the wavelet analysis as a destination for many applications. The performance of a particular technique depends on the wavelet coefficients arrived after applying the wavelet transform. The coefficients for a specific input signal depends on the basis functions used in the wavelet transform. Hence in this paper toward this end, different basis functions and their features are presented. As the image compression task depends on wavelet transform to large extent from few decades, the selection of basis function for image compression should be taken with care. In this paper, the factors influencing the performance of image compression are presented
MP3D: Highly Scalable Video Coding Scheme Based on Matching Pursuit
This paper describes a novel video coding scheme based on a three-dimensional Matching Pursuit algorithm. In addition to good compression performance at low bit rate, the proposed coder allows for flexible spatial, temporal and rate scalability thanks to its progressive coding structure. The Matching Pursuit algorithm generates a sparse decomposition of a video sequence in a series of spatio-temporal atoms, taken from an overcomplete dictionary of three-dimensional basis functions. The dictionary is generated by shifting, scaling and rotating two different mother atoms in order to cover the whole frequency cube. An embedded stream is then produced from the series of atoms. They are first distributed into sets through the set-partitioned position map algorithm (SPPM) to form the index-map, inspired from bit plane encoding. Scalar quantization is then applied to the coefficients which are finally arithmetic coded. A complete MP3D codec has been implemented, and performances are shown to favorably compare to other scalable coders like MPEG-4 FGS and SPIHT-3D. In addition, the MP3D streams offer an incomparable flexibility for multiresolution streaming or adaptive decoding