3,717 research outputs found
Compressive Sensing for Spectroscopy and Polarimetry
We demonstrate through numerical simulations with real data the feasibility
of using compressive sensing techniques for the acquisition of
spectro-polarimetric data. This allows us to combine the measurement and the
compression process into one consistent framework. Signals are recovered thanks
to a sparse reconstruction scheme from projections of the signal of interest
onto appropriately chosen vectors, typically noise-like vectors. The
compressibility properties of spectral lines are analyzed in detail. The
results shown in this paper demonstrate that, thanks to the compressibility
properties of spectral lines, it is feasible to reconstruct the signals using
only a small fraction of the information that is measured nowadays. We
investigate in depth the quality of the reconstruction as a function of the
amount of data measured and the influence of noise. This change of paradigm
also allows us to define new instrumental strategies and to propose
modifications to existing instruments in order to take advantage of compressive
sensing techniques.Comment: 11 pages, 9 figures, accepted for publication in A&
Imaging via Compressive Sampling [Introduction to compressive sampling and recovery via convex programming]
There is an extensive body of literature on image compression, but the central concept is straightforward: we transform the image into an appropriate basis and then code only the important expansion coefficients. The crux is finding a good transform, a problem that has been studied extensively from both a theoretical [14] and practical [25] standpoint. The most notable product of this research is the wavelet transform [9], [16]; switching from sinusoid-based representations to wavelets marked a watershed in image compression and is the essential difference between the classical JPEG [18] and modern JPEG-2000 [22] standards.
Image compression algorithms convert high-resolution images into a relatively small bit streams (while keeping the essential features intact), in effect turning a large digital data set into a substantially smaller one. But is there a way to avoid the large digital data set to begin with? Is there a way we can build the data compression directly into the acquisition? The answer is yes, and is what compressive sampling (CS) is all about
Towards the text compression based feature extraction in high impedance fault detection
High impedance faults of medium voltage overhead lines with covered conductors can be identified by the presence of partial discharges. Despite it is a subject of research for more than 60 years, online partial discharges detection is always a challenge, especially in environment with heavy background noise. In this paper, a new approach for partial discharge pattern recognition is presented. All results were obtained on data, acquired from real 22 kV medium voltage overhead power line with covered conductors. The proposed method is based on a text compression algorithm and it serves as a signal similarity estimation, applied for the first time on partial discharge pattern. Its relevancy is examined by three different variations of classification model. The improvement gained on an already deployed model proves its quality.Web of Science1211art. no. 214
Compressive Mining: Fast and Optimal Data Mining in the Compressed Domain
Real-world data typically contain repeated and periodic patterns. This
suggests that they can be effectively represented and compressed using only a
few coefficients of an appropriate basis (e.g., Fourier, Wavelets, etc.).
However, distance estimation when the data are represented using different sets
of coefficients is still a largely unexplored area. This work studies the
optimization problems related to obtaining the \emph{tightest} lower/upper
bound on Euclidean distances when each data object is potentially compressed
using a different set of orthonormal coefficients. Our technique leads to
tighter distance estimates, which translates into more accurate search,
learning and mining operations \textit{directly} in the compressed domain.
We formulate the problem of estimating lower/upper distance bounds as an
optimization problem. We establish the properties of optimal solutions, and
leverage the theoretical analysis to develop a fast algorithm to obtain an
\emph{exact} solution to the problem. The suggested solution provides the
tightest estimation of the -norm or the correlation. We show that typical
data-analysis operations, such as k-NN search or k-Means clustering, can
operate more accurately using the proposed compression and distance
reconstruction technique. We compare it with many other prevalent compression
and reconstruction techniques, including random projections and PCA-based
techniques. We highlight a surprising result, namely that when the data are
highly sparse in some basis, our technique may even outperform PCA-based
compression.
The contributions of this work are generic as our methodology is applicable
to any sequential or high-dimensional data as well as to any orthogonal data
transformation used for the underlying data compression scheme.Comment: 25 pages, 20 figures, accepted in VLD
Performance of Ti-6242 production using nano powder mixed with different dielectrics
The recent scenario of modern manufacturing is tremendously improved in the sense of precision
machining and abstaining from environmental pollution and hazard issues. In the present work, Ti�6242 is machined through wire EDM (WEDM) process with powder mixed dielectric and analyzed the
influence of input parameters and inherent hazard issues. WEDM has different parameters such as peak
current, pulse on time, pulse off time, gap voltage, wire-speed, wire tension, and so on, as well as dielec�trics with powder mixed. These are playing an essential role in WEDM performances to improve the pro�cess efficiency by developing the metal removal rate. Even though the parameter’s influencing, the study
of the nano dielectric effect in the WEDM process is very essential during the machining process due to
the high discharge energy. In the present study, two different dielectric fluids were used, including deio�nised water, and nanopowder dielectrically, and analyzed the data by taking the response surface method
to use program design expert 10. From this study, it is established that dielectric types and powder sig�nificantly improve performances with a proper set of machining parameters and find out the risk factor
associated with the WEDM process
Non-parametric PSF estimation from celestial transit solar images using blind deconvolution
Context: Characterization of instrumental effects in astronomical imaging is
important in order to extract accurate physical information from the
observations. The measured image in a real optical instrument is usually
represented by the convolution of an ideal image with a Point Spread Function
(PSF). Additionally, the image acquisition process is also contaminated by
other sources of noise (read-out, photon-counting). The problem of estimating
both the PSF and a denoised image is called blind deconvolution and is
ill-posed.
Aims: We propose a blind deconvolution scheme that relies on image
regularization. Contrarily to most methods presented in the literature, our
method does not assume a parametric model of the PSF and can thus be applied to
any telescope.
Methods: Our scheme uses a wavelet analysis prior model on the image and weak
assumptions on the PSF. We use observations from a celestial transit, where the
occulting body can be assumed to be a black disk. These constraints allow us to
retain meaningful solutions for the filter and the image, eliminating trivial,
translated and interchanged solutions. Under an additive Gaussian noise
assumption, they also enforce noise canceling and avoid reconstruction
artifacts by promoting the whiteness of the residual between the blurred
observations and the cleaned data.
Results: Our method is applied to synthetic and experimental data. The PSF is
estimated for the SECCHI/EUVI instrument using the 2007 Lunar transit, and for
SDO/AIA using the 2012 Venus transit. Results show that the proposed
non-parametric blind deconvolution method is able to estimate the core of the
PSF with a similar quality to parametric methods proposed in the literature. We
also show that, if these parametric estimations are incorporated in the
acquisition model, the resulting PSF outperforms both the parametric and
non-parametric methods.Comment: 31 pages, 47 figure
- …