1,046 research outputs found
A Hierarchical Bayesian Model for Frame Representation
In many signal processing problems, it may be fruitful to represent the
signal under study in a frame. If a probabilistic approach is adopted, it
becomes then necessary to estimate the hyper-parameters characterizing the
probability distribution of the frame coefficients. This problem is difficult
since in general the frame synthesis operator is not bijective. Consequently,
the frame coefficients are not directly observable. This paper introduces a
hierarchical Bayesian model for frame representation. The posterior
distribution of the frame coefficients and model hyper-parameters is derived.
Hybrid Markov Chain Monte Carlo algorithms are subsequently proposed to sample
from this posterior distribution. The generated samples are then exploited to
estimate the hyper-parameters and the frame coefficients of the target signal.
Validation experiments show that the proposed algorithms provide an accurate
estimation of the frame coefficients and hyper-parameters. Application to
practical problems of image denoising show the impact of the resulting Bayesian
estimation on the recovered signal quality
Noise Removal in Microarray Images Using Variational Mode Decomposition Technique
Microarray technology allows the simultaneous monitoring of thousands of genes in parallel. Based on the gene expression measurements, microarray technology have proven powerful in gene expression profiling for discovering new types of diseases and for predicting the type of a disease. Enhancement, Gridding, Segmentation and Intensity extraction are important steps in microarray image analysis. This paper presents a noise removal method in microarray images based on Variational Mode Decomposition (VMD). VMD is a signal processing method which decomposes any input signal into discrete number of sub-signals (called Variational Mode Functions) with each mode chosen to be its band width in spectral domain. First the noisy image is processed using 2-D VMD to produce 2-D VMFs. Then Discrete Wavelet Transform (DWT) thresholding technique is applied to each VMF for denoising. The denoised microarray image is reconstructed by the summation of VMFs. This method is named as 2-D VMD and DWT thresholding method. The proposed method is compared with DWT thresholding and BEMD and DWT thresholding methods. The qualitative and quantitative analysis shows that 2-D VMD and DWT thresholding method produces better noise removal than other two methods
Combining local regularity estimation and total variation optimization for scale-free texture segmentation
Texture segmentation constitutes a standard image processing task, crucial to
many applications. The present contribution focuses on the particular subset of
scale-free textures and its originality resides in the combination of three key
ingredients: First, texture characterization relies on the concept of local
regularity ; Second, estimation of local regularity is based on new multiscale
quantities referred to as wavelet leaders ; Third, segmentation from local
regularity faces a fundamental bias variance trade-off: In nature, local
regularity estimation shows high variability that impairs the detection of
changes, while a posteriori smoothing of regularity estimates precludes from
locating correctly changes. Instead, the present contribution proposes several
variational problem formulations based on total variation and proximal
resolutions that effectively circumvent this trade-off. Estimation and
segmentation performance for the proposed procedures are quantified and
compared on synthetic as well as on real-world textures
Machine Learning-Based Classification of Pulmonary Diseases through Real-Time Lung Sounds
The study presents a computer-based automated system that employs machine learning to classify pulmonary diseases using lung sound data collected from hospitals. Denoising techniques, such as discrete wavelet transform and variational mode decomposition, are applied to enhance classifier performance. The system combines cepstral features, such as Mel-frequency cepstrum coefficients and gammatone frequency cepstral coefficients, for classification. Four machine learning classifiers, namely the decision tree, k-nearest neighbor, linear discriminant analysis, and random forest, are compared. Evaluation metrics such as accuracy, recall, specificity, and f1 score are employed. This study includes patients affected by chronic obstructive pulmonary disease, asthma, bronchiectasis, and healthy individuals. The results demonstrate that the random forest classifier outperforms the others, achieving an accuracy of 99.72% along with 100% recall, specificity, and f1 scores. The study suggests that the computer-based system serves as a decision-making tool for classifying pulmonary diseases, especially in resource-limited settings
First order algorithms in variational image processing
Variational methods in imaging are nowadays developing towards a quite
universal and flexible tool, allowing for highly successful approaches on tasks
like denoising, deblurring, inpainting, segmentation, super-resolution,
disparity, and optical flow estimation. The overall structure of such
approaches is of the form ; where the functional is a data fidelity term also
depending on some input data and measuring the deviation of from such
and is a regularization functional. Moreover is a (often linear)
forward operator modeling the dependence of data on an underlying image, and
is a positive regularization parameter. While is often
smooth and (strictly) convex, the current practice almost exclusively uses
nonsmooth regularization functionals. The majority of successful techniques is
using nonsmooth and convex functionals like the total variation and
generalizations thereof or -norms of coefficients arising from scalar
products with some frame system. The efficient solution of such variational
problems in imaging demands for appropriate algorithms. Taking into account the
specific structure as a sum of two very different terms to be minimized,
splitting algorithms are a quite canonical choice. Consequently this field has
revived the interest in techniques like operator splittings or augmented
Lagrangians. Here we shall provide an overview of methods currently developed
and recent results as well as some computational studies providing a comparison
of different methods and also illustrating their success in applications.Comment: 60 pages, 33 figure
Data-driven Signal Decomposition Approaches: A Comparative Analysis
Signal decomposition (SD) approaches aim to decompose non-stationary signals
into their constituent amplitude- and frequency-modulated components. This
represents an important preprocessing step in many practical signal processing
pipelines, providing useful knowledge and insight into the data and relevant
underlying system(s) while also facilitating tasks such as noise or artefact
removal and feature extraction. The popular SD methods are mostly data-driven,
striving to obtain inherent well-behaved signal components without making many
prior assumptions on input data. Among those methods include empirical mode
decomposition (EMD) and variants, variational mode decomposition (VMD) and
variants, synchrosqueezed transform (SST) and variants and sliding singular
spectrum analysis (SSA). With the increasing popularity and utility of these
methods in wide-ranging application, it is imperative to gain a better
understanding and insight into the operation of these algorithms, evaluate
their accuracy with and without noise in input data and gauge their sensitivity
against algorithmic parameter changes. In this work, we achieve those tasks
through extensive experiments involving carefully designed synthetic and
real-life signals. Based on our experimental observations, we comment on the
pros and cons of the considered SD algorithms as well as highlighting the best
practices, in terms of parameter selection, for the their successful operation.
The SD algorithms for both single- and multi-channel (multivariate) data fall
within the scope of our work. For multivariate signals, we evaluate the
performance of the popular algorithms in terms of fulfilling the mode-alignment
property, especially in the presence of noise.Comment: Resubmission with changes in the reference lis
- …