544,244 research outputs found
Bayesian Nonparametric Dictionary Learning for Compressed Sensing MRI
We develop a Bayesian nonparametric model for reconstructing magnetic
resonance images (MRI) from highly undersampled k-space data. We perform
dictionary learning as part of the image reconstruction process. To this end,
we use the beta process as a nonparametric dictionary learning prior for
representing an image patch as a sparse combination of dictionary elements. The
size of the dictionary and the patch-specific sparsity pattern are inferred
from the data, in addition to other dictionary learning variables. Dictionary
learning is performed directly on the compressed image, and so is tailored to
the MRI being considered. In addition, we investigate a total variation penalty
term in combination with the dictionary learning model, and show how the
denoising property of dictionary learning removes dependence on regularization
parameters in the noisy setting. We derive a stochastic optimization algorithm
based on Markov Chain Monte Carlo (MCMC) for the Bayesian model, and use the
alternating direction method of multipliers (ADMM) for efficiently performing
total variation minimization. We present empirical results on several MRI,
which show that the proposed regularization framework can improve
reconstruction accuracy over other methods
Task-Driven Dictionary Learning for Hyperspectral Image Classification with Structured Sparsity Constraints
Sparse representation models a signal as a linear combination of a small
number of dictionary atoms. As a generative model, it requires the dictionary
to be highly redundant in order to ensure both a stable high sparsity level and
a low reconstruction error for the signal. However, in practice, this
requirement is usually impaired by the lack of labelled training samples.
Fortunately, previous research has shown that the requirement for a redundant
dictionary can be less rigorous if simultaneous sparse approximation is
employed, which can be carried out by enforcing various structured sparsity
constraints on the sparse codes of the neighboring pixels. In addition,
numerous works have shown that applying a variety of dictionary learning
methods for the sparse representation model can also improve the classification
performance. In this paper, we highlight the task-driven dictionary learning
algorithm, which is a general framework for the supervised dictionary learning
method. We propose to enforce structured sparsity priors on the task-driven
dictionary learning method in order to improve the performance of the
hyperspectral classification. Our approach is able to benefit from both the
advantages of the simultaneous sparse representation and those of the
supervised dictionary learning. We enforce two different structured sparsity
priors, the joint and Laplacian sparsity, on the task-driven dictionary
learning method and provide the details of the corresponding optimization
algorithms. Experiments on numerous popular hyperspectral images demonstrate
that the classification performance of our approach is superior to sparse
representation classifier with structured priors or the task-driven dictionary
learning method
Weakly-supervised Dictionary Learning
We present a probabilistic modeling and inference framework for
discriminative analysis dictionary learning under a weak supervision setting.
Dictionary learning approaches have been widely used for tasks such as
low-level signal denoising and restoration as well as high-level classification
tasks, which can be applied to audio and image analysis. Synthesis dictionary
learning aims at jointly learning a dictionary and corresponding sparse
coefficients to provide accurate data representation. This approach is useful
for denoising and signal restoration, but may lead to sub-optimal
classification performance. By contrast, analysis dictionary learning provides
a transform that maps data to a sparse discriminative representation suitable
for classification. We consider the problem of analysis dictionary learning for
time-series data under a weak supervision setting in which signals are assigned
with a global label instead of an instantaneous label signal. We propose a
discriminative probabilistic model that incorporates both label information and
sparsity constraints on the underlying latent instantaneous label signal using
cardinality control. We present the expectation maximization (EM) procedure for
maximum likelihood estimation (MLE) of the proposed model. To facilitate a
computationally efficient E-step, we propose both a chain and a novel tree
graph reformulation of the graphical model. The performance of the proposed
model is demonstrated on both synthetic and real-world data
Pooling-Invariant Image Feature Learning
Unsupervised dictionary learning has been a key component in state-of-the-art
computer vision recognition architectures. While highly effective methods exist
for patch-based dictionary learning, these methods may learn redundant features
after the pooling stage in a given early vision architecture. In this paper, we
offer a novel dictionary learning scheme to efficiently take into account the
invariance of learned features after the spatial pooling stage. The algorithm
is built on simple clustering, and thus enjoys efficiency and scalability. We
discuss the underlying mechanism that justifies the use of clustering
algorithms, and empirically show that the algorithm finds better dictionaries
than patch-based methods with the same dictionary size
- …
