371 research outputs found
Investigation of feature extraction algorithms and techniques for hyperspectral images.
Doctor of Philosophy (Computer Engineering). University of KwaZulu-Natal. Durban, 2017.Hyperspectral images (HSIs) are remote-sensed images that are characterized
by very high spatial and spectral dimensions and nd applications, for example,
in land cover classi cation, urban planning and management, security and food
processing. Unlike conventional three bands RGB images, their high
dimensional data space creates a challenge for traditional image processing
techniques which are usually based on the assumption that there exists
su cient training samples in order to increase the likelihood of high
classi cation accuracy. However, the high cost and di culty of obtaining
ground truth of hyperspectral data sets makes this assumption unrealistic and
necessitates the introduction of alternative methods for their processing.
Several techniques have been developed in the exploration of the rich spectral
and spatial information in HSIs. Speci cally, feature extraction (FE)
techniques are introduced in the processing of HSIs as a necessary step before
classi cation. They are aimed at transforming the high dimensional data of the
HSI into one of a lower dimension while retaining as much spatial and/or
spectral information as possible. In this research, we develop semi-supervised
FE techniques which combine features of supervised and unsupervised
techniques into a single framework for the processing of HSIs. Firstly, we
developed a feature extraction algorithm known as Semi-Supervised Linear
Embedding (SSLE) for the extraction of features in HSI. The algorithm
combines supervised Linear Discriminant Analysis (LDA) and unsupervised
Local Linear Embedding (LLE) to enhance class discrimination while also
preserving the properties of classes of interest. The technique was developed
based on the fact that LDA extracts features from HSIs by discriminating
between classes of interest and it can only extract C 1 features provided there
are C classes in the image by extracting features that are equivalent to the
number of classes in the HSI. Experiments show that the SSLE algorithm
overcomes the limitation of LDA and extracts features that are equivalent to
ii
iii
the number of classes in HSIs. Secondly, a graphical manifold dimension
reduction (DR) algorithm known as Graph Clustered Discriminant Analysis
(GCDA) is developed. The algorithm is developed to dynamically select labeled
samples from the pool of available unlabeled samples in order to complement
the few available label samples in HSIs. The selection is achieved by entwining
K-means clustering with a semi-supervised manifold discriminant analysis.
Using two HSI data sets, experimental results show that GCDA extracts
features that are equivalent to the number of classes with high classi cation
accuracy when compared with other state-of-the-art techniques. Furthermore,
we develop a window-based partitioning approach to preserve the spatial
properties of HSIs when their features are being extracted. In this approach,
the HSI is partitioned along its spatial dimension into n windows and the
covariance matrices of each window are computed. The covariance matrices of
the windows are then merged into a single matrix through using the Kalman
ltering approach so that the resulting covariance matrix may be used for
dimension reduction. Experiments show that the windowing approach achieves
high classi cation accuracy and preserves the spatial properties of HSIs. For
the proposed feature extraction techniques, Support Vector Machine (SVM)
and Neural Networks (NN) classi cation techniques are employed and their
performances are compared for these two classi ers. The performances of all
proposed FE techniques have also been shown to outperform other
state-of-the-art approaches
Algorithms for feature selection and pattern recognition on Grassmann manifolds
Includes bibliographical references.2015 Summer.This dissertation presents three distinct application-driven research projects united by ideas and topics from geometric data analysis, optimization, computational topology, and machine learning. We first consider hyperspectral band selection problem solved by using sparse support vector machines (SSVMs). A supervised embedded approach is proposed using the property of SSVMs to exhibit a model structure that includes a clearly identifiable gap between zero and non-zero feature vector weights that permits important bands to be definitively selected in conjunction with the classification problem. An SSVM is trained using bootstrap aggregating to obtain a sample of SSVM models to reduce variability in the band selection process. This preliminary sample approach for band selection is followed by a secondary band selection which involves retraining the SSVM to further reduce the set of bands retained. We propose and compare three adaptations of the SSVM band selection algorithm for the multiclass problem. We illustrate the performance of these methods on two benchmark hyperspectral data sets. Second, we propose an approach for capturing the signal variability in data using the framework of the Grassmann manifold (Grassmannian). Labeled points from each class are sampled and used to form abstract points on the Grassmannian. The resulting points have representations as orthonormal matrices and as such do not reside in Euclidean space in the usual sense. There are a variety of metrics which allow us to determine distance matrices that can be used to realize the Grassmannian as an embedding in Euclidean space. Multidimensional scaling (MDS) determines a low dimensional Euclidean embedding of the manifold, preserving or approximating the Grassmannian geometry based on the distance measure. We illustrate that we can achieve an isometric embedding of the Grassmann manifold using the chordal metric while this is not the case with other distances. However, non-isometric embeddings generated by using the smallest principal angle pseudometric on the Grassmannian lead to the best classification results: we observe that as the dimension of the Grassmannian grows, the accuracy of the classification grows to 100% in binary classification experiments. To build a classification model, we use SSVMs to perform simultaneous dimension selection. The resulting classifier selects a subset of dimensions of the embedding without loss in classification performance. Lastly, we present an application of persistent homology to the detection of chemical plumes in hyperspectral movies. The pixels of the raw hyperspectral data cubes are mapped to the geometric framework of the Grassmann manifold where they are analyzed, contrasting our approach with the more standard framework in Euclidean space. An advantage of this approach is that it allows the time slices in a hyperspectral movie to be collapsed to a sequence of points in such a way that some of the key structure within and between the slices is encoded by the points on the Grassmannian. This motivates the search for topological structure, associated with the evolution of the frames of a hyperspectral movie, within the corresponding points on the manifold. The proposed framework affords the processing of large data sets, such as the hyperspectral movies explored in this investigation, while retaining valuable discriminative information. For a particular choice of a distance metric on the Grassmannian, it is possible to generate topological signals that capture changes in the scene after a chemical release
Unsupervised Feature Learning by Autoencoder and Prototypical Contrastive Learning for Hyperspectral Classification
Unsupervised learning methods for feature extraction are becoming more and
more popular. We combine the popular contrastive learning method (prototypical
contrastive learning) and the classic representation learning method
(autoencoder) to design an unsupervised feature learning network for
hyperspectral classification. Experiments have proved that our two proposed
autoencoder networks have good feature learning capabilities by themselves, and
the contrastive learning network we designed can better combine the features of
the two to learn more representative features. As a result, our method
surpasses other comparison methods in the hyperspectral classification
experiments, including some supervised methods. Moreover, our method maintains
a fast feature extraction speed than baseline methods. In addition, our method
reduces the requirements for huge computing resources, separates feature
extraction and contrastive learning, and allows more researchers to conduct
research and experiments on unsupervised contrastive learning
Parsimonious Mahalanobis Kernel for the Classification of High Dimensional Data
The classification of high dimensional data with kernel methods is considered
in this article. Exploit- ing the emptiness property of high dimensional
spaces, a kernel based on the Mahalanobis distance is proposed. The computation
of the Mahalanobis distance requires the inversion of a covariance matrix. In
high dimensional spaces, the estimated covariance matrix is ill-conditioned and
its inversion is unstable or impossible. Using a parsimonious statistical
model, namely the High Dimensional Discriminant Analysis model, the specific
signal and noise subspaces are estimated for each considered class making the
inverse of the class specific covariance matrix explicit and stable, leading to
the definition of a parsimonious Mahalanobis kernel. A SVM based framework is
used for selecting the hyperparameters of the parsimonious Mahalanobis kernel
by optimizing the so-called radius-margin bound. Experimental results on three
high dimensional data sets show that the proposed kernel is suitable for
classifying high dimensional data, providing better classification accuracies
than the conventional Gaussian kernel
Deep Metric Learning Based on Scalable Neighborhood Components for Remote Sensing Scene Characterization
With the development of convolutional neural networks (CNNs), the semantic understanding of remote sensing (RS) scenes has been significantly improved based on their prominent feature encoding capabilities. While many existing deep-learning models focus on designing different architectures, only a few works in the RS field have focused on investigating the performance of the learned feature embeddings and the associated metric space. In particular, two main loss functions have been exploited: the contrastive and the triplet loss. However, the straightforward application of these techniques to RS images may not be optimal in order to capture their neighborhood structures in the metric space due to the insufficient sampling of image pairs or triplets during the training stage and to the inherent semantic complexity of remotely sensed data. To solve these problems, we propose a new deep metric learning approach, which overcomes the limitation on the class discrimination by means of two different components: 1) scalable neighborhood component analysis (SNCA) that aims at discovering the neighborhood structure in the metric space and 2) the cross-entropy loss that aims at preserving the class discrimination capability based on the learned class prototypes. Moreover, in order to preserve feature consistency among all the minibatches during training, a novel optimization mechanism based on momentum update is introduced for minimizing the proposed loss. An extensive experimental comparison (using several state-of-the-art models and two different benchmark data sets) has been conducted to validate the effectiveness of the proposed method from different perspectives, including: 1) classification; 2) clustering; and 3) image retrieval. The related codes of this article will be made publicly available for reproducible research by the community
PerTurbo manifold learning algorithm for weakly labelled hyperspectral image classification
International audienceHyperspectral data analysis has been given a growing attention due to the scientific challenges it raises and the wide set of applications that can benefit from it. Classification of hyperspectral images has been identified as one of the hottest topics in this context, and has been mainly addressed by discriminative methods such as SVM. In this paper, we argue that generative methods, and especially those based on manifold representation of classes in the hyperspectral space, are relevant alternatives to SVM. To illustrate our point, we focus on the recently published PerTurbo algorithm and benchmark against SVM this generative manifold learning algorithm in the context of hyperspectral image classification. This choice is motivated by the fact that PerTurbo is fitted with numerous interesting properties, such as low sensitivity to dimensionality curse, high accuracy in weakly labelled images classification context (few training samples), straightforward extension to on-line setting, and interpretability for the practitioner. The promising results call for an up-to-date interest toward generative algorithms for hyperspectral image classification
- …