1,054 research outputs found
Investigation of feature extraction algorithms and techniques for hyperspectral images.
Doctor of Philosophy (Computer Engineering). University of KwaZulu-Natal. Durban, 2017.Hyperspectral images (HSIs) are remote-sensed images that are characterized
by very high spatial and spectral dimensions and nd applications, for example,
in land cover classi cation, urban planning and management, security and food
processing. Unlike conventional three bands RGB images, their high
dimensional data space creates a challenge for traditional image processing
techniques which are usually based on the assumption that there exists
su cient training samples in order to increase the likelihood of high
classi cation accuracy. However, the high cost and di culty of obtaining
ground truth of hyperspectral data sets makes this assumption unrealistic and
necessitates the introduction of alternative methods for their processing.
Several techniques have been developed in the exploration of the rich spectral
and spatial information in HSIs. Speci cally, feature extraction (FE)
techniques are introduced in the processing of HSIs as a necessary step before
classi cation. They are aimed at transforming the high dimensional data of the
HSI into one of a lower dimension while retaining as much spatial and/or
spectral information as possible. In this research, we develop semi-supervised
FE techniques which combine features of supervised and unsupervised
techniques into a single framework for the processing of HSIs. Firstly, we
developed a feature extraction algorithm known as Semi-Supervised Linear
Embedding (SSLE) for the extraction of features in HSI. The algorithm
combines supervised Linear Discriminant Analysis (LDA) and unsupervised
Local Linear Embedding (LLE) to enhance class discrimination while also
preserving the properties of classes of interest. The technique was developed
based on the fact that LDA extracts features from HSIs by discriminating
between classes of interest and it can only extract C 1 features provided there
are C classes in the image by extracting features that are equivalent to the
number of classes in the HSI. Experiments show that the SSLE algorithm
overcomes the limitation of LDA and extracts features that are equivalent to
ii
iii
the number of classes in HSIs. Secondly, a graphical manifold dimension
reduction (DR) algorithm known as Graph Clustered Discriminant Analysis
(GCDA) is developed. The algorithm is developed to dynamically select labeled
samples from the pool of available unlabeled samples in order to complement
the few available label samples in HSIs. The selection is achieved by entwining
K-means clustering with a semi-supervised manifold discriminant analysis.
Using two HSI data sets, experimental results show that GCDA extracts
features that are equivalent to the number of classes with high classi cation
accuracy when compared with other state-of-the-art techniques. Furthermore,
we develop a window-based partitioning approach to preserve the spatial
properties of HSIs when their features are being extracted. In this approach,
the HSI is partitioned along its spatial dimension into n windows and the
covariance matrices of each window are computed. The covariance matrices of
the windows are then merged into a single matrix through using the Kalman
ltering approach so that the resulting covariance matrix may be used for
dimension reduction. Experiments show that the windowing approach achieves
high classi cation accuracy and preserves the spatial properties of HSIs. For
the proposed feature extraction techniques, Support Vector Machine (SVM)
and Neural Networks (NN) classi cation techniques are employed and their
performances are compared for these two classi ers. The performances of all
proposed FE techniques have also been shown to outperform other
state-of-the-art approaches
Bilinear Random Projections for Locality-Sensitive Binary Codes
Locality-sensitive hashing (LSH) is a popular data-independent indexing
method for approximate similarity search, where random projections followed by
quantization hash the points from the database so as to ensure that the
probability of collision is much higher for objects that are close to each
other than for those that are far apart. Most of high-dimensional visual
descriptors for images exhibit a natural matrix structure. When visual
descriptors are represented by high-dimensional feature vectors and long binary
codes are assigned, a random projection matrix requires expensive complexities
in both space and time. In this paper we analyze a bilinear random projection
method where feature matrices are transformed to binary codes by two smaller
random projection matrices. We base our theoretical analysis on extending
Raginsky and Lazebnik's result where random Fourier features are composed with
random binary quantizers to form locality sensitive binary codes. To this end,
we answer the following two questions: (1) whether a bilinear random projection
also yields similarity-preserving binary codes; (2) whether a bilinear random
projection yields performance gain or loss, compared to a large linear
projection. Regarding the first question, we present upper and lower bounds on
the expected Hamming distance between binary codes produced by bilinear random
projections. In regards to the second question, we analyze the upper and lower
bounds on covariance between two bits of binary codes, showing that the
correlation between two bits is small. Numerical experiments on MNIST and
Flickr45K datasets confirm the validity of our method.Comment: 11 pages, 23 figures, CVPR-201
THRIVE: Threshold Homomorphic encryption based secure and privacy preserving bIometric VErification system
In this paper, we propose a new biometric verification and template
protection system which we call the THRIVE system. The system includes novel
enrollment and authentication protocols based on threshold homomorphic
cryptosystem where the private key is shared between a user and the verifier.
In the THRIVE system, only encrypted binary biometric templates are stored in
the database and verification is performed via homomorphically randomized
templates, thus, original templates are never revealed during the
authentication stage. The THRIVE system is designed for the malicious model
where the cheating party may arbitrarily deviate from the protocol
specification. Since threshold homomorphic encryption scheme is used, a
malicious database owner cannot perform decryption on encrypted templates of
the users in the database. Therefore, security of the THRIVE system is enhanced
using a two-factor authentication scheme involving the user's private key and
the biometric data. We prove security and privacy preservation capability of
the proposed system in the simulation-based model with no assumption. The
proposed system is suitable for applications where the user does not want to
reveal her biometrics to the verifier in plain form but she needs to proof her
physical presence by using biometrics. The system can be used with any
biometric modality and biometric feature extraction scheme whose output
templates can be binarized. The overall connection time for the proposed THRIVE
system is estimated to be 336 ms on average for 256-bit biohash vectors on a
desktop PC running with quad-core 3.2 GHz CPUs at 10 Mbit/s up/down link
connection speed. Consequently, the proposed system can be efficiently used in
real life applications
DIMAL: Deep Isometric Manifold Learning Using Sparse Geodesic Sampling
This paper explores a fully unsupervised deep learning approach for computing
distance-preserving maps that generate low-dimensional embeddings for a certain
class of manifolds. We use the Siamese configuration to train a neural network
to solve the problem of least squares multidimensional scaling for generating
maps that approximately preserve geodesic distances. By training with only a
few landmarks, we show a significantly improved local and nonlocal
generalization of the isometric mapping as compared to analogous non-parametric
counterparts. Importantly, the combination of a deep-learning framework with a
multidimensional scaling objective enables a numerical analysis of network
architectures to aid in understanding their representation power. This provides
a geometric perspective to the generalizability of deep learning.Comment: 10 pages, 11 Figure
- …