963 research outputs found
Block-level discrete cosine transform coefficients for autonomic face recognition
This dissertation presents a novel method of autonomic face recognition based on the recently proposed biologically plausible network of networks (NoN) model of information processing. The NoN model is based on locally parallel and globally coordinated transformations. In the NoN architecture, the neurons or computational units form distributed networks, which themselves link to form larger networks. In the general case, an n-level hierarchy of nested distributed networks is constructed. This models the structures in the cerebral cortex described by Mountcastle and the architecture based on that proposed for information processing by Sutton. In the implementation proposed in the dissertation, the image is processed by a nested family of locally operating networks along with a hierarchically superior network that classifies the information from each of the local networks. The implementation of this approach helps obtain sensitivity to the contrast sensitivity function (CSF) in the middle of the spectrum, as is true for the human vision system. The input images are divided into blocks to define the local regions of processing. The two-dimensional Discrete Cosine Transform (DCT), a spatial frequency transform, is used to transform the data into the frequency domain. Thereafter, statistical operators that calculate various functions of spatial frequency in the block are used to produce a block-level DCT coefficient. The image is now transformed into a variable length vector that is trained with respect to the data set. The classification was done by the use of a backpropagation neural network. The proposed method yields excellent results on a benchmark database. The results of the experiments yielded a maximum of 98.5% recognition accuracy and an average of 97.4% recognition accuracy. An advanced version of the method where the local processing is done on offset blocks has also been developed. This has validated the NoN approach and further research using local processing as well as more advanced global operators is likely to yield even better results
Learning Fast Algorithms for Linear Transforms Using Butterfly Factorizations
Fast linear transforms are ubiquitous in machine learning, including the
discrete Fourier transform, discrete cosine transform, and other structured
transformations such as convolutions. All of these transforms can be
represented by dense matrix-vector multiplication, yet each has a specialized
and highly efficient (subquadratic) algorithm. We ask to what extent
hand-crafting these algorithms and implementations is necessary, what
structural priors they encode, and how much knowledge is required to
automatically learn a fast algorithm for a provided structured transform.
Motivated by a characterization of fast matrix-vector multiplication as
products of sparse matrices, we introduce a parameterization of
divide-and-conquer methods that is capable of representing a large class of
transforms. This generic formulation can automatically learn an efficient
algorithm for many important transforms; for example, it recovers the Cooley-Tukey FFT algorithm to machine precision, for dimensions up to
. Furthermore, our method can be incorporated as a lightweight
replacement of generic matrices in machine learning pipelines to learn
efficient and compressible transformations. On a standard task of compressing a
single hidden-layer network, our method exceeds the classification accuracy of
unconstrained matrices on CIFAR-10 by 3.9 points---the first time a structured
approach has done so---with 4X faster inference speed and 40X fewer parameters
A survey of face detection, extraction and recognition
The goal of this paper is to present a critical survey of existing literatures on human face recognition over the last 4-5 years. Interest and research activities in face recognition have increased significantly over the past few years, especially after the American airliner tragedy on September 11 in 2001. While this growth largely is driven by growing application demands, such as static matching of controlled photographs as in mug shots matching, credit card verification to surveillance video images, identification for law enforcement and authentication for banking and security system access, advances in signal analysis techniques, such as wavelets and neural networks, are also important catalysts. As the number of proposed techniques increases, survey and evaluation becomes important
DCT and DST Filtering with Sparse Graph Operators
Graph filtering is a fundamental tool in graph signal processing. Polynomial
graph filters (PGFs), defined as polynomials of a fundamental graph operator,
can be implemented in the vertex domain, and usually have a lower complexity
than frequency domain filter implementations. In this paper, we focus on the
design of filters for graphs with graph Fourier transform (GFT) corresponding
to a discrete trigonometric transform (DTT), i.e., one of 8 types of discrete
cosine transforms (DCT) and 8 discrete sine transforms (DST). In this case, we
show that multiple sparse graph operators can be identified, which allows us to
propose a generalization of PGF design: multivariate polynomial graph filter
(MPGF). First, for the widely used DCT-II (type-2 DCT), we characterize a set
of sparse graph operators that share the DCT-II matrix as their common
eigenvector matrix. This set contains the well-known connected line graph.
These sparse operators can be viewed as graph filters operating in the DCT
domain, which allows us to approximate any DCT graph filter by a MPGF, leading
to a design with more degrees of freedom than the conventional PGF approach.
Then, we extend those results to all of the 16 DTTs as well as their 2D
versions, and show how their associated sets of multiple graph operators can be
determined. We demonstrate experimentally that ideal low-pass and exponential
DCT/DST filters can be approximated with higher accuracy with similar runtime
complexity. Finally, we apply our method to transform-type selection in a video
codec, AV1, where we demonstrate significant encoding time savings, with a
negligible compression loss.Comment: 16 pages, 11 figures, 5 table
Data comparison schemes for Pattern Recognition in Digital Images using Fractals
Pattern recognition in digital images is a common problem with application in
remote sensing, electron microscopy, medical imaging, seismic imaging and
astrophysics for example. Although this subject has been researched for over
twenty years there is still no general solution which can be compared with the
human cognitive system in which a pattern can be recognised subject to
arbitrary orientation and scale.
The application of Artificial Neural Networks can in principle provide a very
general solution providing suitable training schemes are implemented.
However, this approach raises some major issues in practice. First, the CPU
time required to train an ANN for a grey level or colour image can be very
large especially if the object has a complex structure with no clear geometrical
features such as those that arise in remote sensing applications. Secondly,
both the core and file space memory required to represent large images and
their associated data tasks leads to a number of problems in which the use of
virtual memory is paramount.
The primary goal of this research has been to assess methods of image data
compression for pattern recognition using a range of different compression
methods. In particular, this research has resulted in the design and
implementation of a new algorithm for general pattern recognition based on
the use of fractal image compression.
This approach has for the first time allowed the pattern recognition problem to
be solved in a way that is invariant of rotation and scale. It allows both ANNs
and correlation to be used subject to appropriate pre-and post-processing
techniques for digital image processing on aspect for which a dedicated
programmer's work bench has been developed using X-Designer
Fault Diagnosis of HVDC Systems Using Machine Learning Based Methods
With the development of high-power electronic technology, HVDC system is applied in the power system because of advantages in large-capacity and long-distance transmission, stability, and flexibility. Therefore, as the guarantee of reliable operating of HVDC system, fault diagnosis of the HVDC system is of great significance. In the current variety methods used in fault diagnosis, Machine Learning based methods have become a hotspot. To this end, the performance of several commonly used machine learning classifiers is compared in HVDC system. First of all, nine faults both in AC systems and DC systems of the HVDC system are set in the HVDC model in Simulink. Therefore, 10 operating states corresponding to the faults and normal operating are considered as the output classes of classifier. Seven parameters, such as DC voltage and DC current, are selected as fault feature parameters of each sample. By simulating the HVDC system in 10 operating states (including normal operating state) correspondingly, 20000 samples, each containing seven parameters, be obtained during the fault period. Then, the training sample set and the test sample set are established by 80% and 20% of the whole sample set. Subsequently, Decision Trees, the Support Vector Machine (SVM), K-Nearest Neighborhood Classifier (KNN), Ensemble classifiers, Discriminant Analysis, Backward Propagation Neural Network (BP-NN), long Short-Term Memory Neural Network (LSTM-NN), Extreme Learning Machine (ELM) was trained and tested. The accuracy of testing is used as the performance index of the model. In particular, for BP-NN, the impact of different transfer functions and learning rules combinations on the accuracy of the model was tested. For ELM, the impact of different activation functions on accuracy is tested. The results have shown that ELM and Bagged Trees have the best performance in HVDC fault diagnosis. The accuracy of these two methods are 92.23% and 96.5% respectively. However, in order to achieve better accuracy in ELM model, a large number of hidden layer nodes are set so that training time increases sharply
Fault Diagnosis of HVDC Systems Using Machine Learning Based Methods
With the development of high-power electronic technology, HVDC system is applied in the power system because of advantages in large-capacity and long-distance transmission, stability, and flexibility. Therefore, as the guarantee of reliable operating of HVDC system, fault diagnosis of the HVDC system is of great significance. In the current variety methods used in fault diagnosis, Machine Learning based methods have become a hotspot. To this end, the performance of several commonly used machine learning classifiers is compared in HVDC system. First of all, nine faults both in AC systems and DC systems of the HVDC system are set in the HVDC model in Simulink. Therefore, 10 operating states corresponding to the faults and normal operating are considered as the output classes of classifier. Seven parameters, such as DC voltage and DC current, are selected as fault feature parameters of each sample. By simulating the HVDC system in 10 operating states (including normal operating state) correspondingly, 20000 samples, each containing seven parameters, be obtained during the fault period. Then, the training sample set and the test sample set are established by 80% and 20% of the whole sample set. Subsequently, Decision Trees, the Support Vector Machine (SVM), K-Nearest Neighborhood Classifier (KNN), Ensemble classifiers, Discriminant Analysis, Backward Propagation Neural Network (BP-NN), long Short-Term Memory Neural Network (LSTM-NN), Extreme Learning Machine (ELM) was trained and tested. The accuracy of testing is used as the performance index of the model. In particular, for BP-NN, the impact of different transfer functions and learning rules combinations on the accuracy of the model was tested. For ELM, the impact of different activation functions on accuracy is tested. The results have shown that ELM and Bagged Trees have the best performance in HVDC fault diagnosis. The accuracy of these two methods are 92.23% and 96.5% respectively. However, in order to achieve better accuracy in ELM model, a large number of hidden layer nodes are set so that training time increases sharply
- …