963 research outputs found

    Block-level discrete cosine transform coefficients for autonomic face recognition

    Get PDF
    This dissertation presents a novel method of autonomic face recognition based on the recently proposed biologically plausible network of networks (NoN) model of information processing. The NoN model is based on locally parallel and globally coordinated transformations. In the NoN architecture, the neurons or computational units form distributed networks, which themselves link to form larger networks. In the general case, an n-level hierarchy of nested distributed networks is constructed. This models the structures in the cerebral cortex described by Mountcastle and the architecture based on that proposed for information processing by Sutton. In the implementation proposed in the dissertation, the image is processed by a nested family of locally operating networks along with a hierarchically superior network that classifies the information from each of the local networks. The implementation of this approach helps obtain sensitivity to the contrast sensitivity function (CSF) in the middle of the spectrum, as is true for the human vision system. The input images are divided into blocks to define the local regions of processing. The two-dimensional Discrete Cosine Transform (DCT), a spatial frequency transform, is used to transform the data into the frequency domain. Thereafter, statistical operators that calculate various functions of spatial frequency in the block are used to produce a block-level DCT coefficient. The image is now transformed into a variable length vector that is trained with respect to the data set. The classification was done by the use of a backpropagation neural network. The proposed method yields excellent results on a benchmark database. The results of the experiments yielded a maximum of 98.5% recognition accuracy and an average of 97.4% recognition accuracy. An advanced version of the method where the local processing is done on offset blocks has also been developed. This has validated the NoN approach and further research using local processing as well as more advanced global operators is likely to yield even better results

    Learning Fast Algorithms for Linear Transforms Using Butterfly Factorizations

    Get PDF
    Fast linear transforms are ubiquitous in machine learning, including the discrete Fourier transform, discrete cosine transform, and other structured transformations such as convolutions. All of these transforms can be represented by dense matrix-vector multiplication, yet each has a specialized and highly efficient (subquadratic) algorithm. We ask to what extent hand-crafting these algorithms and implementations is necessary, what structural priors they encode, and how much knowledge is required to automatically learn a fast algorithm for a provided structured transform. Motivated by a characterization of fast matrix-vector multiplication as products of sparse matrices, we introduce a parameterization of divide-and-conquer methods that is capable of representing a large class of transforms. This generic formulation can automatically learn an efficient algorithm for many important transforms; for example, it recovers the O(NlogN)O(N \log N) Cooley-Tukey FFT algorithm to machine precision, for dimensions NN up to 10241024. Furthermore, our method can be incorporated as a lightweight replacement of generic matrices in machine learning pipelines to learn efficient and compressible transformations. On a standard task of compressing a single hidden-layer network, our method exceeds the classification accuracy of unconstrained matrices on CIFAR-10 by 3.9 points---the first time a structured approach has done so---with 4X faster inference speed and 40X fewer parameters

    A survey of face detection, extraction and recognition

    Get PDF
    The goal of this paper is to present a critical survey of existing literatures on human face recognition over the last 4-5 years. Interest and research activities in face recognition have increased significantly over the past few years, especially after the American airliner tragedy on September 11 in 2001. While this growth largely is driven by growing application demands, such as static matching of controlled photographs as in mug shots matching, credit card verification to surveillance video images, identification for law enforcement and authentication for banking and security system access, advances in signal analysis techniques, such as wavelets and neural networks, are also important catalysts. As the number of proposed techniques increases, survey and evaluation becomes important

    DCT and DST Filtering with Sparse Graph Operators

    Full text link
    Graph filtering is a fundamental tool in graph signal processing. Polynomial graph filters (PGFs), defined as polynomials of a fundamental graph operator, can be implemented in the vertex domain, and usually have a lower complexity than frequency domain filter implementations. In this paper, we focus on the design of filters for graphs with graph Fourier transform (GFT) corresponding to a discrete trigonometric transform (DTT), i.e., one of 8 types of discrete cosine transforms (DCT) and 8 discrete sine transforms (DST). In this case, we show that multiple sparse graph operators can be identified, which allows us to propose a generalization of PGF design: multivariate polynomial graph filter (MPGF). First, for the widely used DCT-II (type-2 DCT), we characterize a set of sparse graph operators that share the DCT-II matrix as their common eigenvector matrix. This set contains the well-known connected line graph. These sparse operators can be viewed as graph filters operating in the DCT domain, which allows us to approximate any DCT graph filter by a MPGF, leading to a design with more degrees of freedom than the conventional PGF approach. Then, we extend those results to all of the 16 DTTs as well as their 2D versions, and show how their associated sets of multiple graph operators can be determined. We demonstrate experimentally that ideal low-pass and exponential DCT/DST filters can be approximated with higher accuracy with similar runtime complexity. Finally, we apply our method to transform-type selection in a video codec, AV1, where we demonstrate significant encoding time savings, with a negligible compression loss.Comment: 16 pages, 11 figures, 5 table

    Data comparison schemes for Pattern Recognition in Digital Images using Fractals

    Get PDF
    Pattern recognition in digital images is a common problem with application in remote sensing, electron microscopy, medical imaging, seismic imaging and astrophysics for example. Although this subject has been researched for over twenty years there is still no general solution which can be compared with the human cognitive system in which a pattern can be recognised subject to arbitrary orientation and scale. The application of Artificial Neural Networks can in principle provide a very general solution providing suitable training schemes are implemented. However, this approach raises some major issues in practice. First, the CPU time required to train an ANN for a grey level or colour image can be very large especially if the object has a complex structure with no clear geometrical features such as those that arise in remote sensing applications. Secondly, both the core and file space memory required to represent large images and their associated data tasks leads to a number of problems in which the use of virtual memory is paramount. The primary goal of this research has been to assess methods of image data compression for pattern recognition using a range of different compression methods. In particular, this research has resulted in the design and implementation of a new algorithm for general pattern recognition based on the use of fractal image compression. This approach has for the first time allowed the pattern recognition problem to be solved in a way that is invariant of rotation and scale. It allows both ANNs and correlation to be used subject to appropriate pre-and post-processing techniques for digital image processing on aspect for which a dedicated programmer's work bench has been developed using X-Designer

    Fault Diagnosis of HVDC Systems Using Machine Learning Based Methods

    Get PDF
    With the development of high-power electronic technology, HVDC system is applied in the power system because of advantages in large-capacity and long-distance transmission, stability, and flexibility. Therefore, as the guarantee of reliable operating of HVDC system, fault diagnosis of the HVDC system is of great significance. In the current variety methods used in fault diagnosis, Machine Learning based methods have become a hotspot. To this end, the performance of several commonly used machine learning classifiers is compared in HVDC system. First of all, nine faults both in AC systems and DC systems of the HVDC system are set in the HVDC model in Simulink. Therefore, 10 operating states corresponding to the faults and normal operating are considered as the output classes of classifier. Seven parameters, such as DC voltage and DC current, are selected as fault feature parameters of each sample. By simulating the HVDC system in 10 operating states (including normal operating state) correspondingly, 20000 samples, each containing seven parameters, be obtained during the fault period. Then, the training sample set and the test sample set are established by 80% and 20% of the whole sample set. Subsequently, Decision Trees, the Support Vector Machine (SVM), K-Nearest Neighborhood Classifier (KNN), Ensemble classifiers, Discriminant Analysis, Backward Propagation Neural Network (BP-NN), long Short-Term Memory Neural Network (LSTM-NN), Extreme Learning Machine (ELM) was trained and tested. The accuracy of testing is used as the performance index of the model. In particular, for BP-NN, the impact of different transfer functions and learning rules combinations on the accuracy of the model was tested. For ELM, the impact of different activation functions on accuracy is tested. The results have shown that ELM and Bagged Trees have the best performance in HVDC fault diagnosis. The accuracy of these two methods are 92.23% and 96.5% respectively. However, in order to achieve better accuracy in ELM model, a large number of hidden layer nodes are set so that training time increases sharply

    Fault Diagnosis of HVDC Systems Using Machine Learning Based Methods

    Get PDF
    With the development of high-power electronic technology, HVDC system is applied in the power system because of advantages in large-capacity and long-distance transmission, stability, and flexibility. Therefore, as the guarantee of reliable operating of HVDC system, fault diagnosis of the HVDC system is of great significance. In the current variety methods used in fault diagnosis, Machine Learning based methods have become a hotspot. To this end, the performance of several commonly used machine learning classifiers is compared in HVDC system. First of all, nine faults both in AC systems and DC systems of the HVDC system are set in the HVDC model in Simulink. Therefore, 10 operating states corresponding to the faults and normal operating are considered as the output classes of classifier. Seven parameters, such as DC voltage and DC current, are selected as fault feature parameters of each sample. By simulating the HVDC system in 10 operating states (including normal operating state) correspondingly, 20000 samples, each containing seven parameters, be obtained during the fault period. Then, the training sample set and the test sample set are established by 80% and 20% of the whole sample set. Subsequently, Decision Trees, the Support Vector Machine (SVM), K-Nearest Neighborhood Classifier (KNN), Ensemble classifiers, Discriminant Analysis, Backward Propagation Neural Network (BP-NN), long Short-Term Memory Neural Network (LSTM-NN), Extreme Learning Machine (ELM) was trained and tested. The accuracy of testing is used as the performance index of the model. In particular, for BP-NN, the impact of different transfer functions and learning rules combinations on the accuracy of the model was tested. For ELM, the impact of different activation functions on accuracy is tested. The results have shown that ELM and Bagged Trees have the best performance in HVDC fault diagnosis. The accuracy of these two methods are 92.23% and 96.5% respectively. However, in order to achieve better accuracy in ELM model, a large number of hidden layer nodes are set so that training time increases sharply
    corecore