22,795 research outputs found
PENGENALAN WAJAH MENGGUNAKAN METODE CONVOLUTIONAL NEURAL NETWORK-RESTRICTED BOLTZMANN MACHINE BERBASIS PRINCIPAL COMPONENT ANALYSIS
Teknologi pengenalan wajah berpotensi untuk diterapkan pada berbagai bidang dalam kehidupan sehari-hari. Penelitian ini melakukan pengembangan teknologi pengenalan wajah dengan mengusulkan metode Convolutional Neural Network-Restricted Boltzmann Machine (CNN-RBM) berbasis Principal Component Analysis (PCA) menggunakan set data Labeled Faces in the Wild (LFW). CNN-RBM berbasis PCA memanfaatkan PCA sebagai pereduksi dimensi pada input, kemudian menggunakan CNN sebagai ekstraksi fitur, dan menggunakan RBM pada tahap klasifikasi wajah. Hasil eksperimen membuktikan bahwa CNN-RBM berbasis PCA mampu mengungguli baseline dengan peningkatan akurasi sebesar 1,6%.
Face recognition technology can be applied in various fields of in everyday life. This research develops face recognition technology using Convolutional Neural Network-Restricted Boltzmann Machine (CNN-RBM) based on Principal Component Analysis (PCA) using labeled Faces in the Wild (LFW) set data. PCN-based CNN-RBM uses PCA as a dimension reduction in input, then uses CNN as a feature extraction, and uses RBM in face classification. The experimental results prove that PCN-based CNN-RBM was able to outperform the baseline with 1,6% accuracy improvement
CIFAR-10: KNN-based Ensemble of Classifiers
In this paper, we study the performance of different classifiers on the
CIFAR-10 dataset, and build an ensemble of classifiers to reach a better
performance. We show that, on CIFAR-10, K-Nearest Neighbors (KNN) and
Convolutional Neural Network (CNN), on some classes, are mutually exclusive,
thus yield in higher accuracy when combined. We reduce KNN overfitting using
Principal Component Analysis (PCA), and ensemble it with a CNN to increase its
accuracy. Our approach improves our best CNN model from 93.33% to 94.03%
Distance Metric Learning using Graph Convolutional Networks: Application to Functional Brain Networks
Evaluating similarity between graphs is of major importance in several
computer vision and pattern recognition problems, where graph representations
are often used to model objects or interactions between elements. The choice of
a distance or similarity metric is, however, not trivial and can be highly
dependent on the application at hand. In this work, we propose a novel metric
learning method to evaluate distance between graphs that leverages the power of
convolutional neural networks, while exploiting concepts from spectral graph
theory to allow these operations on irregular graphs. We demonstrate the
potential of our method in the field of connectomics, where neuronal pathways
or functional connections between brain regions are commonly modelled as
graphs. In this problem, the definition of an appropriate graph similarity
function is critical to unveil patterns of disruptions associated with certain
brain disorders. Experimental results on the ABIDE dataset show that our method
can learn a graph similarity metric tailored for a clinical application,
improving the performance of a simple k-nn classifier by 11.9% compared to a
traditional distance metric.Comment: International Conference on Medical Image Computing and
Computer-Assisted Interventions (MICCAI) 201
Class-Weighted Convolutional Features for Visual Instance Search
Image retrieval in realistic scenarios targets large dynamic datasets of
unlabeled images. In these cases, training or fine-tuning a model every time
new images are added to the database is neither efficient nor scalable.
Convolutional neural networks trained for image classification over large
datasets have been proven effective feature extractors for image retrieval. The
most successful approaches are based on encoding the activations of
convolutional layers, as they convey the image spatial information. In this
paper, we go beyond this spatial information and propose a local-aware encoding
of convolutional features based on semantic information predicted in the target
image. To this end, we obtain the most discriminative regions of an image using
Class Activation Maps (CAMs). CAMs are based on the knowledge contained in the
network and therefore, our approach, has the additional advantage of not
requiring external information. In addition, we use CAMs to generate object
proposals during an unsupervised re-ranking stage after a first fast search.
Our experiments on two public available datasets for instance retrieval,
Oxford5k and Paris6k, demonstrate the competitiveness of our approach
outperforming the current state-of-the-art when using off-the-shelf models
trained on ImageNet. The source code and model used in this paper are publicly
available at http://imatge-upc.github.io/retrieval-2017-cam/.Comment: To appear in the British Machine Vision Conference (BMVC), September
201
Deep learning architectures for Computer Vision
Deep learning has become part of many state-of-the-art systems in multiple disciplines (specially in computer vision and speech processing). In this thesis Convolutional Neural Networks are used to solve the problem of recognizing people in images, both for verification and identification. Two different architectures, AlexNet and VGG19, both winners of the ILSVRC, have been fine-tuned and tested with four datasets: Labeled Faces in the Wild, FaceScrub, YouTubeFaces and Google UPC, a dataset generated at the UPC. Finally, with the features extracted from these fine-tuned networks, some verifications algorithms have been tested including Support Vector Machines, Joint Bayesian and Advanced Joint Bayesian formulation. The results of this work show that an Area Under the Receiver Operating Characteristic curve of 99.6% can be obtained, close to the state-of-the-art performance.El aprendizaje profundo se ha convertido en parte de muchos sistemas en el estado del arte de múltiples ámbitos (especialmente en visión por computador y procesamiento de voz). En esta tesis se utilizan las Redes Neuronales Convolucionales para resolver el problema de reconocer a personas en imágenes, tanto para verificación como para identificación. Dos arquitecturas diferentes, AlexNet y VGG19, ambas ganadores del ILSVRC, han sido afinadas y probadas con cuatro conjuntos de datos: Labeled Faces in the Wild, FaceScrub, YouTubeFaces y Google UPC, un conjunto generado en la UPC. Finalmente con las características extraídas de las redes afinadas, se han probado diferentes algoritmos de verificación, incluyendo Maquinas de Soporte Vectorial, Joint Bayesian y Advanced Joint Bayesian. Los resultados de este trabajo muestran que el Área Bajo la Curva de la Característica Operativa del Receptor puede llegar a ser del 99.6%, cercana al valor del estado del arte.L’aprenentatge profund s’ha convertit en una part importat de molts sistemes a l’estat de
l’art de múltiples àmbits (especialment de la visió per computador i el processament de
veu). A aquesta tesi s’utilitzen les Xarxes Neuronals Convolucionals per a resoldre el
problema de reconèixer persones a imatges, tant per verificació com per identificatió.
Dos arquitectures diferents, AlexNet i VGG19, les dues guanyadores del ILSVRC, han
sigut afinades i provades amb quatre bases de dades: Labeled Faces in the Wild,
FaceScrub, YouTubeFaces i Google UPC, un conjunt generat a la UPC.
Finalment, amb les característiques extretes de les xarxes afinades, s’han provat diferents
algoritmes de verificació, incloent Màquines de Suport Vectorial, Joint Bayesian i Advanced
Joint Bayesian. Els resultats d’aquest treball mostres que un Àrea Baix la Curva de la
Característica Operativa del Receptor por arribar a ser del 99.6%, propera al valor de l’estat
de l’art
- …