40,508 research outputs found
A Hybrid Stacked CNN Model with Weighted Average Ensembling for Effective Face Recognition
The discipline of computer vision has given a lot of attention to facial recognition. Automated face recognition is extensively employed in various practical scenarios, including systems for streamlining immigration checkpoints, intelligent monitoring of visual data, and authentication of personal identity. Depending on the situation, it may be divided into the two separate duties of facial verification and face identification. This study proposed a hybrid stacked CNN model for face recognition system. The models have mastered the art of making their own inferences. The models are then combined to predict a class value using a cutting-edge method called weighted average ensembling. A more accurate estimate should be produced by the new assembly process. Pre-trained CNN models are used in our proposed method: AlexNet, Resnet50V2, and VGG-19, Yolov5, VGG-16 and ResNet50. When applied to Tufts dataset images, our suggested model successfully achieved 98.05% accuracy. We have also used the Discrete Wavelet Transform (DWT) method for denoising, SegNet for Image segmentation for better performance of the model proposed
A Novel Hybrid Biometric Electronic Voting System: Integrating Finger Print and Face Recognition
A novel hybrid design based electronic voting system is proposed, implemented
and analyzed. The proposed system uses two voter verification techniques to
give better results in comparison to single identification based systems.
Finger print and facial recognition based methods are used for voter
identification. Cross verification of a voter during an election process
provides better accuracy than single parameter identification method. The
facial recognition system uses Viola-Jones algorithm along with rectangular
Haar feature selection method for detection and extraction of features to
develop a biometric template and for feature extraction during the voting
process. Cascaded machine learning based classifiers are used for comparing the
features for identity verification using GPCA (Generalized Principle Component
Analysis) and K-NN (K-Nearest Neighbor). It is accomplished through comparing
the Eigen-vectors of the extracted features with the biometric template
pre-stored in the election regulatory body database. The results of the
proposed system show that the proposed cascaded design based system performs
better than the systems using other classifiers or separate schemes i.e. facial
or finger print based schemes. The proposed system will be highly useful for
real time applications due to the reason that it has 91% accuracy under nominal
light in terms of facial recognition. with bags of paper votes. The central
station compiles and publishes the names of winners and losers through
television and radio stations. This method is useful only if the whole process
is completed in a transparent way. However, there are some drawbacks to this
system. These include higher expenses, longer time to complete the voting
process, fraudulent practices by the authorities administering elections as
well as malpractices by the voters [1]. These challenges result in manipulated
election results
Timing is everything: A spatio-temporal approach to the analysis of facial actions
This thesis presents a fully automatic facial expression analysis system based on the Facial Action
Coding System (FACS). FACS is the best known and the most commonly used system to describe
facial activity in terms of facial muscle actions (i.e., action units, AUs). We will present our research
on the analysis of the morphological, spatio-temporal and behavioural aspects of facial expressions.
In contrast with most other researchers in the field who use appearance based techniques, we use a
geometric feature based approach. We will argue that that approach is more suitable for analysing
facial expression temporal dynamics. Our system is capable of explicitly exploring the temporal
aspects of facial expressions from an input colour video in terms of their onset (start), apex (peak)
and offset (end).
The fully automatic system presented here detects 20 facial points in the first frame and tracks them
throughout the video. From the tracked points we compute geometry-based features which serve as
the input to the remainder of our systems. The AU activation detection system uses GentleBoost
feature selection and a Support Vector Machine (SVM) classifier to find which AUs were present in an
expression. Temporal dynamics of active AUs are recognised by a hybrid GentleBoost-SVM-Hidden
Markov model classifier. The system is capable of analysing 23 out of 27 existing AUs with high
accuracy.
The main contributions of the work presented in this thesis are the following: we have created a
method for fully automatic AU analysis with state-of-the-art recognition results. We have proposed
for the first time a method for recognition of the four temporal phases of an AU. We have build the
largest comprehensive database of facial expressions to date. We also present for the first time in the
literature two studies for automatic distinction between posed and spontaneous expressions
Fast Landmark Localization with 3D Component Reconstruction and CNN for Cross-Pose Recognition
Two approaches are proposed for cross-pose face recognition, one is based on
the 3D reconstruction of facial components and the other is based on the deep
Convolutional Neural Network (CNN). Unlike most 3D approaches that consider
holistic faces, the proposed approach considers 3D facial components. It
segments a 2D gallery face into components, reconstructs the 3D surface for
each component, and recognizes a probe face by component features. The
segmentation is based on the landmarks located by a hierarchical algorithm that
combines the Faster R-CNN for face detection and the Reduced Tree Structured
Model for landmark localization. The core part of the CNN-based approach is a
revised VGG network. We study the performances with different settings on the
training set, including the synthesized data from 3D reconstruction, the
real-life data from an in-the-wild database, and both types of data combined.
We investigate the performances of the network when it is employed as a
classifier or designed as a feature extractor. The two recognition approaches
and the fast landmark localization are evaluated in extensive experiments, and
compared to stateof-the-art methods to demonstrate their efficacy.Comment: 14 pages, 12 figures, 4 table
LEARNet Dynamic Imaging Network for Micro Expression Recognition
Unlike prevalent facial expressions, micro expressions have subtle,
involuntary muscle movements which are short-lived in nature. These minute
muscle movements reflect true emotions of a person. Due to the short duration
and low intensity, these micro-expressions are very difficult to perceive and
interpret correctly. In this paper, we propose the dynamic representation of
micro-expressions to preserve facial movement information of a video in a
single frame. We also propose a Lateral Accretive Hybrid Network (LEARNet) to
capture micro-level features of an expression in the facial region. The LEARNet
refines the salient expression features in accretive manner by incorporating
accretion layers (AL) in the network. The response of the AL holds the hybrid
feature maps generated by prior laterally connected convolution layers.
Moreover, LEARNet architecture incorporates the cross decoupled relationship
between convolution layers which helps in preserving the tiny but influential
facial muscle change information. The visual responses of the proposed LEARNet
depict the effectiveness of the system by preserving both high- and micro-level
edge features of facial expression. The effectiveness of the proposed LEARNet
is evaluated on four benchmark datasets: CASME-I, CASME-II, CAS(ME)^2 and SMIC.
The experimental results after investigation show a significant improvement of
4.03%, 1.90%, 1.79% and 2.82% as compared with ResNet on CASME-I, CASME-II,
CAS(ME)^2 and SMIC datasets respectively.Comment: Dynamic imaging, accretion, lateral, micro expression recognitio
- …