1,038 research outputs found
Deep Fisher Discriminant Learning for Mobile Hand Gesture Recognition
Gesture recognition becomes a popular analytics tool for extracting the characteristics of user movement and enables numerous practical applications in the biometrics field. Despite recent advances in this technique, complex user interaction and the limited amount of data pose serious challenges to existing methods. In this paper, we present a novel approach for hand gesture recognition based on user interaction on mobile devices. We have developed two deep models by integrating Bidirectional Long-Short Term Memory (BiLSTM) network and Bidirectional Gated Recurrent Unit (BiGRU) with Fisher criterion, termed as F-BiLSTM and F-BiGRU respectively. These two Fisher discriminative models can classify user’s gesture effectively by analyzing the corresponding acceleration and angular velocity data of hand motion. In addition, we build a large Mobile Gesture Database (MGD) containing 5547 sequences of 12 gestures. With extensive experiments, we demonstrate the superior performance of the proposed method compared to the state-of-the-art BiLSTM and BiGRU on MGD database and two other benchmark databases (i.e., BUAA mobile gesture and SmartWatch gesture). The source code and MGD database will be made publicly available at https://github.com/bczhangbczhang/Fisher-Discriminant-LSTM
Automated Face Recognition: Challenges and Solutions
Automated face recognition (AFR) aims to identify people in images or videos using pattern recognition techniques. Automated face recognition is widely used in applications ranging from social media to advanced authentication systems. Whilst techniques for face recognition are well established, the automatic recognition of faces captured by digital cameras in unconstrained, real‐world environment is still very challenging, since it involves important variations in both acquisition conditions as well as in facial expressions and in pose changes. Thus, this chapter introduces the topic of computer automated face recognition in light of the main challenges in that research field and the developed solutions and applications based on image processing and artificial intelligence methods
SCLAiR : Supervised Contrastive Learning for User and Device Independent Airwriting Recognition
Airwriting Recognition is the problem of identifying letters written in free
space with finger movement. It is essentially a specialized case of gesture
recognition, wherein the vocabulary of gestures corresponds to letters as in a
particular language. With the wide adoption of smart wearables in the general
population, airwriting recognition using motion sensors from a smart-band can
be used as a medium of user input for applications in Human-Computer
Interaction. There has been limited work in the recognition of in-air
trajectories using motion sensors, and the performance of the techniques in the
case when the device used to record signals is changed has not been explored
hitherto. Motivated by these, a new paradigm for device and user-independent
airwriting recognition based on supervised contrastive learning is proposed. A
two stage classification strategy is employed, the first of which involves
training an encoder network with supervised contrastive loss. In the subsequent
stage, a classification head is trained with the encoder weights kept frozen.
The efficacy of the proposed method is demonstrated through experiments on a
publicly available dataset and also with a dataset recorded in our lab using a
different device. Experiments have been performed in both supervised and
unsupervised settings and compared against several state-of-the-art domain
adaptation techniques. Data and the code for our implementation will be made
available at https://github.com/ayushayt/SCLAiR
ImAiR: Airwriting Recognition framework using Image Representation of IMU Signals
The problem of Airwriting Recognition is focused on identifying letters
written by movement of finger in free space. It is a type of gesture
recognition where the dictionary corresponds to letters in a specific language.
In particular, airwriting recognition using sensor data from wrist-worn devices
can be used as a medium of user input for applications in Human-Computer
Interaction (HCI). Recognition of in-air trajectories using such wrist-worn
devices is limited in literature and forms the basis of the current work. In
this paper, we propose an airwriting recognition framework by first encoding
the time-series data obtained from a wearable Inertial Measurement Unit (IMU)
on the wrist as images and then utilizing deep learning-based models for
identifying the written alphabets. The signals recorded from 3-axis
accelerometer and gyroscope in IMU are encoded as images using different
techniques such as Self Similarity Matrix (SSM), Gramian Angular Field (GAF)
and Markov Transition Field (MTF) to form two sets of 3-channel images. These
are then fed to two separate classification models and letter prediction is
made based on an average of the class conditional probabilities obtained from
the two models. Several standard model architectures for image classification
such as variants of ResNet, DenseNet, VGGNet, AlexNet and GoogleNet have been
utilized. Experiments performed on two publicly available datasets demonstrate
the efficacy of the proposed strategy. The code for our implementation will be
made available at https://github.com/ayushayt/ImAiR
- …