5,292 research outputs found

    Recurrent Attention Models for Depth-Based Person Identification

    Get PDF
    We present an attention-based model that reasons on human body shape and motion dynamics to identify individuals in the absence of RGB information, hence in the dark. Our approach leverages unique 4D spatio-temporal signatures to address the identification problem across days. Formulated as a reinforcement learning task, our model is based on a combination of convolutional and recurrent neural networks with the goal of identifying small, discriminative regions indicative of human identity. We demonstrate that our model produces state-of-the-art results on several published datasets given only depth images. We further study the robustness of our model towards viewpoint, appearance, and volumetric changes. Finally, we share insights gleaned from interpretable 2D, 3D, and 4D visualizations of our model's spatio-temporal attention.Comment: Computer Vision and Pattern Recognition (CVPR) 201

    Influencing brain waves by evoked potentials as biometric approach: taking stock of the last six years of research

    Get PDF
    The scientific advances of recent years have made available to anyone affordable hardware devices capable of doing something unthinkable until a few years ago, the reading of brain waves. It means that through small wearable devices it is possible to perform an electroencephalography (EEG), albeit with less potential than those offered by high-cost professional devices. Such devices make it possible for researchers a huge number of experiments that were once impossible in many areas due to the high costs of the necessary hardware. Many studies in the literature explore the use of EEG data as a biometric approach for people identification, but, unfortunately, it presents problems mainly related to the difficulty of extracting unique and stable patterns from users, despite the adoption of sophisticated techniques. An approach to face this problem is based on the evoked potentials (EPs), external stimuli applied during the EEG reading, a noninvasive technique used for many years in clinical routine, in combination with other diagnostic tests, to evaluate the electrical activity related to some areas of the brain and spinal cord to diagnose neurological disorders. In consideration of the growing number of works in the literature that combine the EEG and EP approaches for biometric purposes, this work aims to evaluate the practical feasibility of such approaches as reliable biometric instruments for user identification by surveying the state of the art of the last 6 years, also providing an overview of the elements and concepts related to this research area

    SaS-BCI: A New Strategy to Predict Image Memorability and use Mental Imagery as a Brain-Based Biometric Authentication

    Get PDF
    Security authentication is one of the most important levels of information security. Nowadays, human biometric techniques are the most secure methods for authentication purposes that cover the problems of older types of authentication like passwords and pins. There are many advantages of recent biometrics in terms of security; however, they still have some disadvantages. Progresses in technology made some specific devices, which make it possible to copy and make a fake human biometric because they are all visible and touchable. According to this matter, there is a need for a new biometric to cover the issues of other types. Brainwave is human data, which uses them as a new type of security authentication that has engaged many researchers. There are some research and experiments, which are investigating and testing EEG signals to find the uniqueness of human brainwave. Some researchers achieved high accuracy rates in this area by applying different signal acquisition techniques, feature extraction and classifications using Brain–Computer Interface (BCI). One of the important parts of any BCI processes is the way that brainwaves could be acquired and recorded. A new Signal Acquisition Strategy is presented in this paper for the process of authorization and authentication of brain signals specifically. This is to predict image memorability from the user’s brain to use mental imagery as a visualization pattern for security authentication. Therefore, users can authenticate themselves with visualizing a specific picture in their minds. In conclusion, we can see that brainwaves can be different according to the mental tasks, which it would make it harder using them for authentication process. There are many signal acquisition strategies and signal processing for brain-based authentication that by using the right methods, a higher level of accuracy rate could be achieved which is suitable for using brain signal as another biometric security authentication

    Multi-modal association learning using spike-timing dependent plasticity (STDP)

    Get PDF
    We propose an associative learning model that can integrate facial images with speech signals to target a subject in a reinforcement learning (RL) paradigm. Through this approach, the rules of learning will involve associating paired stimuli (stimulus–stimulus, i.e., face–speech), which is also known as predictor-choice pairs. Prior to a learning simulation, we extract the features of the biometrics used in the study. For facial features, we experiment by using two approaches: principal component analysis (PCA)-based Eigenfaces and singular value decomposition (SVD). For speech features, we use wavelet packet decomposition (WPD). The experiments show that the PCA-based Eigenfaces feature extraction approach produces better results than SVD. We implement the proposed learning model by using the Spike- Timing-Dependent Plasticity (STDP) algorithm, which depends on the time and rate of pre-post synaptic spikes. The key contribution of our study is the implementation of learning rules via STDP and firing rate in spatiotemporal neural networks based on the Izhikevich spiking model. In our learning, we implement learning for response group association by following the reward-modulated STDP in terms of RL, wherein the firing rate of the response groups determines the reward that will be given. We perform a number of experiments that use existing face samples from the Olivetti Research Laboratory (ORL) dataset, and speech samples from TIDigits. After several experiments and simulations are performed to recognize a subject, the results show that the proposed learning model can associate the predictor (face) with the choice (speech) at optimum performance rates of 77.26% and 82.66% for training and testing, respectively. We also perform learning by using real data, that is, an experiment is conducted on a sample of face–speech data, which have been collected in a manner similar to that of the initial data. The performance results are 79.11% and 77.33% for training and testing, respectively. Based on these results, the proposed learning model can produce high learning performance in terms of combining heterogeneous data (face–speech). This finding opens possibilities to expand RL in the field of biometric authenticatio

    An overview of machine learning and 5G for people with disabilities

    Get PDF
    Currently, over a billion people, including children (or about 15% of the world’s population), are estimated to be living with disability, and this figure is going to increase to beyond two billion by 2050. People with disabilities generally experience poorer levels of health, fewer achievements in education, fewer economic opportunities, and higher rates of poverty. Artificial intelligence and 5G can make major contributions towards the assistance of people with disabilities, so they can achieve a good quality of life. In this paper, an overview of machine learning and 5G for people with disabilities is provided. For this purpose, the proposed 5G network slicing architecture for disabled people is introduced. Different application scenarios and their main benefits are considered to illustrate the interaction of machine learning and 5G. Critical challenges have been identified and addressed.This work has been supported by the Agencia Estatal de Investigación of Ministerio de Ciencia e Innovación of Spain under project PID2019-108713RB-C51 MCIN/ AEI /10.13039/501100011033.Postprint (published version
    • …
    corecore