1,334 research outputs found

    Integrating RBF-based Neural Network Face Expression Recognition in Access System

    Get PDF
    Biometric recognition system such as facial recognition system was widely developed over the past few years. Facial recognition system is commonly used in security system to allow user to protect their privilege. The normal security like key or password is no longer relevant as people prefer an easier and flexible way. Therefore, this paper presents a better and easier way of security system that can recognize the user successfully and give the matching percentage. By using Radial Basis Function Neural Network in MATLAB, a face recognition system can be created. The RBF system will be trained by data as reference, input image will undergo the same process and the data obtained will be used to match with the data in the RBF to obtain the matching percentage. A suitable matching percentage reference was chosen from this analysis as the minimum require matching to access the security system where error rate is one of the main concerns where it is the unwanted result that might occur. Different threshold number, spread value, and sizes of dimension also tested, the differences on the output matching result were observed. By using the microcontroller to control a relay to control the magnetic door lock, the system was able to successfully control the door lock

    Deep Reflectance Maps

    Get PDF
    Undoing the image formation process and therefore decomposing appearance into its intrinsic properties is a challenging task due to the under-constraint nature of this inverse problem. While significant progress has been made on inferring shape, materials and illumination from images only, progress in an unconstrained setting is still limited. We propose a convolutional neural architecture to estimate reflectance maps of specular materials in natural lighting conditions. We achieve this in an end-to-end learning formulation that directly predicts a reflectance map from the image itself. We show how to improve estimates by facilitating additional supervision in an indirect scheme that first predicts surface orientation and afterwards predicts the reflectance map by a learning-based sparse data interpolation. In order to analyze performance on this difficult task, we propose a new challenge of Specular MAterials on SHapes with complex IllumiNation (SMASHINg) using both synthetic and real images. Furthermore, we show the application of our method to a range of image-based editing tasks on real images.Comment: project page: http://homes.esat.kuleuven.be/~krematas/DRM

    End-to-end 3D face reconstruction with deep neural networks

    Full text link
    Monocular 3D facial shape reconstruction from a single 2D facial image has been an active research area due to its wide applications. Inspired by the success of deep neural networks (DNN), we propose a DNN-based approach for End-to-End 3D FAce Reconstruction (UH-E2FAR) from a single 2D image. Different from recent works that reconstruct and refine the 3D face in an iterative manner using both an RGB image and an initial 3D facial shape rendering, our DNN model is end-to-end, and thus the complicated 3D rendering process can be avoided. Moreover, we integrate in the DNN architecture two components, namely a multi-task loss function and a fusion convolutional neural network (CNN) to improve facial expression reconstruction. With the multi-task loss function, 3D face reconstruction is divided into neutral 3D facial shape reconstruction and expressive 3D facial shape reconstruction. The neutral 3D facial shape is class-specific. Therefore, higher layer features are useful. In comparison, the expressive 3D facial shape favors lower or intermediate layer features. With the fusion-CNN, features from different intermediate layers are fused and transformed for predicting the 3D expressive facial shape. Through extensive experiments, we demonstrate the superiority of our end-to-end framework in improving the accuracy of 3D face reconstruction.Comment: Accepted to CVPR1

    INTELLIGENT VISION-BASED NAVIGATION SYSTEM

    Get PDF
    This thesis presents a complete vision-based navigation system that can plan and follow an obstacle-avoiding path to a desired destination on the basis of an internal map updated with information gathered from its visual sensor. For vision-based self-localization, the system uses new floor-edges-specific filters for detecting floor edges and their pose, a new algorithm for determining the orientation of the robot, and a new procedure for selecting the initial positions in the self-localization procedure. Self-localization is based on matching visually detected features with those stored in a prior map. For planning, the system demonstrates for the first time a real-world application of the neural-resistive grid method to robot navigation. The neural-resistive grid is modified with a new connectivity scheme that allows the representation of the collision-free space of a robot with finite dimensions via divergent connections between the spatial memory layer and the neuro-resistive grid layer. A new control system is proposed. It uses a Smith Predictor architecture that has been modified for navigation applications and for intermittent delayed feedback typical of artificial vision. A receding horizon control strategy is implemented using Normalised Radial Basis Function nets as path encoders, to ensure continuous motion during the delay between measurements. The system is tested in a simplified environment where an obstacle placed anywhere is detected visually and is integrated in the path planning process. The results show the validity of the control concept and the crucial importance of a robust vision-based self-localization process

    Learning a Hierarchical Latent-Variable Model of 3D Shapes

    Full text link
    We propose the Variational Shape Learner (VSL), a generative model that learns the underlying structure of voxelized 3D shapes in an unsupervised fashion. Through the use of skip-connections, our model can successfully learn and infer a latent, hierarchical representation of objects. Furthermore, realistic 3D objects can be easily generated by sampling the VSL's latent probabilistic manifold. We show that our generative model can be trained end-to-end from 2D images to perform single image 3D model retrieval. Experiments show, both quantitatively and qualitatively, the improved generalization of our proposed model over a range of tasks, performing better or comparable to various state-of-the-art alternatives.Comment: Accepted as oral presentation at International Conference on 3D Vision (3DV), 201

    Artificial Neural Networks as Decision-Makers for Stereo Matching

    Get PDF
    This paper investigates the use of artificial neural networks to help making a decision on matching of stereo images. An image matching technique based on extracting features from segmented regions is adopted in this work, and a neural network framework is applied for region matching of stereo photographs. Two types of neural networks are used, the radial basis network, (RB) for learning clustering, and the back propagation (BP) network for learning image matching. The (RB) neural network is to cluster the regions according to the locations of their centered points. For each region, the BP network uses differential features as input training data. While training and testing the system, multiple features are extracted and used for enhancing the accuracy of the matching process. Features include (compactness, Euler number, and invariant moments) for each region. Results obtained from the neural networks (namely; clustering and initial matching array) are used to select the best matching pair. Results are showing a good matching accuracy
    • …
    corecore