345 research outputs found

    Multimodal Three Dimensional Scene Reconstruction, The Gaussian Fields Framework

    Get PDF
    The focus of this research is on building 3D representations of real world scenes and objects using different imaging sensors. Primarily range acquisition devices (such as laser scanners and stereo systems) that allow the recovery of 3D geometry, and multi-spectral image sequences including visual and thermal IR images that provide additional scene characteristics. The crucial technical challenge that we addressed is the automatic point-sets registration task. In this context our main contribution is the development of an optimization-based method at the core of which lies a unified criterion that solves simultaneously for the dense point correspondence and transformation recovery problems. The new criterion has a straightforward expression in terms of the datasets and the alignment parameters and was used primarily for 3D rigid registration of point-sets. However it proved also useful for feature-based multimodal image alignment. We derived our method from simple Boolean matching principles by approximation and relaxation. One of the main advantages of the proposed approach, as compared to the widely used class of Iterative Closest Point (ICP) algorithms, is convexity in the neighborhood of the registration parameters and continuous differentiability, allowing for the use of standard gradient-based optimization techniques. Physically the criterion is interpreted in terms of a Gaussian Force Field exerted by one point-set on the other. Such formulation proved useful for controlling and increasing the region of convergence, and hence allowing for more autonomy in correspondence tasks. Furthermore, the criterion can be computed with linear complexity using recently developed Fast Gauss Transform numerical techniques. In addition, we also introduced a new local feature descriptor that was derived from visual saliency principles and which enhanced significantly the performance of the registration algorithm. The resulting technique was subjected to a thorough experimental analysis that highlighted its strength and showed its limitations. Our current applications are in the field of 3D modeling for inspection, surveillance, and biometrics. However, since this matching framework can be applied to any type of data, that can be represented as N-dimensional point-sets, the scope of the method is shown to reach many more pattern analysis applications

    NIRFaceNet: A Convolutional Neural Network for Near-Infrared Face Identification

    Get PDF
    Near-infrared (NIR) face recognition has attracted increasing attention because of its advantage of illumination invariance. However, traditional face recognition methods based on NIR are designed for and tested in cooperative-user applications. In this paper, we present a convolutional neural network (CNN) for NIR face recognition (specifically face identification) in non-cooperative-user applications. The proposed NIRFaceNet is modified from GoogLeNet, but has a more compact structure designed specifically for the Chinese Academy of Sciences Institute of Automation (CASIA) NIR database and can achieve higher identification rates with less training time and less processing time. The experimental results demonstrate that NIRFaceNet has an overall advantage compared to other methods in the NIR face recognition domain when image blur and noise are present. The performance suggests that the proposed NIRFaceNet method may be more suitable for non-cooperative-user applications

    An Extensive Review on Spectral Imaging in Biometric Systems: Challenges and Advancements

    Full text link
    Spectral imaging has recently gained traction for face recognition in biometric systems. We investigate the merits of spectral imaging for face recognition and the current challenges that hamper the widespread deployment of spectral sensors for face recognition. The reliability of conventional face recognition systems operating in the visible range is compromised by illumination changes, pose variations and spoof attacks. Recent works have reaped the benefits of spectral imaging to counter these limitations in surveillance activities (defence, airport security checks, etc.). However, the implementation of this technology for biometrics, is still in its infancy due to multiple reasons. We present an overview of the existing work in the domain of spectral imaging for face recognition, different types of modalities and their assessment, availability of public databases for sake of reproducible research as well as evaluation of algorithms, and recent advancements in the field, such as, the use of deep learning-based methods for recognizing faces from spectral images

    Performances of proposed normalization algorithm for iris recognition

    Get PDF
    Iris recognition has very high recognition accuracy in comparison with many other biometric features. The iris pattern is not the same even right and left eye of the same person. It is different and unique. This paper proposes an algorithm to recognize people based on iris images. The algorithm consists of three stages. In the first stage, the segmentation process is using circular Hough transforms to find the region of interest (ROI) of given eye images. After that, a proposed normalization algorithm is to generate the polar images than to enhance the polar images using a modified Daugman’s Rubber sheet model. The last step of the proposed algorithm is to divide the enhance the polar image to be 16 divisions of the iris region. The normalized image is 16 small constant dimensions. The Gray-Level Co-occurrence Matrices (GLCM) technique calculates and extracts the normalized image’s texture feature. Here, the features extracted are contrast, correlation, energy, and homogeneity of the iris. In the last stage, a classification technique, discriminant analysis (DA), is employed for analysis of the proposed normalization algorithm. We have compared the proposed normalization algorithm to the other nine normalization algorithms. The DA technique produces an excellent classification performance with 100% accuracy. We also compare our results with previous results and find out that the proposed iris recognition algorithm is an effective system to detect and recognize person digitally, thus it can be used for security in the building, airports, and other automation in many applications

    Connected Attribute Filtering Based on Contour Smoothness

    Get PDF

    Reduced-order modelling for high-speed aerial weapon aerodynamics

    Get PDF
    In this work a high-fidelity low-cost surrogate of a computational fluid dynamics analysis tool was developed. This computational tool is composed of general and physics- based approximation methods by which three dimensional high-speed aerodynamic flow- field predictions are made with high efficiency and an accuracy which is comparable with that of CFD. The tool makes use of reduced-basis methods that are suitable for both linear and non-linear problems, whereby the basis vectors are computed via the proper orthogonal decomposition (POD) of a training dataset or a set of observations. The surrogate model was applied to two flow problems related to high-speed weapon aerodynamics. Comparisons of surrogate model predictions with high-fidelity CFD simulations suggest that POD-based reduced-order modelling together with response surface methods provide a reliable and robust approach for efficient and accurate predictions. In contrast to the many modelling efforts reported in the literature, this surrogate model provides access to information about the whole flow-field. In an attempt to reduce the up-front cost necessary to generate the training dataset from which the surrogate model is subsequently developed, a variable-fidelity POD- based reduced-order modelling method is proposed in this work for the first time. In this model, the scalar coefficients which are obtained by projecting the solution vectors onto the basis vectors, are mapped between spaces of low and high fidelities, to achieve high- fidelity predictions with complete flow-field information. In general, this technique offers an automatic way of fusing variable-fidelity data through interpolation and extrapolation schemes together with reduced-order modelling (ROM). Furthermore, a study was undertaken to investigate the possibility of modelling the transonic flow over an aerofoil using a kernel POD–based reduced-order modelling method. By using this type of ROM it was noticed that the weak non-linear features of the transonic flow are accurately modelled using a small number of basis vectors. The strong non-linear features are only modelled accurately by using a large number of basis vectors

    Signal processing algorithms for enhanced image fusion performance and assessment

    Get PDF
    The dissertation presents several signal processing algorithms for image fusion in noisy multimodal conditions. It introduces a novel image fusion method which performs well for image sets heavily corrupted by noise. As opposed to current image fusion schemes, the method has no requirements for a priori knowledge of the noise component. The image is decomposed with Chebyshev polynomials (CP) being used as basis functions to perform fusion at feature level. The properties of CP, namely fast convergence and smooth approximation, renders it ideal for heuristic and indiscriminate denoising fusion tasks. Quantitative evaluation using objective fusion assessment methods show favourable performance of the proposed scheme compared to previous efforts on image fusion, notably in heavily corrupted images. The approach is further improved by incorporating the advantages of CP with a state-of-the-art fusion technique named independent component analysis (ICA), for joint-fusion processing based on region saliency. Whilst CP fusion is robust under severe noise conditions, it is prone to eliminating high frequency information of the images involved, thereby limiting image sharpness. Fusion using ICA, on the other hand, performs well in transferring edges and other salient features of the input images into the composite output. The combination of both methods, coupled with several mathematical morphological operations in an algorithm fusion framework, is considered a viable solution. Again, according to the quantitative metrics the results of our proposed approach are very encouraging as far as joint fusion and denoising are concerned. Another focus of this dissertation is on a novel metric for image fusion evaluation that is based on texture. The conservation of background textural details is considered important in many fusion applications as they help define the image depth and structure, which may prove crucial in many surveillance and remote sensing applications. Our work aims to evaluate the performance of image fusion algorithms based on their ability to retain textural details from the fusion process. This is done by utilising the gray-level co-occurrence matrix (GLCM) model to extract second-order statistical features for the derivation of an image textural measure, which is then used to replace the edge-based calculations in an objective-based fusion metric. Performance evaluation on established fusion methods verifies that the proposed metric is viable, especially for multimodal scenarios

    3D Face Recognition

    Get PDF
    • 

    corecore