620 research outputs found

    3D human pose estimation from depth maps using a deep combination of poses

    Full text link
    Many real-world applications require the estimation of human body joints for higher-level tasks as, for example, human behaviour understanding. In recent years, depth sensors have become a popular approach to obtain three-dimensional information. The depth maps generated by these sensors provide information that can be employed to disambiguate the poses observed in two-dimensional images. This work addresses the problem of 3D human pose estimation from depth maps employing a Deep Learning approach. We propose a model, named Deep Depth Pose (DDP), which receives a depth map containing a person and a set of predefined 3D prototype poses and returns the 3D position of the body joints of the person. In particular, DDP is defined as a ConvNet that computes the specific weights needed to linearly combine the prototypes for the given input. We have thoroughly evaluated DDP on the challenging 'ITOP' and 'UBC3V' datasets, which respectively depict realistic and synthetic samples, defining a new state-of-the-art on them.Comment: Accepted for publication at "Journal of Visual Communication and Image Representation

    Eye Size and Shape in Relation to Refractive Error in Children:A Magnetic Resonance Imaging Study

    Get PDF
    Purpose: The purpose of this study was to determine the association between eye shape and volume measured with magnetic resonance imaging (MRI) and optical biometry and with spherical equivalent (SE) in children. Methods: For this study, there were 3637 10-year-old children from a population-based birth-cohort study that underwent optical biometry (IOL-master 500) and T2-weighted MRI scanning (height, width, and volume). Cycloplegic refractive error was determined by automated refraction. The MRI images of the eyes were segmented using an automated algorithm combining atlas registration with voxel classification. Associations among optical biometry, anthropometry, MRI measurements, and RE were tested using Pearson correlation. Differences between refractive error groups were tested using ANOVA. Results: The mean volume of the posterior segment was 6350 (±680) mm3. Myopic eyes (SE ≤ -0.5 diopters [D]) had 470 mm3 (P &lt; 0.001) and 970 mm3 (P &lt; 0.001) larger posterior segment volume than emmetropic and hyperopic eyes (SE ≥ +2.0D), respectively. The majority of eyes (77.1%) had an oblate shape, but 47.4% of myopic eyes had a prolate shape versus 3.9% of hyperopic eyes. The correlation between SE and MRI-derived posterior segment length (r -0.51, P &lt; 0.001) was stronger than the correlation with height (r -0.30, P &lt; 0.001) or width of the eye (r -0.10, P &lt; 0.001). Conclusions: In this study, eye shape at 10 years of age was predominantly oblate, even in eyes with myopia. Of all MRI measurements, posterior segment length was most prominently associated with SE. Whether eye shape predicts future myopia development or progression should be investigated in longitudinal studies.</p

    Eye Size and Shape in Relation to Refractive Error in Children:A Magnetic Resonance Imaging Study

    Get PDF
    Purpose: The purpose of this study was to determine the association between eye shape and volume measured with magnetic resonance imaging (MRI) and optical biometry and with spherical equivalent (SE) in children. Methods: For this study, there were 3637 10-year-old children from a population-based birth-cohort study that underwent optical biometry (IOL-master 500) and T2-weighted MRI scanning (height, width, and volume). Cycloplegic refractive error was determined by automated refraction. The MRI images of the eyes were segmented using an automated algorithm combining atlas registration with voxel classification. Associations among optical biometry, anthropometry, MRI measurements, and RE were tested using Pearson correlation. Differences between refractive error groups were tested using ANOVA. Results: The mean volume of the posterior segment was 6350 (±680) mm3. Myopic eyes (SE ≤ -0.5 diopters [D]) had 470 mm3 (P &lt; 0.001) and 970 mm3 (P &lt; 0.001) larger posterior segment volume than emmetropic and hyperopic eyes (SE ≥ +2.0D), respectively. The majority of eyes (77.1%) had an oblate shape, but 47.4% of myopic eyes had a prolate shape versus 3.9% of hyperopic eyes. The correlation between SE and MRI-derived posterior segment length (r -0.51, P &lt; 0.001) was stronger than the correlation with height (r -0.30, P &lt; 0.001) or width of the eye (r -0.10, P &lt; 0.001). Conclusions: In this study, eye shape at 10 years of age was predominantly oblate, even in eyes with myopia. Of all MRI measurements, posterior segment length was most prominently associated with SE. Whether eye shape predicts future myopia development or progression should be investigated in longitudinal studies.</p

    Automating the human action of first-trimester biometry measurement from real-world freehand ultrasound

    Get PDF
    Objective: Automated medical image analysis solutions should closely mimic complete human actions to be useful in clinical practice. However, more often an automated image analysis solution represents only part of a human task, which restricts its practical utility. In the case of ultrasound-based fetal biometry, an automated solution should ideally recognize key fetal structures in freehand video guidance, select a standard plane from a video stream and perform biometry. A complete automated solution should automate all three subactions. Methods: In this article, we consider how to automate the complete human action of first-trimester biometry measurement from real-world freehand ultrasound. In the proposed hybrid convolutional neural network (CNN) architecture design, a classification regression-based guidance model detects and tracks fetal anatomical structures (using visual cues) in the ultrasound video. Several high-quality standard planes that contain the mid-sagittal view of the fetus are sampled at multiple time stamps (using a custom-designed confident-frame detector) based on the estimated probability values associated with predicted anatomical structures that define the biometry plane. Automated semantic segmentation is performed on the selected frames to extract fetal anatomical landmarks. A crown–rump length (CRL) estimate is calculated as the mean CRL from these multiple frames. Results: Our fully automated method has a high correlation with clinical expert CRL measurement (Pearson's p = 0.92, R-squared [R2] = 0.84) and a low mean absolute error of 0.834 (weeks) for fetal age estimation on a test data set of 42 videos. Conclusion: A novel algorithm for standard plane detection employs a quality detection mechanism defined by clinical standards, ensuring precise biometric measurements

    Biometric recognition through gait analysis

    Get PDF
    [EN] The use of people recognition techniques has become critical in some areas. For instance, social or assistive robots carry out collaborative tasks in the robotics field. A robot must know who to work with to deal with such tasks. Using biometric patterns may replace identification cards or codes on access control to critical infrastructures. The usage of Red Green Blue Depth (RGBD) cameras is ubiquitous to solve people recognition. However, this sensor has some constraints, such as they demand high computational capabilities, require the users to face the sensor, or do not regard users' privacy. Furthermore, in the COVID-19 pandemic, masks hide a significant portion of the face. In this work, we present BRITTANY, a biometric recognition tool through gait analysis using Laser Imaging Detection and Ranging (LIDAR) data and a Convolutional Neural Network (CNN). A Proof of Concept (PoC) has been carried out in an indoor environment with five users to evaluate BRITTANY. A new CNN architecture is presented, allowing the classification of aggregated occupancy maps that represent the people's gait. This new architecture has been compared with LeNet-5 and AlexNet through the same datasets. The final system reports an accuracy of 88%.SIInstituto Nacional de Ciberseguridad de Espana (INCIBE)The research described in this article has been funded by the Instituto Nacional de Ciberseguridad de España (INCIBE), under the grant ”ADENDA 4: Detección de nuevas amenazas y patrones desconocidos (Red Regional de Ciencia y Tecnología)”, addendum to the framework agreement INCIBE-Universidad de León, 2019-2021. Miguel Ángel González-Santamarta would like to thank Universidad de León for its funding support for his doctoral studies

    IMI - Clinical Myopia Control Trials and Instrumentation Report

    Get PDF
    The evidence-basis based on existing myopia control trials along with the supporting academic literature were reviewed; this informed recommendations on the outcomes suggested from clinical trials aimed at slowing myopia progression to show the effectiveness of treatments and the impact on patients. These outcomes were classified as primary (refractive error and/or axial length), secondary (patient reported outcomes and treatment compliance), and exploratory (peripheral refraction, accommodative changes, ocular alignment, pupil size, outdoor activity/lighting levels, anterior and posterior segment imaging, and tissue biomechanics). The currently available instrumentation, which the literature has shown to best achieve the primary and secondary outcomes, was reviewed and critiqued. Issues relating to study design and patient selection were also identified. These findings and consensus from the International Myopia Institute members led to final recommendations to inform future instrumentation development and to guide clinical trial protocols

    Design of a Multi-biometric Platform, based on physical traits and physiological measures: Face, Iris, Ear, ECG and EEG

    Get PDF
    Security and safety is one the main concerns both for governments and for private companies in the last years so raising growing interests and investments in the area of biometric recognition and video surveillance, especially after the sad happenings of September 2001. Outlays assessments of the U.S. government for the years 2001-2005 estimate that the homeland security spending climbed from 56.0billionsofdollarsin2001toalmost56.0 billions of dollars in 2001 to almost 100 billion of 2005. In this lapse of time, new pattern recognition techniques have been developed and, even more important, new biometric traits have been investigated and refined; besides the well-known physical and behavioral characteristics, also physiological measures have been studied, so providing more features to enhance discrimination capabilities of individuals. This dissertation proposes the design of a multimodal biometric platform, FAIRY, based on the following biometric traits: ear, face, iris EEG and ECG signals. In the thesis the modular architecture of the platform has been presented, together with the results obtained for the solution to the recognition problems related to the different biometrics and their possible fusion. Finally, an analysis of the pattern recognition issues concerning the area of videosurveillance has been discussed

    Design of a Multi-biometric Platform, based on physical traits and physiological measures: Face, Iris, Ear, ECG and EEG

    Get PDF
    Security and safety is one the main concerns both for governments and for private companies in the last years so raising growing interests and investments in the area of biometric recognition and video surveillance, especially after the sad happenings of September 2001. Outlays assessments of the U.S. government for the years 2001-2005 estimate that the homeland security spending climbed from 56.0billionsofdollarsin2001toalmost56.0 billions of dollars in 2001 to almost 100 billion of 2005. In this lapse of time, new pattern recognition techniques have been developed and, even more important, new biometric traits have been investigated and refined; besides the well-known physical and behavioral characteristics, also physiological measures have been studied, so providing more features to enhance discrimination capabilities of individuals. This dissertation proposes the design of a multimodal biometric platform, FAIRY, based on the following biometric traits: ear, face, iris EEG and ECG signals. In the thesis the modular architecture of the platform has been presented, together with the results obtained for the solution to the recognition problems related to the different biometrics and their possible fusion. Finally, an analysis of the pattern recognition issues concerning the area of videosurveillance has been discussed

    Human Part Segmentation in Depth Images with Annotated Part Positions

    Get PDF
    We present a method of segmenting human parts in depth images, when provided the image positions of the body parts. The goal is to facilitate per-pixel labelling of large datasets of human images, which are used for training and testing algorithms for pose estimation and automatic segmentation. A common technique in image segmentation is to represent an image as a two-dimensional grid graph, with one node for each pixel and edges between neighbouring pixels. We introduce a graph with distinct layers of nodes to model occlusion of the body by the arms. Once the graph is constructed, the annotated part positions are used as seeds for a standard interactive segmentation algorithm. Our method is evaluated on two public datasets containing depth images of humans from a frontal view. It produces a mean per-class accuracy of 93.55% on the first dataset, compared to 87.91% (random forest and graph cuts) and 90.31% (random forest and Markov random field). It also achieves a per-class accuracy of 90.60% on the second dataset. Future work can experiment with various methods for creating the graph layers to accurately model occlusion
    corecore