3 research outputs found

    Towards automated eyewitness descriptions: describing the face, body and clothing for recognition

    No full text
    A fusion approach to person recognition is presented here outlining the automated recognition of targets from human descriptions of face, body and clothing. Three novel results are highlighted. First, the present work stresses the value of comparative descriptions (he is taller than…) over categorical descriptions (he is tall). Second, it stresses the primacy of the face over body and clothing cues for recognition. Third, the present work unequivocally demonstrates the benefit gained through the combination of cues: recognition from face, body and clothing taken together far outstrips recognition from any of the cues in isolation. Moreover, recognition from body and clothing taken together nearly equals the recognition possible from the face alone. These results are discussed with reference to the intelligent fusion of information within police investigations. However, they also signal a potential new era in which automated descriptions could be provided without the need for human witnesses at all

    Analysing comparative soft biometrics from crowdsourced annotations

    No full text
    Soft biometrics enable human description and identification from low-quality surveillance footage. This study premises the design, collection and analysis of a novel crowdsourced dataset of comparative soft biometric body annotations, obtained from a richly diverse set of human annotators. The authors annotate 100 subject images to provide a coherent, in-depth appraisal of the collected annotations and inferred relative labels. The dataset includes gender as a comparative trait and the authors find that comparative labels characteristically contain additional discriminative information over traditional categorical annotations. Using the authors' pragmatic dataset, semantic recognition is performed by inferring relative biometric signatures using a RankSVM algorithm. This demonstrates a practical scenario, reproducing responses from a video surveillance operator searching for an individual. The approach can reliably return the correct match in the top 7% of results with ten comparisons, or top 13% of results using just five sets of subject comparisons
    corecore