2 research outputs found

    Real-time action recognition using a multilayer descriptor with variable size

    Get PDF
    Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)Video analysis technology has become less expensive and more powerful in terms of storage resources and resolution capacity, promoting progress in a wide range of applications. Video-based human action detection has been used for several tasks in surveillance environments, such as forensic investigation, patient monitoring, medical training, accident prevention, and traffic monitoring, among others. We present a method for action identification based on adaptive training of a multilayer descriptor applied to a single classifier. Cumulative motion shapes (CMSs) are extracted according to the number of frames present in the video. Each CMS is employed as a self-sufficient layer in the training stage but belongs to the same descriptor. A robust classification is achieved through individual responses of classifiers for each layer, and the dominant result is used as a final outcome. Experiments are conducted on five public datasets (Weizmann, KTH, MuHAVi, IXMAS, and URADL) to demonstrate the effectiveness of the method in terms of accuracy in real time. (C) 2016 SPIE and IS&TVideo analysis technology has become less expensive and more powerful in terms of storage resources and resolution capacity, promoting progress in a wide range of applications. Video-based human action detection has been used for several tasks in surveill2501FAPESP - FUNDAÇÃO DE AMPARO À PESQUISA DO ESTADO DE SÃO PAULOCNPQ - CONSELHO NACIONAL DE DESENVOLVIMENTO CIENTÍFICO E TECNOLÓGICOFundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)SEM INFORMAÇÃOSEM INFORMAÇÃ

    Analyse du contenu expressif des gestes corporels

    Get PDF
    Nowadays, researches dealing with gesture analysis suffer from a lack of unified mathematical models. On the one hand, gesture formalizations by human sciences remain purely theoretical and are not inclined to any quantification. On the other hand, the commonly used motion descriptors are generally purely intuitive, and limited to the visual aspects of the gesture. In the present work, we retain Laban Movement Analysis (LMA – originally designed for the study of dance movements) as a framework for building our own gesture descriptors, based on expressivity. Two datasets are introduced: the first one is called ORCHESTRE-3D, and is composed of pre-segmented orchestra conductors’ gestures, which have been annotated with the help of lexicon of musical emotions. The second one, HTI 2014-2015, comprises sequences of multiple daily actions. In a first experiment, we define a global feature vector based upon the expressive indices of our model and dedicated to the characterization of the whole gesture. This descriptor is used for action recognition purpose and to discriminate the different emotions of our orchestra conductors’ dataset. In a second approach, the different elements of our expressive model are used as a frame descriptor (e.g., describing the gesture at a given time). The feature space provided by such local characteristics is used to extract key poses of the motion. With the help of such poses, we obtain a per-frame sub-representation of body motions which is available for real-time action recognition purposeAujourd’hui, les recherches portant sur le geste manquent de modĂšles gĂ©nĂ©riques. Les spĂ©cialistes du geste doivent osciller entre une formalisation excessivement conceptuelle et une description purement visuelle du mouvement. Nous reprenons les concepts dĂ©veloppĂ©s par le chorĂ©graphe Rudolf Laban pour l’analyse de la danse classique contemporaine, et proposons leur extension afin d’élaborer un modĂšle gĂ©nĂ©rique du geste basĂ© sur ses Ă©lĂ©ments expressifs. Nous prĂ©sentons Ă©galement deux corpus de gestes 3D que nous avons constituĂ©s. Le premier, ORCHESTRE-3D, se compose de gestes prĂ©-segmentĂ©s de chefs d’orchestre enregistrĂ©s en rĂ©pĂ©tition. Son annotation Ă  l’aide d’émotions musicales est destinĂ©e Ă  l’étude du contenu Ă©motionnel de la direction musicale. Le deuxiĂšme corpus, HTI 2014-2015, propose des sĂ©quences d’actions variĂ©es de la vie quotidienne. Dans une premiĂšre approche de reconnaissance dite « globale », nous dĂ©finissons un descripteur qui se rapporte Ă  l’entiĂšretĂ© du geste. Ce type de caractĂ©risation nous permet de discriminer diverses actions, ainsi que de reconnaĂźtre les diffĂ©rentes Ă©motions musicales que portent les gestes des chefs d’orchestre de notre base ORCHESTRE-3D. Dans une seconde approche dite « dynamique », nous dĂ©finissons un descripteur de trame gestuelle (e.g. dĂ©fini pour tout instant du geste). Les descripteurs de trame sont utilisĂ©s des poses-clĂ©s du mouvement, de sorte Ă  en obtenir Ă  tout instant une reprĂ©sentation simplifiĂ©e et utilisable pour reconnaĂźtre des actions Ă  la volĂ©e. Nous testons notre approche sur plusieurs bases de geste, dont notre propre corpus HTI 2014-201
    corecore