62,479 research outputs found

    Computer-based tracking, analysis, and visualization of linguistically significant nonmanual events in American Sign Language (ASL)

    Full text link
    Our linguistically annotated American Sign Language (ASL) corpora have formed a basis for research to automate detection by computer of essential linguistic information conveyed through facial expressions and head movements. We have tracked head position and facial deformations, and used computational learning to discern specific grammatical markings. Our ability to detect, identify, and temporally localize the occurrence of such markings in ASL videos has recently been improved by incorporation of (1) new techniques for deformable model-based 3D tracking of head position and facial expressions, which provide significantly better tracking accuracy and recover quickly from temporary loss of track due to occlusion; and (2) a computational learning approach incorporating 2-level Conditional Random Fields (CRFs), suited to the multi-scale spatio-temporal characteristics of the data, which analyses not only low-level appearance characteristics, but also the patterns that enable identification of significant gestural components, such as periodic head movements and raised or lowered eyebrows. Here we summarize our linguistically motivated computational approach and the results for detection and recognition of nonmanual grammatical markings; demonstrate our data visualizations, and discuss the relevance for linguistic research; and describe work underway to enable such visualizations to be produced over large corpora and shared publicly on the Web

    3D face tracking and multi-scale, spatio-temporal analysis of linguistically significant facial expressions and head positions in ASL

    Full text link
    Essential grammatical information is conveyed in signed languages by clusters of events involving facial expressions and movements of the head and upper body. This poses a significant challenge for computer-based sign language recognition. Here, we present new methods for the recognition of nonmanual grammatical markers in American Sign Language (ASL) based on: (1) new 3D tracking methods for the estimation of 3D head pose and facial expressions to determine the relevant low-level features; (2) methods for higher-level analysis of component events (raised/lowered eyebrows, periodic head nods and head shakes) used in grammatical markings—with differentiation of temporal phases (onset, core, offset, where appropriate), analysis of their characteristic properties, and extraction of corresponding features; (3) a 2-level learning framework to combine lowand high-level features of differing spatio-temporal scales. This new approach achieves significantly better tracking and recognition results than our previous methods

    Recognition of nonmanual markers in American Sign Language (ASL) using non-parametric adaptive 2D-3D face tracking

    Full text link
    This paper addresses the problem of automatically recognizing linguistically significant nonmanual expressions in American Sign Language from video. We develop a fully automatic system that is able to track facial expressions and head movements, and detect and recognize facial events continuously from video. The main contributions of the proposed framework are the following: (1) We have built a stochastic and adaptive ensemble of face trackers to address factors resulting in lost face track; (2) We combine 2D and 3D deformable face models to warp input frames, thus correcting for any variation in facial appearance resulting from changes in 3D head pose; (3) We use a combination of geometric features and texture features extracted from a canonical frontal representation. The proposed new framework makes it possible to detect grammatically significant nonmanual expressions from continuous signing and to differentiate successfully among linguistically significant expressions that involve subtle differences in appearance. We present results that are based on the use of a dataset containing 330 sentences from videos that were collected and linguistically annotated at Boston University

    Under-explicit and minimally explicit reference: Evidence from a longitudinal case study

    Get PDF
    This chapter reports on a 2 ½ year longitudinal case study of one Korean speaker of English, focusing on the development of her command of accessibility marking in referring to persons. The data are derived from informal, open interviews spanning the entire length of the participant’s enrolment in a Bachelor of Nursing programme in New Zealand. These interviews occurred every few weeks during semester (17 in total), and were typically between 45 minutes to one hour in length. The participant reported that she used these interviews as “a kind of reflective journal”, in which she discussed her classes, interactions with classmates, tutors and others, her assignments, and other experiences in New Zealand. The events she reported are rich in references to individuals. Using a previously reported coding scheme (Ryan, 2015), these data were analysed in relation to pragmatic felicity, particularly concerning the felicity of accessibility marking for referents of varying cognitive status in contexts of topic or focus continuity or shift. These data [yet to be analysed] provide evidence of the developmental progression of the participant’s command of reference in English. This chapter contributes substantially to the literature in several ways. In general, there has been a lack of longitudinal case studies of pragmatic development in any domain, including few – if any – previous longitudinal studies focusing on reference; the present analysis is therefore expected to reveal previously unreported details of the trajectory of pragmatic development in reference. The present study is also one of the few working with oral data that was generated in ways other than an elicited communication task. Finally, the study contributes to the somewhat still contentious issue of to what extent mainstream study in an English-speaking context leads to genuine language gains

    Towards responsive Sensitive Artificial Listeners

    Get PDF
    This paper describes work in the recently started project SEMAINE, which aims to build a set of Sensitive Artificial Listeners – conversational agents designed to sustain an interaction with a human user despite limited verbal skills, through robust recognition and generation of non-verbal behaviour in real-time, both when the agent is speaking and listening. We report on data collection and on the design of a system architecture in view of real-time responsiveness
    • …
    corecore