1,110 research outputs found
Social behavior modeling based on Incremental Discrete Hidden Markov Models
12 pagesInternational audienceModeling multimodal face-to-face interaction is a crucial step in the process of building social robots or users-aware Embodied Conversational Agents (ECA). In this context, we present a novel approach for human behavior analysis and generation based on what we called "Incremental Discrete Hidden Markov Model" (IDHMM). Joint multimodal activities of interlocutors are first modeled by a set of DHMMs that are specific to supposed joint cognitive states of the interlocutors. Respecting a task-specific syntax, the IDHMM is then built from these DHMMs and split into i) a recognition model that will determine the most likely sequence of cognitive states given the multimodal activity of the in- terlocutor, and ii) a generative model that will compute the most likely activity of the speaker given this estimated sequence of cognitive states. Short-Term Viterbi (STV) decoding is used to incrementally recognize and generate behav- ior. The proposed model is applied to parallel speech and gaze data of interact- ing dyads
Modeling Perception-Action Loops: Comparing Sequential Models with Frame-Based Classifiers
International audienceModeling multimodal perception-action loops in face-to-face interactions is a crucial step in the process of building sensory-motor behaviors for social robots or users-aware Embodied Conversational Agents (ECA). In this paper, we compare trainable behavioral models based on sequential models (HMMs) and classifiers (SVMs and Decision Trees) inherently inappropriate to model sequential aspects. These models aim at giving pertinent perception/action skills for robots in order to generate optimal actions given the perceived actions of others and joint goals. We applied these models to parallel speech and gaze data collected from interacting dyads. The challenge was to predict the gaze of one subject given the gaze of the interlocutor and the voice activity of both. We show that Incremental Discrete HMM (IDHMM) generally outperforms classifiers and that injecting input context in the modeling process significantly improves the performances of all algorithms
Shifting embodied participation in multiparty university student meetings
PhD ThesisStudent group work has been used in higher education as an effective means to cultivate
students’ work-related skills and cooperative learning. These encounters of small groups are the
sites where, through talk and other resources, university students get their educational tasks
done as well as acquire essential workplace skills such as problem-solving, team working,
decision-making and leadership. However, settings of educational talk-as-work, such as student
group meetings, remain under-researched (Stokoe, Benwell, & Attenborough, 2013). The
present study therefore attempts to bridge this gap by investigating the professional and
academic abilities of university students to participate in multiparty group meetings, drawing
upon a dataset of video- and audio-recorded meetings from the Newcastle University Corpus of
Academic English (NUCASE). The dataset consists of ten hours of meetings in which a group
of naval architecture undergraduate students work cooperatively on their final year project – to
design and build a wind turbine.
The study applies the methodological approach of conversation analysis (CA) with a
multimodal perspective. It presents a fine-detailed, sequential multimodal analysis of a
collection of cases of speaker transitions, and reveals how meeting participants display
speakership and recipiency with their verbal/vocal and bodily-visual coordination. In this
respect, the present study is the first to offer a systematic collection, as well as a thorough
investigation, of speaker transition and turn-taking practices from a multimodal perspective,
especially with the scope of analysis beyond pre-turn and turn-beginning positions. It shows
how speaker transitions through ‘current speaker selects next’ and ‘next speaker self-selects’
are joint-undertakings not only between the self-selecting/current speaker, and the target
recipient/addressed next speaker, but also among other co-present participants. Especially, by
mobilising the whole set of multimodal resources, participants are able to display their multiple
orientations toward their co-participants, project, pursue and accomplish multiple courses of
action in concurrence, and intricately coordinate their mutual orientation toward the shifting
and emerging participation framework during the transition, establishment and maintenance of
the speakership and recipiency. By presenting the data and analysis, this study extends
ii
boundaries of existing understandings on the temporality, sequentiality and systematicity of
multimodal resources in talk-and-bodies-in-interaction.
The thesis also contributes to interaction research in the particular context of student group
work in higher education contexts, by providing a ‘screenshot’ of students’ academic lives as it
unfolds ‘in flight’. Particularly, it reveals how students competently participate in multiparty
group meetings (e.g., taking and allocating turns), co-construct the unfolding meeting
procedures (e.g., roundtable update discussion), and jointly achieve the local interactional goals
(e.g., sharing work progress, reaching an agreement). Acquiring such skills is, as it argues
above, not only crucial for accomplishing the educational tasks, but also necessary for
preparing university students to fulfill their future workplace expectations. The study therefore
further informs the practices of university students and professional practitioners in multiparty
meetings, and also draws on methodological implications for multimodal CA research
Animating Synthetic Dyadic Conversations With Variations Based on Context and Agent Attributes
Conversations between two people are ubiquitous in many inhabited contexts. The kinds of conversations that occur depend on several factors, including the time, the location of the participating agents, the spatial relationship between the agents, and the type of conversation in which they are engaged. The statistical distribution of dyadic conversations among a population of agents will therefore depend on these factors. In addition, the conversation types, flow, and duration will depend on agent attributes such as interpersonal relationships, emotional state, personal priorities, and socio-cultural proxemics. We present a framework for distributing conversations among virtual embodied agents in a real-time simulation. To avoid generating actual language dialogues, we express variations in the conversational flow by using behavior trees implementing a set of conversation archetypes. The flow of these behavior trees depends in part on the agents’ attributes and progresses based on parametrically estimated transitional probabilities. With the participating agents’ state, a ‘smart event’ model steers the interchange to different possible outcomes as it executes. Example behavior trees are developed for two conversation archetypes: buyer–seller negotiations and simple asking–answering; the model can be readily extended to others. Because the conversation archetype is known to participating agents, they can animate their gestures appropriate to their conversational state. The resulting animated conversations demonstrate reasonable variety and variability within the environmental context. Copyright © 2012 John Wiley & Sons, Ltd
Collaborative Nonverbal Interaction within Virtual Environments
Abstract:Current virtual environments are predominantly visual-spatial, which allows their ‘inhabitants’ the display, either in a conscious or unconscious way, of nonverbal cues during interaction, such as gaze direction, deictic gestures or location. This interchange of nonverbal messages enriches interaction while supports mutual comprehension, fundamental for collaborative work and therefore particularly important in a multiuser virtual environment, that is, a Collaborative Virtual Environment. Different techniques, the media involvement, and automatic detection related collaborative nonverbal interaction are here discussed.Keywords: Collaborative Virtual Environment, nonverbal communication, collaborative interactio
Tailoring Interaction. Sensing Social Signals with Textiles.
Nonverbal behaviour is an important part of conversation and can reveal much about the nature of an interaction. It includes phenomena ranging from large-scale posture shifts to small scale nods. Capturing these often spontaneous phenomena requires unobtrusive sensing techniques that do not interfere with the interaction. We propose an underexploited sensing modality for sensing nonverbal behaviours: textiles. As a material in close contact with the body, they provide ubiquitous, large surfaces that make them a suitable soft interface. Although the literature on nonverbal communication focuses on upper body movements such as gestures, observations of multi-party, seated conversations suggest that sitting postures, leg and foot movements are also systematically related to patterns of social interaction. This thesis addressees the following questions: Can the textiles surrounding us measure social engagement? Can they tell who is speaking, and who, if anyone, is listening? Furthermore, how should wearable textile sensing systems be designed and what behavioural signals could textiles reveal? To address these questions, we have designed and manufactured bespoke chairs and trousers with integrated textile pressure sensors, that are introduced here. The designs are evaluated in three user studies that produce multi-modal datasets for the exploration of fine-grained interactional signals. Two approaches to using these bespoke textile sensors are explored. First, hand crafted sensor patches in chair covers serve to distinguish speakers and listeners. Second, a pressure sensitive matrix in custom-made smart trousers is developed to detect static sitting postures, dynamic bodily movement, as well as basic conversational states. Statistical analyses, machine learning approaches, and ethnographic methods show that by moni- toring patterns of pressure change alone it is possible to not only classify postures with high accuracy, but also to identify a wide range of behaviours reliably in individuals and groups. These findings es- tablish textiles as a novel, wearable sensing system for applications in social sciences, and contribute towards a better understanding of nonverbal communication, especially the significance of posture shifts when seated. If chairs know who is speaking, if our trousers can capture our social engagement, what role can smart textiles have in the future of human interaction? How can we build new ways to map social ecologies and tailor interactions
- …