6,191 research outputs found
A Virtual Conversational Agent for Teens with Autism: Experimental Results and Design Lessons
We present the design of an online social skills development interface for
teenagers with autism spectrum disorder (ASD). The interface is intended to
enable private conversation practice anywhere, anytime using a web-browser.
Users converse informally with a virtual agent, receiving feedback on nonverbal
cues in real-time, and summary feedback. The prototype was developed in
consultation with an expert UX designer, two psychologists, and a pediatrician.
Using the data from 47 individuals, feedback and dialogue generation were
automated using a hidden Markov model and a schema-driven dialogue manager
capable of handling multi-topic conversations. We conducted a study with nine
high-functioning ASD teenagers. Through a thematic analysis of post-experiment
interviews, identified several key design considerations, notably: 1) Users
should be fully briefed at the outset about the purpose and limitations of the
system, to avoid unrealistic expectations. 2) An interface should incorporate
positive acknowledgment of behavior change. 3) Realistic appearance of a
virtual agent and responsiveness are important in engaging users. 4)
Conversation personalization, for instance in prompting laconic users for more
input and reciprocal questions, would help the teenagers engage for longer
terms and increase the system's utility
Recommended from our members
Real-time decoding of question-and-answer speech dialogue using human cortical activity.
Natural communication often occurs in dialogue, differentially engaging auditory and sensorimotor brain regions during listening and speaking. However, previous attempts to decode speech directly from the human brain typically consider listening or speaking tasks in isolation. Here, human participants listened to questions and responded aloud with answers while we used high-density electrocorticography (ECoG) recordings to detect when they heard or said an utterance and to then decode the utterance's identity. Because certain answers were only plausible responses to certain questions, we could dynamically update the prior probabilities of each answer using the decoded question likelihoods as context. We decode produced and perceived utterances with accuracy rates as high as 61% and 76%, respectively (chance is 7% and 20%). Contextual integration of decoded question likelihoods significantly improves answer decoding. These results demonstrate real-time decoding of speech in an interactive, conversational setting, which has important implications for patients who are unable to communicate
- …