2,369 research outputs found
Determining what people feel and think when interacting with humans and machines
Any interactive software program must interpret the users’ actions and come up with an appropriate response that is intelligable and meaningful to the user. In most situations, the options of the user are determined by the software and hardware and the actions that can be carried out are unambiguous. The machine knows what it should do when the user carries out an action. In most cases, the user knows what he has to do by relying on conventions which he may have learned by having had a look at the instruction manual, having them seen performed by somebody else, or which he learned by modifying a previously learned convention. Some, or most, of the times he just finds out by trial and error. In user-friendly interfaces, the user knows, without having to read extensive manuals, what is expected from him and how he can get the machine to do what he wants. An intelligent interface is so-called, because it does not assume the same kind of programming of the user by the machine, but the machine itself can figure out what the user wants and how he wants it without the user having to take all the trouble of telling it to the machine in the way the machine dictates but being able to do it in his own words. Or perhaps by not using any words at all, as the machine is able to read off the intentions of the user by observing his actions and expressions. Ideally, the machine should be able to determine what the user wants, what he expects, what he hopes will happen, and how he feels
Robust Modeling of Epistemic Mental States
This work identifies and advances some research challenges in the analysis of
facial features and their temporal dynamics with epistemic mental states in
dyadic conversations. Epistemic states are: Agreement, Concentration,
Thoughtful, Certain, and Interest. In this paper, we perform a number of
statistical analyses and simulations to identify the relationship between
facial features and epistemic states. Non-linear relations are found to be more
prevalent, while temporal features derived from original facial features have
demonstrated a strong correlation with intensity changes. Then, we propose a
novel prediction framework that takes facial features and their nonlinear
relation scores as input and predict different epistemic states in videos. The
prediction of epistemic states is boosted when the classification of emotion
changing regions such as rising, falling, or steady-state are incorporated with
the temporal features. The proposed predictive models can predict the epistemic
states with significantly improved accuracy: correlation coefficient (CoERR)
for Agreement is 0.827, for Concentration 0.901, for Thoughtful 0.794, for
Certain 0.854, and for Interest 0.913.Comment: Accepted for Publication in Multimedia Tools and Application, Special
Issue: Socio-Affective Technologie
Knowledge Elicitation Methods for Affect Modelling in Education
Research on the relationship between affect and cognition in Artificial Intelligence in Education (AIEd) brings an important dimension to our understanding of how learning occurs and how it can be facilitated. Emotions are crucial to learning, but their nature, the conditions under which they occur, and their exact impact on learning for different learners in diverse contexts still needs to be mapped out. The study of affect during learning can be challenging, because emotions are subjective, fleeting phenomena that are often difficult for learners to report accurately and for observers to perceive reliably. Context forms an integral part of learners’ affect and the study thereof. This review provides a synthesis of the current knowledge elicitation methods that are used to aid the study of learners’ affect and to inform the design of intelligent technologies for learning. Advantages and disadvantages of the specific methods are discussed along with their respective potential for enhancing research in this area, and issues related to the interpretation of data that emerges as the result of their use. References to related research are also provided together with illustrative examples of where the individual methods have been used in the past. Therefore, this review is intended as a resource for methodological decision making for those who want to study emotions and their antecedents in AIEd contexts, i.e. where the aim is to inform the design and implementation of an intelligent learning environment or to evaluate its use and educational efficacy
The Cognitive-Affective-Social Theory of Learning in digital Environments (CASTLE)
For a long time, research on individuals learning in digital environments was primarily based on cognitive-oriented theories. This paper aims at providing evidence that social processes affect individual learning with digital materials. Based on these theories and empirical results, a social-processes-augmented theory is suggested: the Cognitive-Affective-Social Theory of Learning in digital Environments (CASTLE). This CASTLE postulates that social cues in digital materials activate social schemata in learners leading to enhanced (para-)social, motivational, emotional, and metacognitive processes. To substantiate this theory, socio-cognitive theories are used, which predict social influences on learning with digital materials. Besides, previous empirical findings are presented assuming that with a rising number of social cues in digital materials, the influence of social processes increases. Finally, consequences regarding the design of digital learning media are discussed
A Novel Multimodal Approach for Studying the Dynamics of Curiosity in Small Group Learning
Curiosity is a vital metacognitive skill in educational contexts, leading to
creativity, and a love of learning. And while many school systems increasingly
undercut curiosity by teaching to the test, teachers are increasingly
interested in how to evoke curiosity in their students to prepare them for a
world in which lifelong learning and reskilling will be more and more
important. One aspect of curiosity that has received little attention, however,
is the role of peers in eliciting curiosity. We present what we believe to be
the first theoretical framework that articulates an integrated socio-cognitive
account of curiosity that ties observable behaviors in peers to underlying
curiosity states. We make a bipartite distinction between individual and
interpersonal functions that contribute to curiosity, and multimodal behaviors
that fulfill these functions. We validate the proposed framework by leveraging
a longitudinal latent variable modeling approach. Findings confirm a positive
predictive relationship between the latent variables of individual and
interpersonal functions and curiosity, with the interpersonal functions
exercising a comparatively stronger influence. Prominent behavioral
realizations of these functions are also discovered in a data-driven manner. We
instantiate the proposed theoretical framework in a set of strategies and
tactics that can be incorporated into learning technologies to indicate, evoke,
and scaffold curiosity. This work is a step towards designing learning
technologies that can recognize and evoke moment-by-moment curiosity during
learning in social contexts and towards a more complete multimodal learning
analytics. The underlying rationale is applicable more generally for developing
computer support for other metacognitive and socio-emotional skills.Comment: arXiv admin note: text overlap with arXiv:1704.0748
- …