3,345 research outputs found
Affective games:a multimodal classification system
Affective gaming is a relatively new field of research that exploits human emotions to influence gameplay for an enhanced player experience. Changes in player’s psychology reflect on their behaviour and physiology, hence recognition of such variation is a core element in affective games. Complementary sources of affect offer more reliable recognition, especially in contexts where one modality is partial or unavailable. As a multimodal recognition system, affect-aware games are subject to the practical difficulties met by traditional trained classifiers. In addition, inherited game-related challenges in terms of data collection and performance arise while attempting to sustain an acceptable level of immersion. Most existing scenarios employ sensors that offer limited freedom of movement resulting in less realistic experiences. Recent advances now offer technology that allows players to communicate more freely and naturally with the game, and furthermore, control it without the use of input devices. However, the affective game industry is still in its infancy and definitely needs to catch up with the current life-like level of adaptation provided by graphics and animation
Recommended from our members
Occupant-Facade interaction: a review and classification scheme
The interest in occupant interaction with building controls and automation systems is growing due to the wider availability of embedded sensing devices and automated or intelligent building components that can integrate building control strategies with occupant-centred data and lead to greater occupant satisfaction and reduction in energy consumption. An area of particular interest is the interaction strategies between occupants and the so called automated facades, such as dynamic shading devices and switchable glazing. Occupant-Facade interactions are often disruptive and source of dissatisfaction because of conflicts between competing requirements, e.g. energy-efficiency and indoor environmental quality. To solve these conflicts, expertise from several disciplines is required, including Behavioural Science and Building Physics, but the absence of common research frameworks impedes knowledge transfer between different fields of expertise. This paper reviews existing multi-disciplinary research on occupant interaction with facades, buildings and automation systems and provides a new classification scheme of Occupant-Facade interaction. The scheme is based on an extensive review of interactive scenarios between occupants and facades that are summarised in this paper. The classification scheme was found to be successful in: 1) capturing the multidisciplinary nature of interactive scenarios by clarifying relationships between components; 2) identifying similarities and characteristics among interactive scenarios; 3) understanding research gaps. The classification scheme proposed in this paper has the potential to be a useful tool for the multi-disciplinary research community in this field. The review also showed that more research is needed to characterise the holistic and multi-disciplinary effect of occupant interaction with intelligent building components.EPSRC Doctoral Training Accoun
Edge-centric Optimization of Multi-modal ML-driven eHealth Applications
Smart eHealth applications deliver personalized and preventive digital
healthcare services to clients through remote sensing, continuous monitoring,
and data analytics. Smart eHealth applications sense input data from multiple
modalities, transmit the data to edge and/or cloud nodes, and process the data
with compute intensive machine learning (ML) algorithms. Run-time variations
with continuous stream of noisy input data, unreliable network connection,
computational requirements of ML algorithms, and choice of compute placement
among sensor-edge-cloud layers affect the efficiency of ML-driven eHealth
applications. In this chapter, we present edge-centric techniques for optimized
compute placement, exploration of accuracy-performance trade-offs, and
cross-layered sense-compute co-optimization for ML-driven eHealth applications.
We demonstrate the practical use cases of smart eHealth applications in
everyday settings, through a sensor-edge-cloud framework for an objective pain
assessment case study
An Actor-Centric Approach to Facial Animation Control by Neural Networks For Non-Player Characters in Video Games
Game developers increasingly consider the degree to which character animation emulates facial expressions found in cinema. Employing animators and actors to produce cinematic facial animation by mixing motion capture and hand-crafted animation is labor intensive and therefore expensive. Emotion corpora and neural network controllers have shown promise toward developing autonomous animation that does not rely on motion capture. Previous research and practice in disciplines of Computer Science, Psychology and the Performing Arts have provided frameworks on which to build a workflow toward creating an emotion AI system that can animate the facial mesh of a 3d non-player character deploying a combination of related theories and methods. However, past investigations and their resulting production methods largely ignore the emotion generation systems that have evolved in the performing arts for more than a century. We find very little research that embraces the intellectual process of trained actors as complex collaborators from which to understand and model the training of a neural network for character animation. This investigation demonstrates a workflow design that integrates knowledge from the performing arts and the affective branches of the social and biological sciences. Our workflow begins at the stage of developing and annotating a fictional scenario with actors, to producing a video emotion corpus, to designing training and validating a neural network, to analyzing the emotion data annotation of the corpus and neural network, and finally to determining resemblant behavior of its autonomous animation control of a 3d character facial mesh. The resulting workflow includes a method for the development of a neural network architecture whose initial efficacy as a facial emotion expression simulator has been tested and validated as substantially resemblant to the character behavior developed by a human actor
Affect state recognition for adaptive human robot interaction in learning environments
Previous studies of robots used in learning environments suggest that the interaction between learner and robot is able to enhance the learning procedure towards a better engagement of the learner. Moreover, intelligent robots can also adapt their behavior during a learning process according to certain criteria resulting in increasing cognitive learning gains. Motivated by these results, we propose a novel Human Robot Interaction framework where the robot adjusts its behavior to the affect state of the learner. Our framework uses the theory of flow to label different affect states (i.e., engagement, boredom and frustration) and adapt the robot's actions. Based on the automatic recognition of these states, through visual cues, our method adapt the learning actions taking place at this moment and performed by the robot. This results in keeping the learner at most times engaged in the learning process. In order to recognizing the affect state of the user a two step approach is followed. Initially we recognize the facial expressions of the learner and therefore we map these to an affect state. Our algorithm perform well even in situations where the environment is noisy due to the presence of more than one person and/or situations where the face is partially occluded
I feel you: the design and evaluation of a domotic affect-sensitive spoken conversational agent
We describe the work on infusion of emotion into a limited-task autonomous spoken conversational agent situated in the domestic environment, using a need-inspired task-independent emotion model (NEMO). In order to demonstrate the generation of affect through the use of the model, we describe the work of integrating it with a natural-language mixed-initiative HiFi-control spoken conversational agent (SCA). NEMO and the host system communicate externally, removing the need for the Dialog Manager to be modified, as is done in most existing dialog systems, in order to be adaptive. The first part of the paper concerns the integration between NEMO and the host agent. The second part summarizes the work on automatic affect prediction, namely, frustration and contentment, from dialog features, a non-conventional source, in the attempt of moving towards a more user-centric approach. The final part reports the evaluation results obtained from a user study, in which both versions of the agent (non-adaptive and emotionally-adaptive) were compared. The results provide substantial evidences with respect to the benefits of adding emotion in a spoken conversational agent, especially in mitigating users' frustrations and, ultimately, improving their satisfaction
Face2Multi-modal: in-vehicle multi-modal predictors via facial expressions
Towards intelligent Human-Vehicle Interaction systems and innovative Human-Vehicle Interaction designs, in-vehicle drivers' physiological data has been explored as an essential data source. However, equipping multiple biosensors is considered the limited extent of user-friendliness and impractical during the driving procedure. The lack of a proper approach to access physiological data has hindered wider applications of advanced biosignal-driven designs in practice (e.g. monitoring systems and etc.). Hence, the demand for a user-friendly approach to measuring drivers' body statuses has become more intense. In this Work-In-Progress, we present Face2Multi-modal, an In-vehicle multi-modal Data Streams Predictors through facial expressions only. More specifically, we have explored the estimations of Heart Rate, Skin Conductance, and Vehicle Speed of the drivers. We believe Face2Multi-modal provides a user-friendly alternative to acquiring drivers' physiological status and vehicle status, which could serve as the building block for many current or future personalized Human-Vehicle Interaction designs. More details and updates about the project Face2Multi-modal is online at https://github.com/unnc-ucc/Face2Multimodal/
Deliberative Context-Aware Ambient Intelligence System for Assisted Living Homes
Monitoring wellbeing and stress is one of the problems covered by ambient
intelligence, as stress is a significant cause of human illnesses directly
affecting our emotional state. The primary aim was to propose a deliberation
architecture for an ambient intelligence healthcare application. The
architecture provides a plan for comforting stressed seniors suffering from
negative emotions in an assisted living home and executes the plan considering
the environment's dynamic nature. Literature was reviewed to identify the
convergence between deliberation and ambient intelligence and the latter's
latest healthcare trends. A deliberation function was designed to achieve
context-aware dynamic human-robot interaction, perception, planning
capabilities, reactivity, and context-awareness with regard to the environment.
A number of experimental case studies in a simulated assisted living home
scenario were conducted to demonstrate the approach's behavior and validity.
The proposed methods were validated to show classification accuracy. The
validation showed that the deliberation function has effectively achieved its
deliberative objectives
- …