667 research outputs found
Mechatronic design of the Twente humanoid head
This paper describes the mechatronic design of the Twente humanoid head, which has been realized in the purpose of having a research platform for human-machine interaction. The design features a fast, four degree of freedom neck, with long range of motion, and a vision system with three degrees of freedom, mimicking the eyes. To achieve fast target tracking, two degrees of freedom in the neck are combined in a differential drive, resulting in a low moving mass and the possibility to use powerful actuators. The performance of the neck has been optimized by minimizing backlash in the mechanisms, and using gravity compensation. The vision system is based on a saliency algorithm that uses the camera images to determine where the humanoid head should look at, i.e. the focus of attention computed according to biological studies. The motion control algorithm receives, as input, the output of the vision algorithm and controls the humanoid head to focus on and follow the target point. The control architecture exploits the redundancy of the system to show human-like motions while looking at a target. The head has a translucent plastic cover, onto which an internal LED system projects the mouth and the eyebrows, realizing human-like facial expressions
Design of a Realistic Robotic Head based on Action Coding System
Producción CientíficaIn this paper, the development of a robotic head able to move
and show di erent emotions is addressed. The movement and emotion
generation system has been designed following the human facial muscu-
lature. Starting from the Facial Action Coding System (FACS), we have
built a 26 actions units model that is able to produce the most relevant
movements and emotions of a real human head. The whole work has
been carried out in two steps. In the rst step, a mechanical skeleton
has been designed and built, in which the di erent actuators have been
inserted. In the second step, a two-layered silicon skin has been manu-
factured, on which the di erent actuators have been inserted following
the real muscle-insertions, for performing the di erent movements and
gestures. The developed head has been integrated in a high level be-
havioural architecture, and pilot experiments with 10 users regarding
emotion recognition and mimicking have been carried out.Junta de Castilla y León (Programa de apoyo a proyectos de investigación-Ref. VA036U14)Junta de Castilla y León (programa de apoyo a proyectos de investigación - Ref. VA013A12-2)Ministerio de Economía, Industria y Competitividad (Grant DPI2014-56500-R
Multimodal Man-machine Interface and Virtual Reality for Assistive Medical Systems
The results of research the intelligence multimodal man-machine interface and virtual reality means for
assistive medical systems including computers and mechatronic systems (robots) are discussed. The gesture
translation for disability peoples, the learning-by-showing technology and virtual operating room with 3D
visualization are presented in this report and were announced at International exhibition "Intelligent and Adaptive
Robots–2005"
Developing an Affect-Aware Rear-Projected Robotic Agent
Social (or Sociable) robots are designed to interact with people in a natural and interpersonal manner. They are becoming an integrated part of our daily lives and have achieved positive outcomes in several applications such as education, health care, quality of life, entertainment, etc. Despite significant progress towards the development of realistic social robotic agents, a number of problems remain to be solved. First, current social robots either lack enough ability to have deep social interaction with human, or they are very expensive to build and maintain. Second, current social robots have yet to reach the full emotional and social capabilities necessary for rich and robust interaction with human beings. To address these problems, this dissertation presents the development of a low-cost, flexible, affect-aware rear-projected robotic agent (called ExpressionBot), that is designed to support verbal and non-verbal communication between the robot and humans, with the goal of closely modeling the dynamics of natural face-to-face communication.
The developed robotic platform uses state-of-the-art character animation technologies to create an animated human face (aka avatar) that is capable of showing facial expressions, realistic eye movement, and accurate visual speech, and then project this avatar onto a face-shaped translucent mask. The mask and the projector are then rigged onto a neck mechanism that can move like a human head. Since an animation is projected onto a mask, the robotic face is highly flexible research tool, mechanically simple, and low-cost to design, build and maintain compared with mechatronic and android faces. The results of our comprehensive Human-Robot Interaction (HRI) studies illustrate the benefits and values of the proposed rear-projected robotic platform over a virtual-agent with the same animation displayed on a 2D computer screen. The results indicate that ExpressionBot is well accepted by users, with some advantages in expressing facial expressions more accurately and perceiving mutual eye gaze contact.
To improve social capabilities of the robot and create an expressive and empathic social agent (affect-aware) which is capable of interpreting users\u27 emotional facial expressions, we developed a new Deep Neural Networks (DNN) architecture for Facial Expression Recognition (FER). The proposed DNN was initially trained on seven well-known publicly available databases, and obtained significantly better than, or comparable to, traditional convolutional neural networks or other state-of-the-art methods in both accuracy and learning time. Since the performance of the automated FER system highly depends on its training data, and the eventual goal of the proposed robotic platform is to interact with users in an uncontrolled environment, a database of facial expressions in the wild (called AffectNet) was created by querying emotion-related keywords from different search engines. AffectNet contains more than 1M images with faces and 440,000 manually annotated images with facial expressions, valence, and arousal. Two DNNs were trained on AffectNet to classify the facial expression images and predict the value of valence and arousal. Various evaluation metrics show that our deep neural network approaches trained on AffectNet can perform better than conventional machine learning methods and available off-the-shelf FER systems.
We then integrated this automated FER system into spoken dialog of our robotic platform to extend and enrich the capabilities of ExpressionBot beyond spoken dialog and create an affect-aware robotic agent that can measure and infer users\u27 affect and cognition. Three social/interaction aspects (task engagement, being empathic, and likability of the robot) are measured in an experiment with the affect-aware robotic agent. The results indicate that users rated our affect-aware agent as empathic and likable as a robot in which user\u27s affect is recognized by a human (WoZ).
In summary, this dissertation presents the development and HRI studies of a perceptive, and expressive, conversational, rear-projected, life-like robotic agent (aka ExpressionBot or Ryan) that models natural face-to-face communication between human and emapthic agent. The results of our in-depth human-robot-interaction studies show that this robotic agent can serve as a model for creating the next generation of empathic social robots
Design of a Virtual Assistant to Improve Interaction Between the Audience and the Presenter
This article presents a novel design of a Virtual Assistant as part of a human-machine interaction system to improve communication between the presenter and the audience that can be used in education or general presentations for improving interaction during the presentations (e.g., auditoriums with 200 people). The main goal of the proposed model is the design of a framework of interaction to increase the level of attention of the public in key aspects of the presentation. In this manner, the collaboration between the presenter and Virtual Assistant could improve the level of learning among the public. The design of the Virtual Assistant relies on non-anthropomorphic forms with ‘live’ characteristics generating an intuitive and self-explainable interface. A set of intuitive and useful virtual interactions to support the presenter was designed. This design was validated from various types of the public with a psychological study based on a discrete emotions’ questionnaire confirming the adequacy of the proposed solution. The human-machine interaction system supporting the Virtual Assistant should automatically recognize the attention level of the audience from audiovisual resources and synchronize the Virtual Assistant with the presentation. The system involves a complex artificial intelligence architecture embracing perception of high-level features from audio and video, knowledge representation, and reasoning for pervasive and affective computing and reinforcement learning to teach the intelligent agent to decide on the best strategy to increase the level of attention of the audience
A Retro-Projected Robotic Head for Social Human-Robot Interaction
As people respond strongly to faces and facial features, both con-
sciously and subconsciously, faces are an essential aspect of social
robots. Robotic faces and heads until recently belonged to one of the
following categories: virtual, mechatronic or animatronic. As an orig-
inal contribution to the field of human-robot interaction, I present the
R-PAF technology (Retro-Projected Animated Faces): a novel robotic
head displaying a real-time, computer-rendered face, retro-projected
from within the head volume onto a mask, as well as its driving soft-
ware designed with openness and portability to other hybrid robotic
platforms in mind.
The work constitutes the first implementation of a non-planar mask
suitable for social human-robot interaction, comprising key elements
of social interaction such as precise gaze direction control, facial ex-
pressions and blushing, and the first demonstration of an interactive
video-animated facial mask mounted on a 5-axis robotic arm. The
LightHead robot, a R-PAF demonstrator and experimental platform,
has demonstrated robustness both in extended controlled and uncon-
trolled settings. The iterative hardware and facial design, details of the
three-layered software architecture and tools, the implementation of
life-like facial behaviours, as well as improvements in social-emotional
robotic communication are reported. Furthermore, a series of evalua-
tions present the first study on human performance in reading robotic
gaze and another first on user’s ethnic preference towards a robot face
Expressivity in Natural and Artificial Systems
Roboticists are trying to replicate animal behavior in artificial systems.
Yet, quantitative bounds on capacity of a moving platform (natural or
artificial) to express information in the environment are not known. This paper
presents a measure for the capacity of motion complexity -- the expressivity --
of articulated platforms (both natural and artificial) and shows that this
measure is stagnant and unexpectedly limited in extant robotic systems. This
analysis indicates trends in increasing capacity in both internal and external
complexity for natural systems while artificial, robotic systems have increased
significantly in the capacity of computational (internal) states but remained
more or less constant in mechanical (external) state capacity. This work
presents a way to analyze trends in animal behavior and shows that robots are
not capable of the same multi-faceted behavior in rich, dynamic environments as
natural systems.Comment: Rejected from Nature, after review and appeal, July 4, 2018
(submitted May 11, 2018
- …