3,641 research outputs found
Artificial Empathy in Marketing Interactions: Bridging the Human-AI Gap in Affective and Social Customer Experience
Artificial intelligence (AI) continues to transform firm-customer interactions. However, current AI marketing agents are often perceived as cold and uncaring and can be poor substitutes for human-based interactions. Addressing this issue, this article argues that artificial empathy needs to become an important design consideration in the next generation of AI marketing applications. Drawing from research in diverse disciplines, we develop a systematic framework for integrating artificial empathy into AI-enabled marketing interactions. We elaborate on the key components of artificial empathy and how each component can be implemented in AI marketing agents. We further explicate and test how artificial empathy generates value for both customers and firms by bridging the AI-human gap in affective and social customer experience. Recognizing that artificial empathy may not always be desirable or relevant, we identify the requirements for artificial empathy to create value and deduce situations where it is unnecessary and, in some cases, harmful
Bringing Human Robot Interaction towards _Trust and Social Engineering
Robots started their journey in books and movies; nowadays, they are becoming an
important part of our daily lives: from industrial robots, passing through entertainment
robots, and reaching social robotics in fields like healthcare or education.
An important aspect of social robotics is the human counterpart, therefore, there is
an interaction between the humans and robots. Interactions among humans are often
taken for granted as, since children, we learn how to interact with each other. In robotics,
this interaction is still very immature, however, critical for a successful incorporation of
robots in society. Human robot interaction (HRI) is the domain that works on improving
these interactions.
HRI encloses many aspects, and a significant one is trust. Trust is the assumption that
somebody or something is good and reliable; and it is critical for a developed society.
Therefore, in a society where robots can part, the trust they could generate will be essential
for cohabitation.
A downside of trust is overtrusting an entity; in other words, an insufficient alignment
of the projected trust and the expectations of a morally correct behaviour. This effect
could negatively influence and damage the interactions between agents. In the case of
humans, it is usually exploited by scammers, conmen or social engineers - who take
advantage of the people's overtrust in order to manipulate them into performing actions
that may not be beneficial for the victims.
This thesis tries to shed light on the development of trust towards robots, how this
trust could become overtrust and be exploited by social engineering techniques. More
precisely, the following experiments have been carried out: (i) Treasure Hunt, in which
the robot followed a social engineering framework where it gathered personal
information from the participants, improved the trust and rapport with them, and at the
end, it exploited that trust manipulating participants into performing a risky action.
(ii) Wicked Professor, in which a very human-like robot tried to enforce its authority to
make participants obey socially inappropriate requests. Most of the participants realized
that the requests were morally wrong, but eventually, they succumbed to the robot'sauthority while holding the robot as morally responsible. (iii) Detective iCub, in which it
was evaluated whether the robot could be endowed with the ability to detect when the
human partner was lying. Deception detection is an essential skill for social engineers and
professionals in the domain of education, healthcare and security. The robot achieved
75% of accuracy in the lie detection. There were also found slight differences in the
behaviour exhibited by the participants when interacting with a human or a robot
interrogator.
Lastly, this thesis approaches the topic of privacy - a fundamental human value. With
the integration of robotics and technology in our society, privacy will be affected in ways
we are not used. Robots have sensors able to record and gather all kind of data, and it is
possible that this information is transmitted via internet without the knowledge of the
user. This is an important aspect to consider since a violation in privacy can heavily
impact the trust.
Summarizing, this thesis shows that robots are able to establish and improve trust
during an interaction, to take advantage of overtrust and to misuse it by applying different
types of social engineering techniques, such as manipulation and authority. Moreover,
robots can be enabled to pick up different human cues to detect deception, which can
help both, social engineers and professionals in the human sector. Nevertheless, it is of
the utmost importance to make roboticists, programmers, entrepreneurs, lawyers,
psychologists, and other sectors involved, aware that social robots can be highly beneficial
for humans, but they could also be exploited for malicious purposes
Producing Acoustic-Prosodic Entrainment in a Robotic Learning Companion to Build Learner Rapport
abstract: With advances in automatic speech recognition, spoken dialogue systems are assuming increasingly social roles. There is a growing need for these systems to be socially responsive, capable of building rapport with users. In human-human interactions, rapport is critical to patient-doctor communication, conflict resolution, educational interactions, and social engagement. Rapport between people promotes successful collaboration, motivation, and task success. Dialogue systems which can build rapport with their user may produce similar effects, personalizing interactions to create better outcomes.
This dissertation focuses on how dialogue systems can build rapport utilizing acoustic-prosodic entrainment. Acoustic-prosodic entrainment occurs when individuals adapt their acoustic-prosodic features of speech, such as tone of voice or loudness, to one another over the course of a conversation. Correlated with liking and task success, a dialogue system which entrains may enhance rapport. Entrainment, however, is very challenging to model. People entrain on different features in many ways and how to design entrainment to build rapport is unclear. The first goal of this dissertation is to explore how acoustic-prosodic entrainment can be modeled to build rapport.
Towards this goal, this work presents a series of studies comparing, evaluating, and iterating on the design of entrainment, motivated and informed by human-human dialogue. These models of entrainment are implemented in the dialogue system of a robotic learning companion. Learning companions are educational agents that engage students socially to increase motivation and facilitate learning. As a learning companion’s ability to be socially responsive increases, so do vital learning outcomes. A second goal of this dissertation is to explore the effects of entrainment on concrete outcomes such as learning in interactions with robotic learning companions.
This dissertation results in contributions both technical and theoretical. Technical contributions include a robust and modular dialogue system capable of producing prosodic entrainment and other socially-responsive behavior. One of the first systems of its kind, the results demonstrate that an entraining, social learning companion can positively build rapport and increase learning. This dissertation provides support for exploring phenomena like entrainment to enhance factors such as rapport and learning and provides a platform with which to explore these phenomena in future work.Dissertation/ThesisDoctoral Dissertation Computer Science 201
Choosing a story for measuring language development in Cantonese-speaking children
Also available in print.Thesis (B.Sc)--University of Hong Kong, 2006."A dissertation submitted in partial fulfilment of the requirements for the Bachelor of Science (Speech and Hearing Sciences), The University of Hong Kong, June 30, 2006."published_or_final_versionSpeech and Hearing SciencesBachelorBachelor of Science in Speech and Hearing Science
Perceiving Sociable Technology: Exploring the Role of Anthropomorphism and Agency Perception on Human-Computer Interaction (HCI)
With the arrival of personal assistants and other AI-enabled autonomous technologies, social interactions with smart devices have become a part of our daily lives. Therefore, it becomes increasingly important to understand how these social interactions emerge, and why users appear to be influenced by them. For this reason, I explore questions on what the antecedents and consequences of this phenomenon, known as anthropomorphism, are as described in the extant literature from fields ranging from information systems to social neuroscience. I critically analyze those empirical studies directly measuring anthropomorphism and those referring to it without a corresponding measurement. Through a grounded theory approach, I identify common themes and use them to develop models for the antecedents and consequences of anthropomorphism. The results suggest anthropomorphism possesses both conscious and non-conscious components with varying implications. While conscious attributions are shown to vary based on individual differences, non-conscious attributions emerge whenever a technology exhibits apparent reasoning such as through non-verbal behavior like peer-to-peer mirroring or verbal paralinguistic and backchanneling cues. Anthropomorphism has been shown to affect users’ self-perceptions, perceptions of the technology, how users interact with the technology, and the users’ performance. Examples include changes in a users’ trust on the technology, conformity effects, bonding, and displays of empathy. I argue these effects emerge from changes in users’ perceived agency, and their self- and social- identity similarly to interactions between humans. Afterwards, I critically examine current theories on anthropomorphism and present propositions about its nature based on the results of the empirical literature. Subsequently, I introduce a two-factor model of anthropomorphism that proposes how an individual anthropomorphizes a technology is dependent on how the technology was initially perceived (top-down and rational or bottom-up and automatic), and whether it exhibits a capacity for agency or experience. I propose that where a technology lays along this spectrum determines how individuals relates to it, creating shared agency effects, or changing the users’ social identity. For this reason, anthropomorphism is a powerful tool that can be leveraged to support future interactions with smart technologies
Robotic motion learning framework to promote social engagement
Abstract Imitation is a powerful component of communication between people, and it poses an important implication in improving the quality of interaction in the field of human–robot interaction (HRI). This paper discusses a novel framework designed to improve human–robot interaction through robotic imitation of a participant’s gestures. In our experiment, a humanoid robotic agent socializes with and plays games with a participant. For the experimental group, the robot additionally imitates one of the participant’s novel gestures during a play session. We hypothesize that the robot’s use of imitation will increase the participant’s openness towards engaging with the robot. Experimental results from a user study of 12 subjects show that post-imitation, experimental subjects displayed a more positive emotional state, had higher instances of mood contagion towards the robot, and interpreted the robot to have a higher level of autonomy than their control group counterparts did. These results point to an increased participant interest in engagement fueled by personalized imitation during interaction
Automatic Context-Driven Inference of Engagement in HMI: A Survey
An integral part of seamless human-human communication is engagement, the
process by which two or more participants establish, maintain, and end their
perceived connection. Therefore, to develop successful human-centered
human-machine interaction applications, automatic engagement inference is one
of the tasks required to achieve engaging interactions between humans and
machines, and to make machines attuned to their users, hence enhancing user
satisfaction and technology acceptance. Several factors contribute to
engagement state inference, which include the interaction context and
interactants' behaviours and identity. Indeed, engagement is a multi-faceted
and multi-modal construct that requires high accuracy in the analysis and
interpretation of contextual, verbal and non-verbal cues. Thus, the development
of an automated and intelligent system that accomplishes this task has been
proven to be challenging so far. This paper presents a comprehensive survey on
previous work in engagement inference for human-machine interaction, entailing
interdisciplinary definition, engagement components and factors, publicly
available datasets, ground truth assessment, and most commonly used features
and methods, serving as a guide for the development of future human-machine
interaction interfaces with reliable context-aware engagement inference
capability. An in-depth review across embodied and disembodied interaction
modes, and an emphasis on the interaction context of which engagement
perception modules are integrated sets apart the presented survey from existing
surveys
- …