228,550 research outputs found

    Reasoning about Emotional Agents

    Get PDF
    In this paper we are concerned with reasoning about agents with emotions. To be more precise: we aim at a logical account of emotional agents. The very topic may already raise some eyebrows. Reasoning / rationality and emotions seem opposites, and reasoning about emotions or a logic of emotional agents seems a contradiction in terms. However, emotions and rationality are known to be more interconnected than one may suspect. There is psychological evidence that having emotions may help one to do reasoning and tasks for which rationality seems to be the only factor [1]. Moreover, work by e.g. Sloman [5] shows that one may think of designing agentbased systems where these agents show some kind of emotions, and, even more importantly, display behaviour dependent on their emotional state. It is exactly in this sense that we aim at looking at emotional agents: artificial systems that are designed in such a manner that emotions play a role. Also in psychology emotions are viewed as a structuring mechanism. Emotions are held to help human beings to choose from a myriad of possible actions in response to what happens in ou

    Emotional agents at the square lattice

    Full text link
    We introduce and investigate by numerical simulations a number of models of emotional agents at the square lattice. Our models describe the most general features of emotions such as the spontaneous emotional arousal, emotional relaxation, and transfers of emotions between different agents. Group emotions in the considered models are periodically fluctuating between two opposite valency levels and as result the mean value of such group emotions is zero. The oscillations amplitude depends strongly on probability ps of the individual spontaneous arousal. For small values of relaxation times tau we observed a stochastic resonance, i.e. the signal to noise ratio SNR is maximal for a non-zero ps parameter. The amplitude increases with the probability p of local affective interactions while the mean oscillations period increases with the relaxation time tau and is only weakly dependent on other system parameters. Presence of emotional antenna can enhance positive or negative emotions and for the optimal transition probability the antenna can change agents emotions at longer distances. The stochastic resonance was also observed for the influence of emotions on task execution efficiency.Comment: 28 pages, 19 figures, 3 table

    An Agent-Based Model of Collective Emotions in Online Communities

    Full text link
    We develop a agent-based framework to model the emergence of collective emotions, which is applied to online communities. Agents individual emotions are described by their valence and arousal. Using the concept of Brownian agents, these variables change according to a stochastic dynamics, which also considers the feedback from online communication. Agents generate emotional information, which is stored and distributed in a field modeling the online medium. This field affects the emotional states of agents in a non-linear manner. We derive conditions for the emergence of collective emotions, observable in a bimodal valence distribution. Dependent on a saturated or a superlinear feedback between the information field and the agent's arousal, we further identify scenarios where collective emotions only appear once or in a repeated manner. The analytical results are illustrated by agent-based computer simulations. Our framework provides testable hypotheses about the emergence of collective emotions, which can be verified by data from online communities.Comment: European Physical Journal B (in press), version 2 with extended introduction, clarification

    The perception of emotion in artificial agents

    Get PDF
    Given recent technological developments in robotics, artificial intelligence and virtual reality, it is perhaps unsurprising that the arrival of emotionally expressive and reactive artificial agents is imminent. However, if such agents are to become integrated into our social milieu, it is imperative to establish an understanding of whether and how humans perceive emotion in artificial agents. In this review, we incorporate recent findings from social robotics, virtual reality, psychology, and neuroscience to examine how people recognize and respond to emotions displayed by artificial agents. First, we review how people perceive emotions expressed by an artificial agent, such as facial and bodily expressions and vocal tone. Second, we evaluate the similarities and differences in the consequences of perceived emotions in artificial compared to human agents. Besides accurately recognizing the emotional state of an artificial agent, it is critical to understand how humans respond to those emotions. Does interacting with an angry robot induce the same responses in people as interacting with an angry person? Similarly, does watching a robot rejoice when it wins a game elicit similar feelings of elation in the human observer? Here we provide an overview of the current state of emotion expression and perception in social robotics, as well as a clear articulation of the challenges and guiding principles to be addressed as we move ever closer to truly emotional artificial agents

    Do Embodied Conversational Agents Know When to Smile?

    Get PDF
    We survey the role of humor in particular domains of human-to-human interaction with the aim of seeing whether it is useful for embodied conversational agents to integrate humor capabilities in their models of intelligence, emotions and interaction (verbal and nonverbal) Therefore we first look at the current state of the art of research in embodied conversational agents, affective computing and verbal and nonverbal interaction. We adhere to the 'Computers Are Social Actors' paradigm to assume that human conversational partners of embodied conversational agents assign human properties to these agents, including humor appreciation

    Motivations, Values and Emotions: 3 sides of the same coin

    Get PDF
    This position paper speaks to the interrelationships between the three concepts of motivations, values, and emotion. Motivations prime actions, values serve to choose between motivations, emotions provide a common currency for values, and emotions implement motivations. While conceptually distinct, the three are so pragmatically intertwined as to differ primarily from our taking different points of view. To make these points more transparent, we briefly describe the three in the context a cognitive architecture, the LIDA model, for software agents and robots that models human cognition, including a developmental period. We also compare the LIDA model with other models of cognition, some involving learning and emotions. Finally, we conclude that artificial emotions will prove most valuable as implementers of motivations in situations requiring learning and development
    • ā€¦
    corecore