13 research outputs found

    Developmental changes in perceived moral standing of robots

    Get PDF
    We live in an age where robots are increasingly present in the social and moral world. Here, we explore how children and adults think about the mental lives and moral standing of robots. In Experiment 1 (N = 116), we found that children granted humans and robots with more mental life and vulnerability to harm than an anthropomorphized control (i.e., a toy bear). In Experiment 2 (N = 157), we found that, relative to children, adults ascribed less mental life and vulnerability to harm to robots. In Experiment 3 (N = 152), we modified our experiment to be within-subjects and measured beliefs concerning moral standing. Though younger children again appeared willing to assign mental capacities — particularly those related to experience (e.g., being capable of experiencing hunger) — to robots, older children and adults did so to a lesser degree. This diminished attribution of mental life tracked with diminished ratings of robot moral standing. This informs ongoing debates concerning emerging attitudes about artificial life

    Doing the right thing for the right reason: Evaluating artificial moral cognition by probing cost insensitivity

    Full text link
    Is it possible to evaluate the moral cognition of complex artificial agents? In this work, we take a look at one aspect of morality: `doing the right thing for the right reasons.' We propose a behavior-based analysis of artificial moral cognition which could also be applied to humans to facilitate like-for-like comparison. Morally-motivated behavior should persist despite mounting cost; by measuring an agent's sensitivity to this cost, we gain deeper insight into underlying motivations. We apply this evaluation to a particular set of deep reinforcement learning agents, trained by memory-based meta-reinforcement learning. Our results indicate that agents trained with a reward function that includes other-regarding preferences perform helping behavior in a way that is less sensitive to increasing cost than agents trained with more self-interested preferences.Comment: 11 pages, 3 figure

    Mind perception and moral judgment

    No full text

    Immutable morality: Even God could not change some moral facts

    No full text
    The idea that morality depends on God is a widely held belief. This belief entails that the moral “facts” could be otherwise because, in principle, God could change them. Yet, some moral propositions seem so obviously true (e.g., the immorality of killing someone just for pleasure) that it is hard to imagine how they could be otherwise. In two experiments, we investigated people’s intuitions about the immutability of moral facts. Participants judged whether it was even possible, or possible for God, to change moral, logical, and physical facts. In both experiments, people judged that altering some moral facts was impossible—not even God could turn morally wrong acts into morally right acts. Strikingly, people thought that God could make physically impossible and logically impossible events occur. These results demonstrate the strength of people’s metaethical commitments and shed light on the nature of morality and its centrality to thinking and reasoning

    Qualtrics Survey

    No full text

    Supplemental Online Materials

    No full text

    Immutable Morality

    No full text

    Procedure, Variables, Sampling

    No full text

    Project Overview

    No full text

    Predictions and Hypotheses

    No full text
    corecore