4 research outputs found
Augmented Utilitarianism for AGI Safety
In the light of ongoing progresses of research on artificial intelligent
systems exhibiting a steadily increasing problem-solving ability, the
identification of practicable solutions to the value alignment problem in AGI
Safety is becoming a matter of urgency. In this context, one preeminent
challenge that has been addressed by multiple researchers is the adequate
formulation of utility functions or equivalents reliably capturing human
ethical conceptions. However, the specification of suitable utility functions
harbors the risk of "perverse instantiation" for which no final consensus on
responsible proactive countermeasures has been achieved so far. Amidst this
background, we propose a novel socio-technological ethical framework denoted
Augmented Utilitarianism which directly alleviates the perverse instantiation
problem. We elaborate on how augmented by AI and more generally science and
technology, it might allow a society to craft and update ethical utility
functions while jointly undergoing a dynamical ethical enhancement. Further, we
elucidate the need to consider embodied simulations in the design of utility
functions for AGIs aligned with human values. Finally, we discuss future
prospects regarding the usage of the presented scientifically grounded ethical
framework and mention possible challenges
Requisite Variety in Ethical Utility Functions for AI Value Alignment
Being a complex subject of major importance in AI Safety research, value
alignment has been studied from various perspectives in the last years.
However, no final consensus on the design of ethical utility functions
facilitating AI value alignment has been achieved yet. Given the urgency to
identify systematic solutions, we postulate that it might be useful to start
with the simple fact that for the utility function of an AI not to violate
human ethical intuitions, it trivially has to be a model of these intuitions
and reflect their variety whereby the most accurate models pertaining to
human entities being biological organisms equipped with a brain constructing
concepts like moral judgements, are scientific models. Thus, in order to better
assess the variety of human morality, we perform a transdisciplinary analysis
applying a security mindset to the issue and summarizing variety-relevant
background knowledge from neuroscience and psychology. We complement this
information by linking it to augmented utilitarianism as a suitable ethical
framework. Based on that, we propose first practical guidelines for the design
of approximate ethical goal functions that might better capture the variety of
human moral judgements. Finally, we conclude and address future possible
challenges.Comment: IJCAI 2019 AI Safety Worksho
Augmented Utilitarianism for AGI Safety
In the light of ongoing progresses of research on artificial intelligent systems exhibiting a steadily increasing problem-solving ability, the identification of practicable solutions to the value alignment problem in AGI Safety is becoming a matter of urgency. In this context, one preeminent challenge that has been addressed by multiple researchers is the adequate formulation of utility functions or equivalents reliably capturing human ethical conceptions. However, the specification of suitable utility functions harbors the risk of “perverse instantiation” for which no final consensus on responsible proactive countermeasures has been achieved so far. Amidst this background, we propose a novel non-normative socio-technological ethical framework denoted Augmented Utilitarianism which directly alleviates the perverse instantiation problem. We elaborate on how augmented by AI and more generally science and technology, it might allow a society to craft and update ethical utility functions while jointly undergoing a dynamical ethical enhancement. Further, we elucidate the need to consider embodied simulations in the design of utility functions for AGIs aligned with human values. Finally, we discuss future prospects regarding the usage of the presented scientifically grounded ethical framework and mention possible challenges
Augmented Utilitarianism for AGI Safety
In the light of ongoing progresses of research on artificial intelligent systems exhibiting a steadily increasing problem-solving ability, the identification of practicable solutions to the value alignment problem in AGI Safety is becoming a matter of urgency. In this context, one preeminent challenge that has been addressed by multiple researchers is the adequate formulation of utility functions or equivalents reliably capturing human ethical conceptions. However, the specification of suitable utility functions harbors the risk of “perverse instantiation” for which no final consensus on responsible proactive countermeasures has been achieved so far. Amidst this background, we propose a novel non-normative socio-technological ethical framework denoted Augmented Utilitarianism which directly alleviates the perverse instantiation problem. We elaborate on how augmented by AI and more generally science and technology, it might allow a society to craft and update ethical utility functions while jointly undergoing a dynamical ethical enhancement. Further, we elucidate the need to consider embodied simulations in the design of utility functions for AGIs aligned with human values. Finally, we discuss future prospects regarding the usage of the presented scientifically grounded ethical framework and mention possible challenges