12,274 research outputs found
A Case for Machine Ethics in Modeling Human-Level Intelligent Agents
This paper focuses on the research field of machine ethics and how it relates to a technological singularity—a hypothesized, futuristic event where artificial machines will have greater-than-human-level intelligence. One problem related to the singularity centers on the issue of whether human values and norms would survive such an event. To somehow ensure this, a number of artificial intelligence researchers have opted to focus on the development of artificial moral agents, which refers to machines capable of moral reasoning, judgment, and decision-making. To date, different frameworks on how to arrive at these agents have been put forward. However, there seems to be no hard consensus as to which framework would likely yield a positive result. With the body of work that they have contributed in the study of moral agency, philosophers may contribute to the growing literature on artificial moral agency. While doing so, they could also think about how the said concept could affect other important philosophical concepts
Dynamic Models Applied to Value Learning in Artificial Intelligence
Experts in Artificial Intelligence (AI) development predict that advances in
the development of intelligent systems and agents will reshape vital areas in
our society. Nevertheless, if such an advance is not made prudently and
critically-reflexively, it can result in negative outcomes for humanity. For
this reason, several researchers in the area are trying to develop a robust,
beneficial, and safe concept of AI for the preservation of humanity and the
environment. Currently, several of the open problems in the field of AI
research arise from the difficulty of avoiding unwanted behaviors of
intelligent agents and systems, and at the same time specifying what we want
such systems to do, especially when we look for the possibility of intelligent
agents acting in several domains over the long term. It is of utmost importance
that artificial intelligent agents have their values aligned with human values,
given the fact that we cannot expect an AI to develop human moral values simply
because of its intelligence, as discussed in the Orthogonality Thesis. Perhaps
this difficulty comes from the way we are addressing the problem of expressing
objectives, values, and ends, using representational cognitive methods. A
solution to this problem would be the dynamic approach proposed by Dreyfus,
whose phenomenological philosophy shows that the human experience of
being-in-the-world in several aspects is not well represented by the symbolic
or connectionist cognitive method, especially in regards to the question of
learning values. A possible approach to this problem would be to use
theoretical models such as SED (situated embodied dynamics) to address the
values learning problem in AI.Comment: 18 pages, no figures, published, translated versio
Impact of Artificial Intelligence on Businesses: from Research, Innovation, Market Deployment to Future Shifts in Business Models
The fast pace of artificial intelligence (AI) and automation is propelling
strategists to reshape their business models. This is fostering the integration
of AI in the business processes but the consequences of this adoption are
underexplored and need attention. This paper focuses on the overall impact of
AI on businesses - from research, innovation, market deployment to future
shifts in business models. To access this overall impact, we design a
three-dimensional research model, based upon the Neo-Schumpeterian economics
and its three forces viz. innovation, knowledge, and entrepreneurship. The
first dimension deals with research and innovation in AI. In the second
dimension, we explore the influence of AI on the global market and the
strategic objectives of the businesses and finally, the third dimension
examines how AI is shaping business contexts. Additionally, the paper explores
AI implications on actors and its dark sides.Comment: 38 pages, 10 figures, 3 tables. A part of this work has been
presented in DIGITS 201
Big data analytics:Computational intelligence techniques and application areas
Big Data has significant impact in developing functional smart cities and supporting modern societies. In this paper, we investigate the importance of Big Data in modern life and economy, and discuss challenges arising from Big Data utilization. Different computational intelligence techniques have been considered as tools for Big Data analytics. We also explore the powerful combination of Big Data and Computational Intelligence (CI) and identify a number of areas, where novel applications in real world smart city problems can be developed by utilizing these powerful tools and techniques. We present a case study for intelligent transportation in the context of a smart city, and a novel data modelling methodology based on a biologically inspired universal generative modelling approach called Hierarchical Spatial-Temporal State Machine (HSTSM). We further discuss various implications of policy, protection, valuation and commercialization related to Big Data, its applications and deployment
Superintelligence cannot be contained: Lessons from Computability Theory
Superintelligence is a hypothetical agent that possesses intelligence far
surpassing that of the brightest and most gifted human minds. In light of
recent advances in machine intelligence, a number of scientists, philosophers
and technologists have revived the discussion about the potential catastrophic
risks entailed by such an entity. In this article, we trace the origins and
development of the neo-fear of superintelligence, and some of the major
proposals for its containment. We argue that such containment is, in principle,
impossible, due to fundamental limits inherent to computing itself. Assuming
that a superintelligence will contain a program that includes all the programs
that can be executed by a universal Turing machine on input potentially as
complex as the state of the world, strict containment requires simulations of
such a program, something theoretically (and practically) infeasible.Comment: 7 pages, 5 figure
Towards a Framework Combining Machine Ethics and Machine Explainability
We find ourselves surrounded by a rapidly increasing number of autonomous and
semi-autonomous systems. Two grand challenges arise from this development:
Machine Ethics and Machine Explainability. Machine Ethics, on the one hand, is
concerned with behavioral constraints for systems, so that morally acceptable,
restricted behavior results; Machine Explainability, on the other hand, enables
systems to explain their actions and argue for their decisions, so that human
users can understand and justifiably trust them.
In this paper, we try to motivate and work towards a framework combining
Machine Ethics and Machine Explainability. Starting from a toy example, we
detect various desiderata of such a framework and argue why they should and how
they could be incorporated in autonomous systems. Our main idea is to apply a
framework of formal argumentation theory both, for decision-making under
ethical constraints and for the task of generating useful explanations given
only limited knowledge of the world. The result of our deliberations can be
described as a first version of an ethically motivated, principle-governed
framework combining Machine Ethics and Machine ExplainabilityComment: In Proceedings CREST 2018, arXiv:1901.0007
Assessing the impact of machine intelligence on human behaviour: an interdisciplinary endeavour
This document contains the outcome of the first Human behaviour and machine
intelligence (HUMAINT) workshop that took place 5-6 March 2018 in Barcelona,
Spain. The workshop was organized in the context of a new research programme at
the Centre for Advanced Studies, Joint Research Centre of the European
Commission, which focuses on studying the potential impact of artificial
intelligence on human behaviour. The workshop gathered an interdisciplinary
group of experts to establish the state of the art research in the field and a
list of future research challenges to be addressed on the topic of human and
machine intelligence, algorithm's potential impact on human cognitive
capabilities and decision making, and evaluation and regulation needs. The
document is made of short position statements and identification of challenges
provided by each expert, and incorporates the result of the discussions carried
out during the workshop. In the conclusion section, we provide a list of
emerging research topics and strategies to be addressed in the near future.Comment: Proceedings of 1st HUMAINT (Human Behaviour and Machine Intelligence)
workshop, Barcelona, Spain, March 5-6, 2018, edited by European Commission,
Seville, 2018, JRC111773
https://ec.europa.eu/jrc/communities/community/humaint/document/assessing-impact-machine-intelligence-human-behaviour-interdisciplinary.
arXiv admin note: text overlap with arXiv:1409.3097 by other author
Explanation in Human-AI Systems: A Literature Meta-Review, Synopsis of Key Ideas and Publications, and Bibliography for Explainable AI
This is an integrative review that address the question, "What makes for a
good explanation?" with reference to AI systems. Pertinent literatures are
vast. Thus, this review is necessarily selective. That said, most of the key
concepts and issues are expressed in this Report. The Report encapsulates the
history of computer science efforts to create systems that explain and instruct
(intelligent tutoring systems and expert systems). The Report expresses the
explainability issues and challenges in modern AI, and presents capsule views
of the leading psychological theories of explanation. Certain articles stand
out by virtue of their particular relevance to XAI, and their methods, results,
and key points are highlighted. It is recommended that AI/XAI researchers be
encouraged to include in their research reports fuller details on their
empirical or experimental methods, in the fashion of experimental psychology
research reports: details on Participants, Instructions, Procedures, Tasks,
Dependent Variables (operational definitions of the measures and metrics),
Independent Variables (conditions), and Control Conditions
Consequentialism & Machine Ethics: Towards a Foundational Machine Ethic to Ensure the Right Action of Artificial Moral Agents
In this paper, I argue that Consequentialism represents a kind of ethical theory that is the most plausible to serve as a basis for a machine ethic. First, I outline the concept of an artificial moral agent and the essential properties of Consequentialism. Then, I present a scenario involving autonomous vehicles to illustrate how the features of Consequentialism inform agent action. Thirdly, an alternative Deontological approach will be evaluated and the problem of moral conflict discussed. Finally, two bottom-up approaches to the development of machine ethics are presented and briefly challenged
Artificial consciousness and the consciousness-attention dissociation
Artificial Intelligence is at a turning point, with a substantial increase in projects aiming to implement sophisticated forms of human intelligence in machines. This research attempts to model specific forms of intelligence through brute-force search heuristics and also reproduce features of human perception and cognition, including emotions. Such goals have implications for artificial consciousness, with some arguing that it will be achievable once we overcome short-term engineering challenges. We believe, however, that phenomenal consciousness cannot be implemented in machines. This becomes clear when considering emotions and examining the dissociation between consciousness and attention in humans. While we may be able to program ethical behavior based on rules and machine learning, we will never be able to reproduce emotions or empathy by programming such control systems—these will be merely simulations. Arguments in favor of this claim include considerations about evolution, the neuropsychological aspects of emotions, and the dissociation between attention and consciousness found in humans. Ultimately, we are far from achieving artificial consciousness
- …