142,614 research outputs found
Towards artificial situation awareness by autonomous vehicles
This paper presents a novel approach to artificial situation awareness for an autonomous vehicle operating in complex dynamic environments populated by other agents. A key aspect of situation awareness is the use of mental models to predict future states of the environment, allowing safe and rational routing decisions to be made. We present a technique for predicting future discrete state transitions (such as the commencement of a turn) by other agents, based upon an uncertain mental model. Predictions take the form of univariate Gaussian Probability Density Functions which capture the inherent uncertainty in transition time whilst still providing great benefit to a decision making system. The prediction distributions are compared with Monte Carlo simulations and show an excellent correlation over long prediction horizons
Game Theory Models for the Verification of the Collective Behaviour of Autonomous Cars
The collective of autonomous cars is expected to generate almost optimal
traffic. In this position paper we discuss the multi-agent models and the
verification results of the collective behaviour of autonomous cars. We argue
that non-cooperative autonomous adaptation cannot guarantee optimal behaviour.
The conjecture is that intention aware adaptation with a constraint on
simultaneous decision making has the potential to avoid unwanted behaviour. The
online routing game model is expected to be the basis to formally prove this
conjecture.Comment: In Proceedings FVAV 2017, arXiv:1709.0212
Intelligent Agents for Disaster Management
ALADDIN [1] is a multi-disciplinary project that is developing novel techniques, architectures, and mechanisms for multi-agent systems in uncertain and dynamic environments. The application focus of the project is disaster management. Research within a number of themes is being pursued and this is considering different aspects of the interaction between autonomous agents and the decentralised system architectures that support those interactions. The aim of the research is to contribute to building more robust multi-agent systems for future applications in disaster management and other similar domains
Responsible Autonomy
As intelligent systems are increasingly making decisions that directly affect
society, perhaps the most important upcoming research direction in AI is to
rethink the ethical implications of their actions. Means are needed to
integrate moral, societal and legal values with technological developments in
AI, both during the design process as well as part of the deliberation
algorithms employed by these systems. In this paper, we describe leading ethics
theories and propose alternative ways to ensure ethical behavior by artificial
systems. Given that ethics are dependent on the socio-cultural context and are
often only implicit in deliberation processes, methodologies are needed to
elicit the values held by designers and stakeholders, and to make these
explicit leading to better understanding and trust on artificial autonomous
systems.Comment: IJCAI2017 (International Joint Conference on Artificial Intelligence
Adjustably Autonomous Multi-agent Plan Execution with an Internal Spacecraft Free-Flying Robot Prototype
We present an multi-agent model-based autonomy architecture with monitoring, planning, diagnosis, and execution elements. We discuss an internal spacecraft free-flying robot prototype controlled by an implementation of this architecture and a ground test facility used for development. In addition, we discuss a simplified environment control life support system for the spacecraft domain also controlled by an implementation of this architecture. We discuss adjustable autonomy and how it applies to this architecture. We describe an interface that provides the user situation awareness of both autonomous systems and enables the user to dynamically edit the plans prior to and during execution as well as control these agents at various levels of autonomy. This interface also permits the agents to query the user or request the user to perform tasks to help achieve the commanded goals. We conclude by describing a scenario where these two agents and a human interact to cooperatively detect, diagnose and recover from a simulated spacecraft fault
Mixed Initiative Systems for Human-Swarm Interaction: Opportunities and Challenges
Human-swarm interaction (HSI) involves a number of human factors impacting
human behaviour throughout the interaction. As the technologies used within HSI
advance, it is more tempting to increase the level of swarm autonomy within the
interaction to reduce the workload on humans. Yet, the prospective negative
effects of high levels of autonomy on human situational awareness can hinder
this process. Flexible autonomy aims at trading-off these effects by changing
the level of autonomy within the interaction when required; with
mixed-initiatives combining human preferences and automation's recommendations
to select an appropriate level of autonomy at a certain point of time. However,
the effective implementation of mixed-initiative systems raises fundamental
questions on how to combine human preferences and automation recommendations,
how to realise the selected level of autonomy, and what the future impacts on
the cognitive states of a human are. We explore open challenges that hamper the
process of developing effective flexible autonomy. We then highlight the
potential benefits of using system modelling techniques in HSI by illustrating
how they provide HSI designers with an opportunity to evaluate different
strategies for assessing the state of the mission and for adapting the level of
autonomy within the interaction to maximise mission success metrics.Comment: Author version, accepted at the 2018 IEEE Annual Systems Modelling
Conference, Canberra, Australi
- …