9,201 research outputs found

    Towards Identifying and closing Gaps in Assurance of autonomous Road vehicleS - a collection of Technical Notes Part 1

    Get PDF
    This report provides an introduction and overview of the Technical Topic Notes (TTNs) produced in the Towards Identifying and closing Gaps in Assurance of autonomous Road vehicleS (Tigars) project. These notes aim to support the development and evaluation of autonomous vehicles. Part 1 addresses: Assurance-overview and issues, Resilience and Safety Requirements, Open Systems Perspective and Formal Verification and Static Analysis of ML Systems. Part 2: Simulation and Dynamic Testing, Defence in Depth and Diversity, Security-Informed Safety Analysis, Standards and Guidelines

    Resilience, reliability, and coordination in autonomous multi-agent systems

    Get PDF
    Acknowledgements The research reported in this paper was funded and supported by various grants over the years: Robotics and AI in Nuclear (RAIN) Hub (EP/R026084/1); Future AI and Robotics for Space (FAIR-SPACE) Hub (EP/R026092/1); Offshore Robotics for Certification of Assets (ORCA) Hub (EP/R026173/1); the Royal Academy of Engineering under the Chair in Emerging Technologies scheme; Trustworthy Autonomous Systems “Verifiability Node” (EP/V026801); Scrutable Autonomous Systems (EP/J012084/1); Supporting Security Policy with Effective Digital Intervention (EP/P011829/1); The International Technology Alliance in Network and Information Sciences.Peer reviewedPostprin

    Culture-Based Explainable Human-Agent Deconfliction

    Get PDF
    Law codes and regulations help organise societies for centuries, and as AI systems gain more autonomy, we question how human-agent systems can operate as peers under the same norms, especially when resources are contended. We posit that agents must be accountable and explainable by referring to which rules justify their decisions. The need for explanations is associated with user acceptance and trust. This paper's contribution is twofold: i) we propose an argumentation-based human-agent architecture to map human regulations into a culture for artificial agents with explainable behaviour. Our architecture leans on the notion of argumentative dialogues and generates explanations from the history of such dialogues; and ii) we validate our architecture with a user study in the context of human-agent path deconfliction. Our results show that explanations provide a significantly higher improvement in human performance when systems are more complex. Consequently, we argue that the criteria defining the need of explanations should also consider the complexity of a system. Qualitative findings show that when rules are more complex, explanations significantly reduce the perception of challenge for humans.L3Harris ASV and the Royal Commission for the Exhibition of 185

    Perceptions on proof and the teaching of proof: A comparison across preservice secondary teachers in Australia, USA and Korea

    Get PDF
    Despite the recognised importance of mathematical proof in secondary education, there is a limited but growing body of literature indicating how preservice secondary mathematics teachers (PSMTs) view proof and the teaching of proof. The purpose of this survey research was to investigate how PSMTs in Australia, the USA and Korea perceive of proof in the context of secondary mathematics teaching and learning. PSMTs were able to outline various mathematical and pedagogical aspects of proof, including purposes, characteristics, reasons for teaching and imposed constraints. In addition, PSMTs attended to differing, though overlapping, features of proof when asked to determine the extent to which proposed arguments constituted proofs or to decide which arguments they might present to students
    corecore