What criminal and civil law tells us about Safe RL techniques to generate law-abiding behaviour

Abstract

Safe Reinforcement Learning (Safe RL) aims to produce constrained policies with constraints typically motivated by issues of physical safety. This paper considers the issues that arise from regulatory constraints or issues of legal safety. Without guarantees of safety, autonomous systems or agents (A-bots) trained through RL are expensive or dangerous to train and deploy. Many potential applications for RL involve acting in regulated environments and here existing research is thin. Regulations impose behavioural restrictions which can be more complex than those engendered by considerations of physical safety. They are often inter-temporal, require planning on behalf of the learner and involve concepts of causality and intent. By examining the typical types of laws present in a regulated arena, this paper identifies design features that the RL learning process should possess in order to ensure that it is able to generate legally safe or compliant policies

    Similar works