2 research outputs found
SuperSuit: Simple Microwrappers for Reinforcement Learning Environments
In reinforcement learning, wrappers are universally used to transform the
information that passes between a model and an environment. Despite their
ubiquity, no library exists with reasonable implementations of all popular
preprocessing methods. This leads to unnecessary bugs, code inefficiencies, and
wasted developer time. Accordingly we introduce SuperSuit, a Python library
that includes all popular wrappers, and wrappers that can easily apply lambda
functions to the observations/actions/reward. It's compatible with the standard
Gym environment specification, as well as the PettingZoo specification for
multi-agent environments. The library is available at
https://github.com/PettingZoo-Team/SuperSuit,and can be installed via pip
Poisoning Deep Reinforcement Learning Agents with In-Distribution Triggers
In this paper, we propose a new data poisoning attack and apply it to deep
reinforcement learning agents. Our attack centers on what we call
in-distribution triggers, which are triggers native to the data distributions
the model will be trained on and deployed in. We outline a simple procedure for
embedding these, and other, triggers in deep reinforcement learning agents
following a multi-task learning paradigm, and demonstrate in three common
reinforcement learning environments. We believe that this work has important
implications for the security of deep learning models.Comment: 4 pages, 1 figure, Published at ICLR 2021 Workshop on Security and
Safety in Machine Learning System