Article thumbnail

Differentiable Particle Filters: End-to-End Learning with Algorithmic Priors

By Rico Jonschkowski, Divyam Rastogi and Oliver Brock

Abstract

We present differentiable particle filters (DPFs): a differentiable implementation of the particle filter algorithm with learnable motion and measurement models. Since DPFs are end-to-end differentiable, we can efficiently train their models by optimizing end-to-end state estimation performance, rather than proxy objectives such as model accuracy. DPFs encode the structure of recursive state estimation with prediction and measurement update that operate on a probability distribution over states. This structure represents an algorithmic prior that improves learning performance in state estimation problems while enabling explainability of the learned model. Our experiments on simulated and real data show substantial benefits from end-to- end learning with algorithmic priors, e.g. reducing error rates by ~80%. Our experiments also show that, unlike long short-term memory networks, DPFs learn localization in a policy-agnostic way and thus greatly improve generalization. Source code is available at https://github.com/tu-rbo/differentiable-particle-filters .Comment: Accepted at Robotics: Science and Systems 2018 (http://www.roboticsconference.org

Topics: Computer Science - Machine Learning, Computer Science - Artificial Intelligence, Computer Science - Robotics, Statistics - Machine Learning
Year: 2018
OAI identifier: oai:arXiv.org:1805.11122

Suggested articles


To submit an update or takedown request for this paper, please submit an Update/Correction/Removal Request.