206 research outputs found
PennyLane: Automatic differentiation of hybrid quantum-classical computations
PennyLane is a Python 3 software framework for optimization and machine
learning of quantum and hybrid quantum-classical computations. The library
provides a unified architecture for near-term quantum computing devices,
supporting both qubit and continuous-variable paradigms. PennyLane's core
feature is the ability to compute gradients of variational quantum circuits in
a way that is compatible with classical techniques such as backpropagation.
PennyLane thus extends the automatic differentiation algorithms common in
optimization and machine learning to include quantum and hybrid computations. A
plugin system makes the framework compatible with any gate-based quantum
simulator or hardware. We provide plugins for Strawberry Fields, Rigetti
Forest, Qiskit, Cirq, and ProjectQ, allowing PennyLane optimizations to be run
on publicly accessible quantum devices provided by Rigetti and IBM Q. On the
classical front, PennyLane interfaces with accelerated machine learning
libraries such as TensorFlow, PyTorch, and autograd. PennyLane can be used for
the optimization of variational quantum eigensolvers, quantum approximate
optimization, quantum machine learning models, and many other applications.Comment: Code available at https://github.com/XanaduAI/pennylane/ .
Significant contributions to the code (new features, new plugins, etc.) will
be recognized by the opportunity to be a co-author on this pape
Mining gold from implicit models to improve likelihood-free inference
Simulators often provide the best description of real-world phenomena.
However, they also lead to challenging inverse problems because the density
they implicitly define is often intractable. We present a new suite of
simulation-based inference techniques that go beyond the traditional
Approximate Bayesian Computation approach, which struggles in a
high-dimensional setting, and extend methods that use surrogate models based on
neural networks. We show that additional information, such as the joint
likelihood ratio and the joint score, can often be extracted from simulators
and used to augment the training data for these surrogate models. Finally, we
demonstrate that these new techniques are more sample efficient and provide
higher-fidelity inference than traditional methods.Comment: Code available at
https://github.com/johannbrehmer/simulator-mining-example . v2: Fixed typos.
v3: Expanded discussion, added Lotka-Volterra example. v4: Improved clarit
Deep Generative Modeling of LiDAR Data
Building models capable of generating structured output is a key challenge
for AI and robotics. While generative models have been explored on many types
of data, little work has been done on synthesizing lidar scans, which play a
key role in robot mapping and localization. In this work, we show that one can
adapt deep generative models for this task by unravelling lidar scans into a 2D
point map. Our approach can generate high quality samples, while simultaneously
learning a meaningful latent representation of the data. We demonstrate
significant improvements against state-of-the-art point cloud generation
methods. Furthermore, we propose a novel data representation that augments the
2D signal with absolute positional information. We show that this helps
robustness to noisy and imputed input; the learned model can recover the
underlying lidar scan from seemingly uninformative dataComment: Presented at IROS 201
AutoSimulate: (Quickly) Learning Synthetic Data Generation
Simulation is increasingly being used for generating large labelled datasets
in many machine learning problems. Recent methods have focused on adjusting
simulator parameters with the goal of maximising accuracy on a validation task,
usually relying on REINFORCE-like gradient estimators. However these approaches
are very expensive as they treat the entire data generation, model training,
and validation pipeline as a black-box and require multiple costly objective
evaluations at each iteration. We propose an efficient alternative for optimal
synthetic data generation, based on a novel differentiable approximation of the
objective. This allows us to optimize the simulator, which may be
non-differentiable, requiring only one objective evaluation at each iteration
with a little overhead. We demonstrate on a state-of-the-art photorealistic
renderer that the proposed method finds the optimal data distribution faster
(up to ), with significantly reduced training data generation (up to
) and better accuracy () on real-world test datasets than
previous methods.Comment: ECCV 202
- …