40,147 research outputs found
Online Predictive Optimization Framework for Stochastic Demand-Responsive Transit Services
This study develops an online predictive optimization framework for
dynamically operating a transit service in an area of crowd movements. The
proposed framework integrates demand prediction and supply optimization to
periodically redesign the service routes based on recently observed demand. To
predict demand for the service, we use Quantile Regression to estimate the
marginal distribution of movement counts between each pair of serviced
locations. The framework then combines these marginals into a joint demand
distribution by constructing a Gaussian copula, which captures the structure of
correlation between the marginals. For supply optimization, we devise a linear
programming model, which simultaneously determines the route structure and the
service frequency according to the predicted demand. Importantly, our framework
both preserves the uncertainty structure of future demand and leverages this
for robust route optimization, while keeping both components decoupled. We
evaluate our framework using a real-world case study of autonomous mobility in
a university campus in Denmark. The results show that our framework often
obtains the ground truth optimal solution, and can outperform conventional
methods for route optimization, which do not leverage full predictive
distributions.Comment: 34 pages, 12 figures, 5 table
Chance-Constrained Trajectory Optimization for Safe Exploration and Learning of Nonlinear Systems
Learning-based control algorithms require data collection with abundant
supervision for training. Safe exploration algorithms ensure the safety of this
data collection process even when only partial knowledge is available. We
present a new approach for optimal motion planning with safe exploration that
integrates chance-constrained stochastic optimal control with dynamics learning
and feedback control. We derive an iterative convex optimization algorithm that
solves an \underline{Info}rmation-cost \underline{S}tochastic
\underline{N}onlinear \underline{O}ptimal \underline{C}ontrol problem
(Info-SNOC). The optimization objective encodes both optimal performance and
exploration for learning, and the safety is incorporated as distributionally
robust chance constraints. The dynamics are predicted from a robust regression
model that is learned from data. The Info-SNOC algorithm is used to compute a
sub-optimal pool of safe motion plans that aid in exploration for learning
unknown residual dynamics under safety constraints. A stable feedback
controller is used to execute the motion plan and collect data for model
learning. We prove the safety of rollout from our exploration method and
reduction in uncertainty over epochs, thereby guaranteeing the consistency of
our learning method. We validate the effectiveness of Info-SNOC by designing
and implementing a pool of safe trajectories for a planar robot. We demonstrate
that our approach has higher success rate in ensuring safety when compared to a
deterministic trajectory optimization approach.Comment: Submitted to RA-L 2020, review-
Closed-Loop Statistical Verification of Stochastic Nonlinear Systems Subject to Parametric Uncertainties
This paper proposes a statistical verification framework using Gaussian
processes (GPs) for simulation-based verification of stochastic nonlinear
systems with parametric uncertainties. Given a small number of stochastic
simulations, the proposed framework constructs a GP regression model and
predicts the system's performance over the entire set of possible
uncertainties. Included in the framework is a new metric to estimate the
confidence in those predictions based on the variance of the GP's cumulative
distribution function. This variance-based metric forms the basis of active
sampling algorithms that aim to minimize prediction error through careful
selection of simulations. In three case studies, the new active sampling
algorithms demonstrate up to a 35% improvement in prediction error over other
approaches and are able to correctly identify regions with low prediction
confidence through the variance metric.Comment: 8 pages, submitted to ACC 201
Sequential Monte Carlo pricing of American-style options under stochastic volatility models
We introduce a new method to price American-style options on underlying
investments governed by stochastic volatility (SV) models. The method does not
require the volatility process to be observed. Instead, it exploits the fact
that the optimal decision functions in the corresponding dynamic programming
problem can be expressed as functions of conditional distributions of
volatility, given observed data. By constructing statistics summarizing
information about these conditional distributions, one can obtain high quality
approximate solutions. Although the required conditional distributions are in
general intractable, they can be arbitrarily precisely approximated using
sequential Monte Carlo schemes. The drawback, as with many Monte Carlo schemes,
is potentially heavy computational demand. We present two variants of the
algorithm, one closely related to the well-known least-squares Monte Carlo
algorithm of Longstaff and Schwartz [The Review of Financial Studies 14 (2001)
113-147], and the other solving the same problem using a "brute force" gridding
approach. We estimate an illustrative SV model using Markov chain Monte Carlo
(MCMC) methods for three equities. We also demonstrate the use of our algorithm
by estimating the posterior distribution of the market price of volatility risk
for each of the three equities.Comment: Published in at http://dx.doi.org/10.1214/09-AOAS286 the Annals of
Applied Statistics (http://www.imstat.org/aoas/) by the Institute of
Mathematical Statistics (http://www.imstat.org
Automatic Differentiation Variational Inference
Probabilistic modeling is iterative. A scientist posits a simple model, fits
it to her data, refines it according to her analysis, and repeats. However,
fitting complex models to large data is a bottleneck in this process. Deriving
algorithms for new models can be both mathematically and computationally
challenging, which makes it difficult to efficiently cycle through the steps.
To this end, we develop automatic differentiation variational inference (ADVI).
Using our method, the scientist only provides a probabilistic model and a
dataset, nothing else. ADVI automatically derives an efficient variational
inference algorithm, freeing the scientist to refine and explore many models.
ADVI supports a broad class of models-no conjugacy assumptions are required. We
study ADVI across ten different models and apply it to a dataset with millions
of observations. ADVI is integrated into Stan, a probabilistic programming
system; it is available for immediate use
- …