7,951 research outputs found
On the Sample Size of Random Convex Programs with Structured Dependence on the Uncertainty (Extended Version)
The "scenario approach" provides an intuitive method to address chance
constrained problems arising in control design for uncertain systems. It
addresses these problems by replacing the chance constraint with a finite
number of sampled constraints (scenarios). The sample size critically depends
on Helly's dimension, a quantity always upper bounded by the number of decision
variables. However, this standard bound can lead to computationally expensive
programs whose solutions are conservative in terms of cost and violation
probability. We derive improved bounds of Helly's dimension for problems where
the chance constraint has certain structural properties. The improved bounds
lower the number of scenarios required for these problems, leading both to
improved objective value and reduced computational complexity. Our results are
generally applicable to Randomized Model Predictive Control of chance
constrained linear systems with additive uncertainty and affine disturbance
feedback. The efficacy of the proposed bound is demonstrated on an inventory
management example.Comment: Accepted for publication at Automatic
A scenario approach for non-convex control design
Randomized optimization is an established tool for control design with
modulated robustness. While for uncertain convex programs there exist
randomized approaches with efficient sampling, this is not the case for
non-convex problems. Approaches based on statistical learning theory are
applicable to non-convex problems, but they usually are conservative in terms
of performance and require high sample complexity to achieve the desired
probabilistic guarantees. In this paper, we derive a novel scenario approach
for a wide class of random non-convex programs, with a sample complexity
similar to that of uncertain convex programs and with probabilistic guarantees
that hold not only for the optimal solution of the scenario program, but for
all feasible solutions inside a set of a-priori chosen complexity. We also
address measure-theoretic issues for uncertain convex and non-convex programs.
Among the family of non-convex control- design problems that can be addressed
via randomization, we apply our scenario approach to randomized Model
Predictive Control for chance-constrained nonlinear control-affine systems.Comment: Submitted to IEEE Transactions on Automatic Contro
From Uncertainty Data to Robust Policies for Temporal Logic Planning
We consider the problem of synthesizing robust disturbance feedback policies
for systems performing complex tasks. We formulate the tasks as linear temporal
logic specifications and encode them into an optimization framework via
mixed-integer constraints. Both the system dynamics and the specifications are
known but affected by uncertainty. The distribution of the uncertainty is
unknown, however realizations can be obtained. We introduce a data-driven
approach where the constraints are fulfilled for a set of realizations and
provide probabilistic generalization guarantees as a function of the number of
considered realizations. We use separate chance constraints for the
satisfaction of the specification and operational constraints. This allows us
to quantify their violation probabilities independently. We compute disturbance
feedback policies as solutions of mixed-integer linear or quadratic
optimization problems. By using feedback we can exploit information of past
realizations and provide feasibility for a wider range of situations compared
to static input sequences. We demonstrate the proposed method on two robust
motion-planning case studies for autonomous driving
Data-driven Distributionally Robust Optimization Using the Wasserstein Metric: Performance Guarantees and Tractable Reformulations
We consider stochastic programs where the distribution of the uncertain
parameters is only observable through a finite training dataset. Using the
Wasserstein metric, we construct a ball in the space of (multivariate and
non-discrete) probability distributions centered at the uniform distribution on
the training samples, and we seek decisions that perform best in view of the
worst-case distribution within this Wasserstein ball. The state-of-the-art
methods for solving the resulting distributionally robust optimization problems
rely on global optimization techniques, which quickly become computationally
excruciating. In this paper we demonstrate that, under mild assumptions, the
distributionally robust optimization problems over Wasserstein balls can in
fact be reformulated as finite convex programs---in many interesting cases even
as tractable linear programs. Leveraging recent measure concentration results,
we also show that their solutions enjoy powerful finite-sample performance
guarantees. Our theoretical results are exemplified in mean-risk portfolio
optimization as well as uncertainty quantification.Comment: 42 pages, 10 figure
A Posteriori Probabilistic Bounds of Convex Scenario Programs with Validation Tests
Scenario programs have established themselves as efficient tools towards
decision-making under uncertainty. To assess the quality of scenario-based
solutions a posteriori, validation tests based on Bernoulli trials have been
widely adopted in practice. However, to reach a theoretically reliable
judgement of risk, one typically needs to collect massive validation samples.
In this work, we propose new a posteriori bounds for convex scenario programs
with validation tests, which are dependent on both realizations of support
constraints and performance on out-of-sample validation data. The proposed
bounds enjoy wide generality in that many existing theoretical results can be
incorporated as particular cases. To facilitate practical use, a systematic
approach for parameterizing a posteriori probability bounds is also developed,
which is shown to possess a variety of desirable properties allowing for easy
implementations and clear interpretations. By synthesizing comprehensive
information about support constraints and validation tests, improved risk
evaluation can be achieved for randomized solutions in comparison with existing
a posteriori bounds. Case studies on controller design of aircraft lateral
motion are presented to validate the effectiveness of the proposed a posteriori
bounds
Scalable First-Order Methods for Robust MDPs
Robust Markov Decision Processes (MDPs) are a powerful framework for modeling
sequential decision-making problems with model uncertainty. This paper proposes
the first first-order framework for solving robust MDPs. Our algorithm
interleaves primal-dual first-order updates with approximate Value Iteration
updates. By carefully controlling the tradeoff between the accuracy and cost of
Value Iteration updates, we achieve an ergodic convergence rate of for the best
choice of parameters on ellipsoidal and Kullback-Leibler -rectangular
uncertainty sets, where and is the number of states and actions,
respectively. Our dependence on the number of states and actions is
significantly better (by a factor of ) than that of pure
Value Iteration algorithms. In numerical experiments on ellipsoidal uncertainty
sets we show that our algorithm is significantly more scalable than
state-of-the-art approaches. Our framework is also the first one to solve
robust MDPs with -rectangular KL uncertainty sets
- âŠ