128 research outputs found
The Computation of Approximate Generalized Feedback Nash Equilibria
We present the concept of a Generalized Feedback Nash Equilibrium (GFNE) in
dynamic games, extending the Feedback Nash Equilibrium concept to games in
which players are subject to state and input constraints. We formalize
necessary and sufficient conditions for (local) GFNE solutions at the
trajectory level, which enable the development of efficient numerical methods
for their computation. Specifically, we propose a Newton-style method for
finding game trajectories which satisfy the necessary conditions, which can
then be checked against the sufficiency conditions. We show that the evaluation
of the necessary conditions in general requires computing a series of nested,
implicitly-defined derivatives, which quickly becomes intractable. To this end,
we introduce an approximation to the necessary conditions which is amenable to
efficient evaluation, and in turn, computation of solutions. We term the
solutions to the approximate necessary conditions Generalized Feedback Quasi
Nash Equilibria (GFQNE), and we introduce numerical methods for their
computation. In particular, we develop a Sequential Linear-Quadratic Game
approach, in which a locally approximate LQ game is solved at each iteration.
The development of this method relies on the ability to compute a GFNE to
inequality- and equality-constrained LQ games, and therefore specific methods
for the solution of these special cases are developed in detail. We demonstrate
the effectiveness of the proposed solution approach on a dynamic game arising
in an autonomous driving application
Stochastic Extended LQR for Optimization-Based Motion Planning Under Uncertainty
We introduce a novel optimization-based motion planner, Stochastic Extended LQR (SELQR), which computes a trajectory and associated linear control policy with the objective of minimizing the expected value of a user-defined cost function. SELQR applies to robotic systems that have stochastic non-linear dynamics with motion uncertainty modeled by Gaussian distributions that can be state- and control-dependent. In each iteration, SELQR uses a combination of forward and backward value iteration to estimate the cost-to-come and the cost-to-go for each state along a trajectory. SELQR then locally optimizes each state along the trajectory at each iteration to minimize the expected total cost, which results in smoothed states that are used for dynamics linearization and cost function quadratization. SELQR progressively improves the approximation of the expected total cost, resulting in higher quality plans. For applications with imperfect sensing, we extend SELQR to plan in the robot's belief space. We show that our iterative approach achieves fast and reliable convergence to high-quality plans in multiple simulated scenarios involving a car-like robot, a quadrotor, and a medical steerable needle performing a liver biopsy procedure
Feynman-Kac Numerical Techniques for Stochastic Optimal Control
Three significant advancements are proposed for improving numerical methods in the solution of forward-backward stochastic differential equations (FBSDEs) appearing in the Feynman-Kac representation of the value function in stochastic optimal control (SOC) problems. First, we propose a novel characterization of FBSDE estimators as either on-policy or off-policy, highlighting the intuition for these techniques that the distribution over which value functions are approximated should, to some extent, match the distribution the policies generate.
Second, two novel numerical estimators are proposed for improving the accuracy of single-timestep updates. In the case of LQR problems, we demonstrate both in theory and in numerical simulation that our estimators result in near machine-precision level accuracy, in contrast to previously proposed methods that can potentially diverge on the same problems.
Third, we propose a new method for accelerating the global convergence of FBSDE methods. By the repeated use of the Girsanov change of probability measures, it is demonstrated how a McKean-Markov branched sampling method can be utilized for the forward integration pass, as long as the controlled drift terms are appropriately compensated in the backward integration pass. Subsequently, a numerical approximation of the value function is proposed by solving a series of function approximation problems backwards in time along the edges of a space-filling tree.Ph.D
The application of dynamic virtual prototyping to the development of control systems
A new method of developing control systems on the basis of dynamic virtual prototyping (DVP) has been proposed. The method facilitates the control system development process by means of (1) automating the transition from the conventional DVP of a plant to the dynamic model suitable for the control system design and (2) integrating the development process into the overall lifecycle.
The method comprises the three principal stages: (1) representing the plant by its DVP; (2) simulating the DVP and generating the data-based model of the plant; (3) designing the control system using the generated data-based model. Stages 1 and 2 are supported by DVP systems (e.g. IGRIP, LMS/VirtualLab, MSC SimOffice), stage 3 is accomplished by CACSD.
The proposed development method has been adapted to the class of plants that are linearizable, quasi-stationary, stable or stabilizable without using the analytical model and have lumped parameters. The specifics of applying the conventional methods of identification and control system design in the context of DVP have been revealed and analyzed.
The engineering applicability of the method has been proved by means of developing the control system for fine positioning of a gantry crane load.reviewe
- …