17,639 research outputs found

    Approximate Solutions To Constrained Risk-Sensitive Markov Decision Processes

    Full text link
    This paper considers the problem of finding near-optimal Markovian randomized (MR) policies for finite-state-action, infinite-horizon, constrained risk-sensitive Markov decision processes (CRSMDPs). Constraints are in the form of standard expected discounted cost functions as well as expected risk-sensitive discounted cost functions over finite and infinite horizons. The main contribution is to show that the problem possesses a solution if it is feasible, and to provide two methods for finding an approximate solution in the form of an ultimately stationary (US) MR policy. The latter is achieved through two approximating finite-horizon CRSMDPs which are constructed from the original CRSMDP by time-truncating the original objective and constraint cost functions, and suitably perturbing the constraint upper bounds. The first approximation gives a US policy which is ϵ\epsilon-optimal and feasible for the original problem, while the second approximation gives a near-optimal US policy whose violation of the original constraints is bounded above by a specified ϵ\epsilon. A key step in the proofs is an appropriate choice of a metric that makes the set of infinite-horizon MR policies and the feasible regions of the three CRSMDPs compact, and the objective and constraint functions continuous. A linear-programming-based formulation for solving the approximating finite-horizon CRSMDPs is also given.Comment: 38 page

    A method for pricing American options using semi-infinite linear programming

    Full text link
    We introduce a new approach for the numerical pricing of American options. The main idea is to choose a finite number of suitable excessive functions (randomly) and to find the smallest majorant of the gain function in the span of these functions. The resulting problem is a linear semi-infinite programming problem, that can be solved using standard algorithms. This leads to good upper bounds for the original problem. For our algorithms no discretization of space and time and no simulation is necessary. Furthermore it is applicable even for high-dimensional problems. The algorithm provides an approximation of the value not only for one starting point, but for the complete value function on the continuation set, so that the optimal exercise region and e.g. the Greeks can be calculated. We apply the algorithm to (one- and) multidimensional diffusions and to L\'evy processes, and show it to be fast and accurate

    Adaptive Horizon Model Predictive Control and Al'brekht's Method

    Get PDF
    A standard way of finding a feedback law that stabilizes a control system to an operating point is to recast the problem as an infinite horizon optimal control problem. If the optimal cost and the optmal feedback can be found on a large domain around the operating point then a Lyapunov argument can be used to verify the asymptotic stability of the closed loop dynamics. The problem with this approach is that is usually very difficult to find the optimal cost and the optmal feedback on a large domain for nonlinear problems with or without constraints. Hence the increasing interest in Model Predictive Control (MPC). In standard MPC a finite horizon optimal control problem is solved in real time but just at the current state, the first control action is implimented, the system evolves one time step and the process is repeated. A terminal cost and terminal feedback found by Al'brekht's methoddefined in a neighborhood of the operating point is used to shorten the horizon and thereby make the nonlinear programs easier to solve because they have less decision variables. Adaptive Horizon Model Predictive Control (AHMPC) is a scheme for varying the horizon length of Model Predictive Control (MPC) as needed. Its goal is to achieve stabilization with horizons as small as possible so that MPC methods can be used on faster and/or more complicated dynamic processes.Comment: arXiv admin note: text overlap with arXiv:1602.0861

    Representation of homothetic forward performance processes in stochastic factor models via ergodic and infinite horizon BSDE

    Get PDF
    In an incomplete market, with incompleteness stemming from stochastic factors imperfectly correlated with the underlying stocks, we derive representations of homothetic (power, exponential and logarithmic) forward performance processes in factor-form using ergodic BSDE. We also develop a connection between the forward processes and infinite horizon BSDE, and, moreover, with risk-sensitive optimization. In addition, we develop a connection, for large time horizons, with a family of classical homothetic value function processes with random endowments.Comment: 34 page

    Analysis of unconstrained nonlinear MPC schemes with time varying control horizon

    Get PDF
    For discrete time nonlinear systems satisfying an exponential or finite time controllability assumption, we present an analytical formula for a suboptimality estimate for model predictive control schemes without stabilizing terminal constraints. Based on our formula, we perform a detailed analysis of the impact of the optimization horizon and the possibly time varying control horizon on stability and performance of the closed loop

    A receding horizon generalization of pointwise min-norm controllers

    Get PDF
    Control Lyapunov functions (CLFs) are used in conjunction with receding horizon control to develop a new class of receding horizon control schemes. In the process, strong connections between the seemingly disparate approaches are revealed, leading to a unified picture that ties together the notions of pointwise min-norm, receding horizon, and optimal control. This framework is used to develop a CLF based receding horizon scheme, of which a special case provides an appropriate extension of Sontag's formula. The scheme is first presented as an idealized continuous-time receding horizon control law. The issue of implementation under discrete-time sampling is then discussed as a modification. These schemes are shown to possess a number of desirable theoretical and implementation properties. An example is provided, demonstrating their application to a nonlinear control problem. Finally, stronger connections to both optimal and pointwise min-norm control are proved
    corecore