59,654 research outputs found

    Data-driven Inverse Optimization with Imperfect Information

    Full text link
    In data-driven inverse optimization an observer aims to learn the preferences of an agent who solves a parametric optimization problem depending on an exogenous signal. Thus, the observer seeks the agent's objective function that best explains a historical sequence of signals and corresponding optimal actions. We focus here on situations where the observer has imperfect information, that is, where the agent's true objective function is not contained in the search space of candidate objectives, where the agent suffers from bounded rationality or implementation errors, or where the observed signal-response pairs are corrupted by measurement noise. We formalize this inverse optimization problem as a distributionally robust program minimizing the worst-case risk that the {\em predicted} decision ({\em i.e.}, the decision implied by a particular candidate objective) differs from the agent's {\em actual} response to a random signal. We show that our framework offers rigorous out-of-sample guarantees for different loss functions used to measure prediction errors and that the emerging inverse optimization problems can be exactly reformulated as (or safely approximated by) tractable convex programs when a new suboptimality loss function is used. We show through extensive numerical tests that the proposed distributionally robust approach to inverse optimization attains often better out-of-sample performance than the state-of-the-art approaches

    Multilevel Particle Filters for L\'evy-driven stochastic differential equations

    Full text link
    We develop algorithms for computing expectations of the laws of models associated to stochastic differential equations (SDEs) driven by pure L\'evy processes. We consider filtering such processes and well as pricing of path dependent options. We propose a multilevel particle filter (MLPF) to address the computational issues involved in solving these continuum problems. We show via numerical simulations and theoretical results that under suitable assumptions of the discretization of the underlying driving L\'evy proccess, our proposed method achieves optimal convergence rates. The cost to obtain MSE O(ϵ2)O(\epsilon^2) scales like O(ϵ−2)O(\epsilon^{-2}) for our method, as compared with the standard particle filter O(ϵ−3)O(\epsilon^{-3})

    Feller Processes: The Next Generation in Modeling. Brownian Motion, L\'evy Processes and Beyond

    Get PDF
    We present a simple construction method for Feller processes and a framework for the generation of sample paths of Feller processes. The construction is based on state space dependent mixing of L\'evy processes. Brownian Motion is one of the most frequently used continuous time Markov processes in applications. In recent years also L\'evy processes, of which Brownian Motion is a special case, have become increasingly popular. L\'evy processes are spatially homogeneous, but empirical data often suggest the use of spatially inhomogeneous processes. Thus it seems necessary to go to the next level of generalization: Feller processes. These include L\'evy processes and in particular Brownian motion as special cases but allow spatial inhomogeneities. Many properties of Feller processes are known, but proving the very existence is, in general, very technical. Moreover, an applicable framework for the generation of sample paths of a Feller process was missing. We explain, with practitioners in mind, how to overcome both of these obstacles. In particular our simulation technique allows to apply Monte Carlo methods to Feller processes.Comment: 22 pages, including 4 figures and 8 pages of source code for the generation of sample paths of Feller processe
    • …
    corecore