research

Optimal control of continuous-time Markov chains with noise-free observation

Abstract

We consider an infinite horizon optimal control problem for a continuous-time Markov chain XX in a finite set II with noise-free partial observation. The observation process is defined as Yt=h(Xt)Y_t = h(X_t), t≥0t \geq 0, where hh is a given map defined on II. The observation is noise-free in the sense that the only source of randomness is the process XX itself. The aim is to minimize a discounted cost functional and study the associated value function VV. After transforming the control problem with partial observation into one with complete observation (the separated problem) using filtering equations, we provide a link between the value function vv associated to the latter control problem and the original value function VV. Then, we present two different characterizations of vv (and indirectly of VV): on one hand as the unique fixed point of a suitably defined contraction mapping and on the other hand as the unique constrained viscosity solution (in the sense of Soner) of a HJB integro-differential equation. Under suitable assumptions, we finally prove the existence of an optimal control

    Similar works