50 research outputs found

    Hitting Times in Markov Chains with Restart and their Application to Network Centrality

    Get PDF
    Motivated by applications in telecommunications, computer scienceand physics, we consider a discrete-time Markov process withrestart. At each step the process eitherwith a positive probability restarts from a given distribution, orwith the complementary probability continues according to a Markovtransition kernel. The main contribution of the present work is thatwe obtain an explicit expression for the expectation of the hittingtime (to a given target set) of the process with restart.The formula is convenient when considering the problem of optimizationof the expected hitting time with respect to the restart probability.We illustrate our results with two examplesin uncountable and countable state spaces andwith an application to network centrality

    Extreme occupation measures in Markov decision processes with a cemetery

    Full text link
    In this paper, we consider a Markov decision process (MDP) with a Borel state space X∪{Δ}\textbf{X}\cup\{\Delta\}, where Δ\Delta is an absorbing state (cemetery), and a Borel action space A\textbf{A}. We consider the space of finite occupation measures restricted on X×A\textbf{X}\times \textbf{A}, and the extreme points in it. It is possible that some strategies have infinite occupation measures. Nevertheless, we prove that every finite extreme occupation measure is generated by a deterministic stationary strategy. Then, for this MDP, we consider a constrained problem with total undiscounted criteria and JJ constraints, where the cost functions are nonnegative. By assumption, the strategies inducing infinite occupation measures are not optimal. Then, our second main result is that, under mild conditions, the solution to this constrained MDP is given by a mixture of no more than J+1J+1 occupation measures generated by deterministic stationary strategies

    On reducing a constrained gradual-impulsive control problem for a jump Markov model to a model with gradual control only

    Get PDF
    In this paper we consider a gradual-impulsive control problem for continuous-time Markov decision processes (CTMDPs) with total cost criteria and constraints. We develop a simple and useful method, which reduces the concerned problem to a standard CTMDP problem with gradual control only. This allows us to derive straightforwardly and under a minimal set of conditions the optimality results (sufficient classes of control policies, as well as the existence of stationary optimal policies) for the original constrained gradual-impulsive control problem

    On gradual-impulse control of continuous-time Markov decision processes with multiplicative cost

    Full text link
    In this paper, we consider the gradual-impulse control problem of continuous-time Markov decision processes, where the system performance is measured by the expectation of the exponential utility of the total cost. We prove, under very general conditions on the system primitives, the existence of a deterministic stationary optimal policy out of a more general class of policies. Policies that we consider allow multiple simultaneous impulses, randomized selection of impulses with random effects, relaxed gradual controls, and accumulation of jumps. After characterizing the value function using the optimality equation, we reduce the continuous-time gradual-impulse control problem to an equivalent simple discrete-time Markov decision process, whose action space is the union of the sets of gradual and impulsive actions

    On the equivalence of the integral and differential Bellman equations in impulse control problems

    Get PDF
    When solving optimal impulse control problems, one can use the dynamic programming approach in two different ways: at each time moment, one can make the decision whether to apply a particular type of impulse, leading to the instantaneous change of the state, or apply no impulses at all; or, otherwise, one can plan an impulse after a certain interval, so that the length of that interval is to be optimized along with the type of that impulse. The first method leads to the differential Bellman equation, while the second method leads to the integral Bellman equation. The target of the current article is to prove the equivalence of those Bellman equations. Firstly, we prove that, for the simple deterministic optimal stopping problem, the equations in the integral and differential form are equivalent under very mild conditions. Here, the impulse means that the uncontrolled process is stopped, i.e., sent to the so called cemetery. After that, the obtained result immediately implies the similar equivalence of the Bellman equations for other models of optimal impulse control. Those include abstract dynamical systems, controlled ordinary differential equations, piece-wise deterministic Markov processes and continuous-time Markov decision processes.publishe

    NOTE ON DISCOUNTED CONTINUOUS-TIME MARKOV DECISION PROCESSES WITH A LOWER BOUNDING FUNCTION

    Get PDF
    In this paper, we consider the discounted continuous-time Markov decision process (CTMDP) with a lower bounding function. In this model, the negative part of each cost rate is bounded by the drift function, say ww, whereas the positive part is allowed to be arbitrarily unbounded. Our focus is on the existence of a stationary optimal policy for the discounted CTMDP problems out of the more general class. Both constrained and unconstrained problems are considered. Our investigations are based on a useful transformation for nonhomogeneous Markov pure jump processes that has not yet been widely applied to the study of CTMDPs. This technique was not employed in previous literature, but it clarifies the roles of the imposed conditions in a rather transparent way. As a consequence, we withdraw and weaken several conditions commonly imposed in the literature

    On the structure of optimal solutions in a mathematical programming problem in a convex space

    Get PDF
    We consider an optimization problem in a convex space E with an affine objective function, subject to J affine constraints, where J is a given nonnegative integer. We apply the Feinberg-Shwartz lemma in finite dimensional convex analysis to show that there exists an optimal solution, which is in the form of a convex combination of no more than J+1 extreme points of E. The concerned problem does not seem to fit into the framework of standard convex optimization problems

    Infinite horizon optimal impulsive control with applications to Internet congestion control

    Get PDF
    International audienceWe investigate infinite horizon deterministic optimal control problems with both gradual and impulsive controls, where any finitely many impulses are allowed simultaneously. Both discounted and long run time average criteria are considered. We establish very general and at the same time natural conditions, under which the dynamic programming approach results in an optimal feedback policy. The established theoretical results are applied to the Internet congestion control, and by solving analytically and nontrivially the underlying optimal control problems, we obtain a simple threshold-based active queue management scheme, which takes into account the main parameters of the transmission control protocols, and improves the fairness among the connections in a given network
    corecore