50 research outputs found
Hitting Times in Markov Chains with Restart and their Application to Network Centrality
Motivated by applications in telecommunications, computer scienceand physics,
we consider a discrete-time Markov process withrestart. At each step the
process eitherwith a positive probability restarts from a given distribution,
orwith the complementary probability continues according to a Markovtransition
kernel. The main contribution of the present work is thatwe obtain an explicit
expression for the expectation of the hittingtime (to a given target set) of
the process with restart.The formula is convenient when considering the problem
of optimizationof the expected hitting time with respect to the restart
probability.We illustrate our results with two examplesin uncountable and
countable state spaces andwith an application to network centrality
Extreme occupation measures in Markov decision processes with a cemetery
In this paper, we consider a Markov decision process (MDP) with a Borel state
space , where is an absorbing state
(cemetery), and a Borel action space . We consider the space of
finite occupation measures restricted on , and the
extreme points in it. It is possible that some strategies have infinite
occupation measures. Nevertheless, we prove that every finite extreme
occupation measure is generated by a deterministic stationary strategy. Then,
for this MDP, we consider a constrained problem with total undiscounted
criteria and constraints, where the cost functions are nonnegative. By
assumption, the strategies inducing infinite occupation measures are not
optimal. Then, our second main result is that, under mild conditions, the
solution to this constrained MDP is given by a mixture of no more than
occupation measures generated by deterministic stationary strategies
On reducing a constrained gradual-impulsive control problem for a jump Markov model to a model with gradual control only
In this paper we consider a gradual-impulsive control problem for continuous-time Markov decision processes (CTMDPs) with total cost criteria and constraints. We develop a simple and useful method, which reduces the concerned problem to a standard CTMDP problem with gradual control only. This allows us to derive straightforwardly and under a minimal set of conditions the optimality results (sufficient classes of control policies, as well as the existence of stationary optimal policies) for the original constrained gradual-impulsive control problem
On gradual-impulse control of continuous-time Markov decision processes with multiplicative cost
In this paper, we consider the gradual-impulse control problem of
continuous-time Markov decision processes, where the system performance is
measured by the expectation of the exponential utility of the total cost. We
prove, under very general conditions on the system primitives, the existence of
a deterministic stationary optimal policy out of a more general class of
policies. Policies that we consider allow multiple simultaneous impulses,
randomized selection of impulses with random effects, relaxed gradual controls,
and accumulation of jumps. After characterizing the value function using the
optimality equation, we reduce the continuous-time gradual-impulse control
problem to an equivalent simple discrete-time Markov decision process, whose
action space is the union of the sets of gradual and impulsive actions
On the equivalence of the integral and differential Bellman equations in impulse control problems
When solving optimal impulse control problems, one can use the dynamic programming
approach in two different ways: at each time moment, one can make the decision whether
to apply a particular type of impulse, leading to the instantaneous change of the state, or
apply no impulses at all; or, otherwise, one can plan an impulse after a certain interval, so
that the length of that interval is to be optimized along with the type of that impulse. The
first method leads to the differential Bellman equation, while the second method leads to the
integral Bellman equation. The target of the current article is to prove the equivalence of
those Bellman equations. Firstly, we prove that, for the simple deterministic optimal stopping
problem, the equations in the integral and differential form are equivalent under very mild
conditions. Here, the impulse means that the uncontrolled process is stopped, i.e., sent to the
so called cemetery. After that, the obtained result immediately implies the similar equivalence
of the Bellman equations for other models of optimal impulse control. Those include abstract
dynamical systems, controlled ordinary differential equations, piece-wise deterministic Markov
processes and continuous-time Markov decision processes.publishe
NOTE ON DISCOUNTED CONTINUOUS-TIME MARKOV DECISION PROCESSES WITH A LOWER BOUNDING FUNCTION
In this paper, we consider the discounted continuous-time Markov decision
process (CTMDP) with a lower bounding function. In this model, the negative
part of each cost rate is bounded by the drift function, say , whereas the
positive part is allowed to be arbitrarily unbounded. Our focus is on the
existence of a stationary optimal policy for the discounted CTMDP problems out
of the more general class. Both constrained and unconstrained problems are
considered. Our investigations are based on a useful transformation for
nonhomogeneous Markov pure jump processes that has not yet been widely applied
to the study of CTMDPs. This technique was not employed in previous literature,
but it clarifies the roles of the imposed conditions in a rather transparent
way. As a consequence, we withdraw and weaken several conditions commonly
imposed in the literature
On the structure of optimal solutions in a mathematical programming problem in a convex space
We consider an optimization problem in a convex space E with an affine objective function, subject to J affine constraints, where J is a given nonnegative integer. We apply the Feinberg-Shwartz lemma in finite dimensional convex analysis to show that there exists an optimal solution, which is in the form of a convex combination of no more than J+1 extreme points of E. The concerned problem does not seem to fit into the framework of standard convex optimization problems
Infinite horizon optimal impulsive control with applications to Internet congestion control
International audienceWe investigate infinite horizon deterministic optimal control problems with both gradual and impulsive controls, where any finitely many impulses are allowed simultaneously. Both discounted and long run time average criteria are considered. We establish very general and at the same time natural conditions, under which the dynamic programming approach results in an optimal feedback policy. The established theoretical results are applied to the Internet congestion control, and by solving analytically and nontrivially the underlying optimal control problems, we obtain a simple threshold-based active queue management scheme, which takes into account the main parameters of the transmission control protocols, and improves the fairness among the connections in a given network