44,663 research outputs found
Consistency of Feature Markov Processes
We are studying long term sequence prediction (forecasting). We approach this
by investigating criteria for choosing a compact useful state representation.
The state is supposed to summarize useful information from the history. We want
a method that is asymptotically consistent in the sense it will provably
eventually only choose between alternatives that satisfy an optimality property
related to the used criterion. We extend our work to the case where there is
side information that one can take advantage of and, furthermore, we briefly
discuss the active setting where an agent takes actions to achieve desirable
outcomes.Comment: 16 LaTeX page
Maximum likelihood estimation for a general mixture of Markov jump processes
We estimate a general mixture of Markov jump processes. The key novel feature
of the proposed mixture is that the transition intensity matrices of the Markov
processes comprising the mixture are entirely unconstrained. The Markov
processes are mixed with distributions that depend on the initial state of the
mixture process. The new mixture is estimated from its continuously observed
realizations using the EM algorithm, which provides the maximum likelihood (ML)
estimates of the mixture's parameters. To obtain the standard errors of the
estimates of the mixture's parameters, we employ Louis' (1982) general formula
for the observed Fisher information matrix. We also derive the asymptotic
properties of the ML estimators. Simulation study verifies the estimates'
accuracy and confirms the consistency and asymptotic normality of the
estimators. The developed methods are applied to a medical dataset, for which
the likelihood ratio test rejects the constrained mixture in favor of the
proposed unconstrained one. This application exemplifies the usefulness of a
new unconstrained mixture for identification and characterization of
homogeneous subpopulations in a heterogeneous population.Comment: 21 pages, 1 figur
Adaptive Continuous time Markov Chain Approximation Model to General Jump-Diffusions
We propose a non-equidistant Q rate matrix formula and an adaptive numerical algorithm for a continuous time Markov chain to approximate jump-diffusions with affine or non-affine functional specifications. Our approach also accommodates state-dependent jump intensity and jump distribution, a flexibility that is very hard to achieve with other numerical methods. The Kologorov-Smirnov test shows that the proposed Markov chain transition density converges to the one given by the likelihood expansion formula as in Ait-Sahalia (2008). We provide numerical examples for European stock option pricing in Black and Scholes (1973), Merton (1976) and Kou
(2002)
Robot introspection through learned hidden Markov models
In this paper we describe a machine learning approach for acquiring a model of a robot behaviour from raw sensor data. We are interested in automating the acquisition of behavioural models to provide a robot with an introspective capability. We assume that the behaviour of a robot in achieving a task can be modelled as a finite stochastic state transition system. Beginning with data recorded by a robot in the execution of a task, we use unsupervised learning techniques to estimate a hidden Markov model (HMM) that can be used both for predicting and explaining the behaviour of the robot in subsequent executions of the task. We demonstrate that it is feasible to automate the entire process of learning a high quality HMM from the data recorded by the robot during execution of its task.The learned HMM can be used both for monitoring and controlling the behaviour of the robot. The ultimate purpose of our work is to learn models for the full set of tasks associated with a given problem domain, and to integrate these models with a generative task planner. We want to show that these models can be used successfully in controlling the execution of a plan. However, this paper does not develop the planning and control aspects of our work, focussing instead on the learning methodology and the evaluation of a learned model. The essential property of the models we seek to construct is that the most probable trajectory through a model, given the observations made by the robot, accurately diagnoses, or explains, the behaviour that the robot actually performed when making these observations. In the work reported here we consider a navigation task. We explain the learning process, the experimental setup and the structure of the resulting learned behavioural models. We then evaluate the extent to which explanations proposed by the learned models accord with a human observer's interpretation of the behaviour exhibited by the robot in its execution of the task
- …