5 research outputs found
Constrained Active Classification Using Partially Observable Markov Decision Processes
In this work, we study the problem of actively classifying the attributes of
dynamical systems characterized as a finite set of Markov decision process
(MDP) models. We are interested in finding strategies that actively interact
with the dynamical system and observe its reactions so that the attribute of
interest is classified efficiently with high confidence. We present a
decision-theoretic framework based on partially observable Markov decision
processes (POMDPs). The proposed framework relies on assigning a classification
belief (a probability distribution) to the attributes of interest. Given an
initial belief, confidence level over which a classification decision can be
made, a cost bound, safe belief sets, and a finite time horizon, we compute
POMDP strategies leading to classification decisions. We present two different
algorithms to compute such strategies. The first algorithm computes the optimal
strategy exactly by value iteration. To overcome the computational complexity
of computing the exact solutions, we propose a second algorithm is based on
adaptive sampling to approximate the optimal probability of reaching a
classification decision. We illustrate the proposed methodology using examples
from medical diagnosis and privacy-preserving advertising.Comment: arXiv admin note: substantial text overlap with arXiv:1810.0009
Controllability and Stabilization of Kolmogorov Forward Equations for Robotic Swarms
abstract: Numerous works have addressed the control of multi-robot systems for coverage, mapping, navigation, and task allocation problems. In addition to classical microscopic approaches to multi-robot problems, which model the actions and decisions of individual robots, lately, there has been a focus on macroscopic or Eulerian approaches. In these approaches, the population of robots is represented as a continuum that evolves according to a mean-field model, which is directly designed such that the corresponding robot control policies produce target collective behaviours.
This dissertation presents a control-theoretic analysis of three types of mean-field models proposed in the literature for modelling and control of large-scale multi-agent systems, including robotic swarms. These mean-field models are Kolmogorov forward equations of stochastic processes, and their analysis is motivated by the fact that as the number of agents tends to infinity, the empirical measure associated with the agents converges to the solution of these models. Hence, the problem of transporting a swarm of agents from one distribution to another can be posed as a control problem for the forward equation of the process that determines the time evolution of the swarm density.
First, this thesis considers the case in which the agents' states evolve on a finite state space according to a continuous-time Markov chain (CTMC), and the forward equation is an ordinary differential equation (ODE). Defining the agents' task transition rates as the control parameters, the finite-time controllability, asymptotic controllability, and stabilization of the forward equation are investigated. Second, the controllability and stabilization problem for systems of advection-diffusion-reaction partial differential equations (PDEs) is studied in the case where the control parameters include the agents' velocity as well as transition rates. Third, this thesis considers a controllability and optimal control problem for the forward equation in the more general case where the agent dynamics are given by a nonlinear discrete-time control system. Beyond these theoretical results, this thesis also considers numerical optimal transport for control-affine systems. It is shown that finite-volume approximations of the associated PDEs lead to well-posed transport problems on graphs as long as the control system is controllable everywhere.Dissertation/ThesisDoctoral Dissertation Mechanical Engineering 201