13,651 research outputs found
Maximizing Entropy over Markov Processes
International audienceThe channel capacity of a deterministic system with confidential data is an upper bound on the amount of bits of data an attacker can learn from the system. We encode all possible attacks to a system using a probabilistic specification, an Interval Markov Chain. Then the channel capacity computation reduces to finding a model of a specification with highest entropy. Entropy maximization for probabilistic process specifications has not been studied before, even though it is well known in Bayesian inference for discrete distributions. We give a characterization of global entropy of a process as a reward function, a polynomial algorithm to verify the existence of an system maximizing entropy among those respecting a specification, a procedure for the maximization of reward functions over Interval Markov Chains and its application to synthesize an implementation maximizing entropy. We show how to use Interval Markov Chains to model abstractions of deterministic systems with confidential data, and use the above results to compute their channel capacity. These results are a foundation for ongoing work on computing channel capacity for abstractions of programs derived from code
Markov processes follow from the principle of Maximum Caliber
Markov models are widely used to describe processes of stochastic dynamics.
Here, we show that Markov models are a natural consequence of the dynamical
principle of Maximum Caliber. First, we show that when there are different
possible dynamical trajectories in a time-homogeneous process, then the only
type of process that maximizes the path entropy, for any given singlet
statistics, is a sequence of identical, independently distributed (i.i.d.)
random variables, which is the simplest Markov process. If the data is in the
form of sequentially pairwise statistics, then maximizing the caliber dictates
that the process is Markovian with a uniform initial distribution. Furthermore,
if an initial non-uniform dynamical distribution is known, or multiple
trajectories are conditioned on an initial state, then the Markov process is
still the only one that maximizes the caliber. Second, given a model, MaxCal
can be used to compute the parameters of that model. We show that this
procedure is equivalent to the maximum-likelihood method of inference in the
theory of statistics.Comment: 4 page
Estimating ensemble flows on a hidden Markov chain
We propose a new framework to estimate the evolution of an ensemble of
indistinguishable agents on a hidden Markov chain using only aggregate output
data. This work can be viewed as an extension of the recent developments in
optimal mass transport and Schr\"odinger bridges to the finite state space
hidden Markov chain setting. The flow of the ensemble is estimated by solving a
maximum likelihood problem, which has a convex formulation at the
infinite-particle limit, and we develop a fast numerical algorithm for it. We
illustrate in two numerical examples how this framework can be used to track
the flow of identical and indistinguishable dynamical systems.Comment: 8 pages, 4 figure
Write Channel Model for Bit-Patterned Media Recording
We propose a new write channel model for bit-patterned media recording that
reflects the data dependence of write synchronization errors. It is shown that
this model accommodates both substitution-like errors and insertion-deletion
errors whose statistics are determined by an underlying channel state process.
We study information theoretic properties of the write channel model, including
the capacity, symmetric information rate, Markov-1 rate and the zero-error
capacity.Comment: 11 pages, 12 figures, journa
- …