14 research outputs found
Data driven consistency (working title)
We are motivated by applications that need rich model classes to represent
them. Examples of rich model classes include distributions over large,
countably infinite supports, slow mixing Markov processes, etc. But such rich
classes may be too complex to admit estimators that converge to the truth with
convergence rates that can be uniformly bounded over the entire model class as
the sample size increases (uniform consistency). However, these rich classes
may still allow for estimators with pointwise guarantees whose performance can
be bounded in a model dependent way. The pointwise angle of course has the
drawback that the estimator performance is a function of the very unknown model
that is being estimated, and is therefore unknown. Therefore, even if the
estimator is consistent, how well it is doing may not be clear no matter what
the sample size is. Departing from the dichotomy of uniform and pointwise
consistency, a new analysis framework is explored by characterizing rich model
classes that may only admit pointwise guarantees, yet all the information about
the model needed to guage estimator accuracy can be inferred from the sample at
hand. To retain focus, we analyze the universal compression problem in this
data driven pointwise consistency framework.Comment: Working paper. Please email authors for the current versio
On the Compression of Unknown Sources
Ph.D. Thesis. University of Hawaiʻi at Mānoa 2018
Asymptotic Estimates in Information Theory with Non-Vanishing Error Probabilities
This monograph presents a unified treatment of single- and multi-user
problems in Shannon's information theory where we depart from the requirement
that the error probability decays asymptotically in the blocklength. Instead,
the error probabilities for various problems are bounded above by a
non-vanishing constant and the spotlight is shone on achievable coding rates as
functions of the growing blocklengths. This represents the study of asymptotic
estimates with non-vanishing error probabilities.
In Part I, after reviewing the fundamentals of information theory, we discuss
Strassen's seminal result for binary hypothesis testing where the type-I error
probability is non-vanishing and the rate of decay of the type-II error
probability with growing number of independent observations is characterized.
In Part II, we use this basic hypothesis testing result to develop second- and
sometimes, even third-order asymptotic expansions for point-to-point
communication. Finally in Part III, we consider network information theory
problems for which the second-order asymptotics are known. These problems
include some classes of channels with random state, the multiple-encoder
distributed lossless source coding (Slepian-Wolf) problem and special cases of
the Gaussian interference and multiple-access channels. Finally, we discuss
avenues for further research.Comment: Further comments welcom
Characterizing the Asymptotic Per-Symbol Redundancy of Memoryless Sources over Countable Alphabets in Terms of Single-Letter Marginals
The minimum expected number of bits needed to describe a random variable is its entropy, assuming knowledge of the distribution of the random variable. On the other hand, universal compression describes data supposing that the underlying distribution is unknown, but that it belongs to a known set Ρ of distributions. However, since universal descriptions are not matched exactly to the underlying distribution, the number of bits they use on average is higher, and the excess over the entropy used is the redundancy. In this paper, we study the redundancy incurred by the universal description of strings of positive integers (Z+), the strings being generated independently and identically distributed (i.i.d.) according an unknown distribution over Z+ in a known collection P. We first show that if describing a single symbol incurs finite redundancy, then P is tight, but that the converse does not always hold. If a single symbol can be described with finite worst-case regret (a more stringent formulation than redundancy above), then it is known that describing length n i.i.d. strings only incurs vanishing (to zero) redundancy per symbol as n increases. On the contrary, we show it is possible that the description of a single symbol from an unknown distribution of P incurs finite redundancy, yet the description of length n i.i.d. strings incurs a constant (> 0) redundancy per symbol encoded. We then show a sufficient condition on single-letter marginals, such that length n i.i.d. samples will incur vanishing redundancy per symbol encoded
Markov and Semi-markov Chains, Processes, Systems and Emerging Related Fields
This book covers a broad range of research results in the field of Markov and Semi-Markov chains, processes, systems and related emerging fields. The authors of the included research papers are well-known researchers in their field. The book presents the state-of-the-art and ideas for further research for theorists in the fields. Nonetheless, it also provides straightforwardly applicable results for diverse areas of practitioners
LIPIcs, Volume 261, ICALP 2023, Complete Volume
LIPIcs, Volume 261, ICALP 2023, Complete Volum