45,705 research outputs found
Tight Bounds on the R\'enyi Entropy via Majorization with Applications to Guessing and Compression
This paper provides tight bounds on the R\'enyi entropy of a function of a
discrete random variable with a finite number of possible values, where the
considered function is not one-to-one. To that end, a tight lower bound on the
R\'enyi entropy of a discrete random variable with a finite support is derived
as a function of the size of the support, and the ratio of the maximal to
minimal probability masses. This work was inspired by the recently published
paper by Cicalese et al., which is focused on the Shannon entropy, and it
strengthens and generalizes the results of that paper to R\'enyi entropies of
arbitrary positive orders. In view of these generalized bounds and the works by
Arikan and Campbell, non-asymptotic bounds are derived for guessing moments and
lossless data compression of discrete memoryless sources.Comment: The paper was published in the Entropy journal (special issue on
Probabilistic Methods in Information Theory, Hypothesis Testing, and Coding),
vol. 20, no. 12, paper no. 896, November 22, 2018. Online available at
https://www.mdpi.com/1099-4300/20/12/89
Concentration of Measure Inequalities in Information Theory, Communications and Coding (Second Edition)
During the last two decades, concentration inequalities have been the subject
of exciting developments in various areas, including convex geometry,
functional analysis, statistical physics, high-dimensional statistics, pure and
applied probability theory, information theory, theoretical computer science,
and learning theory. This monograph focuses on some of the key modern
mathematical tools that are used for the derivation of concentration
inequalities, on their links to information theory, and on their various
applications to communications and coding. In addition to being a survey, this
monograph also includes various new recent results derived by the authors. The
first part of the monograph introduces classical concentration inequalities for
martingales, as well as some recent refinements and extensions. The power and
versatility of the martingale approach is exemplified in the context of codes
defined on graphs and iterative decoding algorithms, as well as codes for
wireless communication. The second part of the monograph introduces the entropy
method, an information-theoretic technique for deriving concentration
inequalities. The basic ingredients of the entropy method are discussed first
in the context of logarithmic Sobolev inequalities, which underlie the
so-called functional approach to concentration of measure, and then from a
complementary information-theoretic viewpoint based on transportation-cost
inequalities and probability in metric spaces. Some representative results on
concentration for dependent random variables are briefly summarized, with
emphasis on their connections to the entropy method. Finally, we discuss
several applications of the entropy method to problems in communications and
coding, including strong converses, empirical distributions of good channel
codes, and an information-theoretic converse for concentration of measure.Comment: Foundations and Trends in Communications and Information Theory, vol.
10, no 1-2, pp. 1-248, 2013. Second edition was published in October 2014.
ISBN to printed book: 978-1-60198-906-
On the Entropy of Sums of Bernoulli Random Variables via the Chen-Stein Method
This paper considers the entropy of the sum of (possibly dependent and
non-identically distributed) Bernoulli random variables. Upper bounds on the
error that follows from an approximation of this entropy by the entropy of a
Poisson random variable with the same mean are derived. The derivation of these
bounds combines elements of information theory with the Chen-Stein method for
Poisson approximation. The resulting bounds are easy to compute, and their
applicability is exemplified. This conference paper presents in part the first
half of the paper entitled "An information-theoretic perspective of the Poisson
approximation via the Chen-Stein method" (see:arxiv:1206.6811). A
generalization of the bounds that considers the accuracy of the Poisson
approximation for the entropy of a sum of non-negative, integer-valued and
bounded random variables is introduced in the full paper. It also derives lower
bounds on the total variation distance, relative entropy and other measures
that are not considered in this conference paper.Comment: A conference paper of 5 pages that appears in the Proceedings of the
2012 IEEE International Workshop on Information Theory (ITW 2012), pp.
542--546, Lausanne, Switzerland, September 201
Bounds on Information Combining With Quantum Side Information
"Bounds on information combining" are entropic inequalities that determine
how the information (entropy) of a set of random variables can change when
these are combined in certain prescribed ways. Such bounds play an important
role in classical information theory, particularly in coding and Shannon
theory; entropy power inequalities are special instances of them. The arguably
most elementary kind of information combining is the addition of two binary
random variables (a CNOT gate), and the resulting quantities play an important
role in Belief propagation and Polar coding. We investigate this problem in the
setting where quantum side information is available, which has been recognized
as a hard setting for entropy power inequalities.
Our main technical result is a non-trivial, and close to optimal, lower bound
on the combined entropy, which can be seen as an almost optimal "quantum Mrs.
Gerber's Lemma". Our proof uses three main ingredients: (1) a new bound on the
concavity of von Neumann entropy, which is tight in the regime of low pairwise
state fidelities; (2) the quantitative improvement of strong subadditivity due
to Fawzi-Renner, in which we manage to handle the minimization over recovery
maps; (3) recent duality results on classical-quantum-channels due to Renes et
al. We furthermore present conjectures on the optimal lower and upper bounds
under quantum side information, supported by interesting analytical
observations and strong numerical evidence.
We finally apply our bounds to Polar coding for binary-input
classical-quantum channels, and show the following three results: (A) Even
non-stationary channels polarize under the polar transform. (B) The blocklength
required to approach the symmetric capacity scales at most sub-exponentially in
the gap to capacity. (C) Under the aforementioned lower bound conjecture, a
blocklength polynomial in the gap suffices.Comment: 23 pages, 6 figures; v2: small correction
- …