107 research outputs found

    An Exploration of the Role of Principal Inertia Components in Information Theory

    Full text link
    The principal inertia components of the joint distribution of two random variables XX and YY are inherently connected to how an observation of YY is statistically related to a hidden variable XX. In this paper, we explore this connection within an information theoretic framework. We show that, under certain symmetry conditions, the principal inertia components play an important role in estimating one-bit functions of XX, namely f(X)f(X), given an observation of YY. In particular, the principal inertia components bear an interpretation as filter coefficients in the linear transformation of pf(X)Xp_{f(X)|X} into pf(X)Yp_{f(X)|Y}. This interpretation naturally leads to the conjecture that the mutual information between f(X)f(X) and YY is maximized when all the principal inertia components have equal value. We also study the role of the principal inertia components in the Markov chain BXYB^B\rightarrow X\rightarrow Y\rightarrow \widehat{B}, where BB and B^\widehat{B} are binary random variables. We illustrate our results for the setting where XX and YY are binary strings and YY is the result of sending XX through an additive noise binary channel.Comment: Submitted to the 2014 IEEE Information Theory Workshop (ITW

    Privacy Against Statistical Inference

    Full text link
    We propose a general statistical inference framework to capture the privacy threat incurred by a user that releases data to a passive but curious adversary, given utility constraints. We show that applying this general framework to the setting where the adversary uses the self-information cost function naturally leads to a non-asymptotic information-theoretic approach for characterizing the best achievable privacy subject to utility constraints. Based on these results we introduce two privacy metrics, namely average information leakage and maximum information leakage. We prove that under both metrics the resulting design problem of finding the optimal mapping from the user's data to a privacy-preserving output can be cast as a modified rate-distortion problem which, in turn, can be formulated as a convex program. Finally, we compare our framework with differential privacy.Comment: Allerton 2012, 8 page
    corecore