37,753 research outputs found

    Higher Order Derivatives in Costa's Entropy Power Inequality

    Full text link
    Let XX be an arbitrary continuous random variable and ZZ be an independent Gaussian random variable with zero mean and unit variance. For tΒ >Β 0t~>~0, Costa proved that e2h(X+tZ)e^{2h(X+\sqrt{t}Z)} is concave in tt, where the proof hinged on the first and second order derivatives of h(X+tZ)h(X+\sqrt{t}Z). Specifically, these two derivatives are signed, i.e., βˆ‚βˆ‚th(X+tZ)β‰₯0\frac{\partial}{\partial t}h(X+\sqrt{t}Z) \geq 0 and βˆ‚2βˆ‚t2h(X+tZ)≀0\frac{\partial^2}{\partial t^2}h(X+\sqrt{t}Z) \leq 0. In this paper, we show that the third order derivative of h(X+tZ)h(X+\sqrt{t}Z) is nonnegative, which implies that the Fisher information J(X+tZ)J(X+\sqrt{t}Z) is convex in tt. We further show that the fourth order derivative of h(X+tZ)h(X+\sqrt{t}Z) is nonpositive. Following the first four derivatives, we make two conjectures on h(X+tZ)h(X+\sqrt{t}Z): the first is that βˆ‚nβˆ‚tnh(X+tZ)\frac{\partial^n}{\partial t^n} h(X+\sqrt{t}Z) is nonnegative in tt if nn is odd, and nonpositive otherwise; the second is that log⁑J(X+tZ)\log J(X+\sqrt{t}Z) is convex in tt. The first conjecture can be rephrased in the context of completely monotone functions: J(X+tZ)J(X+\sqrt{t}Z) is completely monotone in tt. The history of the first conjecture may date back to a problem in mathematical physics studied by McKean in 1966. Apart from these results, we provide a geometrical interpretation to the covariance-preserving transformation and study the concavity of h(tX+1βˆ’tZ)h(\sqrt{t}X+\sqrt{1-t}Z), revealing its connection with Costa's EPI.Comment: Second version submitted. https://sites.google.com/site/chengfancuhk

    Asymmetry Helps: Eigenvalue and Eigenvector Analyses of Asymmetrically Perturbed Low-Rank Matrices

    Full text link
    This paper is concerned with the interplay between statistical asymmetry and spectral methods. Suppose we are interested in estimating a rank-1 and symmetric matrix Mβ‹†βˆˆRnΓ—n\mathbf{M}^{\star}\in \mathbb{R}^{n\times n}, yet only a randomly perturbed version M\mathbf{M} is observed. The noise matrix Mβˆ’M⋆\mathbf{M}-\mathbf{M}^{\star} is composed of zero-mean independent (but not necessarily homoscedastic) entries and is, therefore, not symmetric in general. This might arise, for example, when we have two independent samples for each entry of M⋆\mathbf{M}^{\star} and arrange them into an {\em asymmetric} data matrix M\mathbf{M}. The aim is to estimate the leading eigenvalue and eigenvector of M⋆\mathbf{M}^{\star}. We demonstrate that the leading eigenvalue of the data matrix M\mathbf{M} can be O(n)O(\sqrt{n}) times more accurate --- up to some log factor --- than its (unadjusted) leading singular value in eigenvalue estimation. Further, the perturbation of any linear form of the leading eigenvector of M\mathbf{M} --- say, entrywise eigenvector perturbation --- is provably well-controlled. This eigen-decomposition approach is fully adaptive to heteroscedasticity of noise without the need of careful bias correction or any prior knowledge about the noise variance. We also provide partial theory for the more general rank-rr case. The takeaway message is this: arranging the data samples in an asymmetric manner and performing eigen-decomposition could sometimes be beneficial.Comment: accepted to Annals of Statistics, 2020. 37 page
    • …
    corecore