3 research outputs found

    Missing gg-mass: Investigating the Missing Parts of Distributions

    Full text link
    Estimating the underlying distribution from \textit{iid} samples is a classical and important problem in statistics. When the alphabet size is large compared to number of samples, a portion of the distribution is highly likely to be unobserved or sparsely observed. The missing mass, defined as the sum of probabilities Pr(x)\text{Pr}(x) over the missing letters xx, and the Good-Turing estimator for missing mass have been important tools in large-alphabet distribution estimation. In this article, given a positive function gg from [0,1][0,1] to the reals, the missing gg-mass, defined as the sum of g(Pr(x))g(\text{Pr}(x)) over the missing letters xx, is introduced and studied. The missing gg-mass can be used to investigate the structure of the missing part of the distribution. Specific applications for special cases such as order-α\alpha missing mass (g(p)=pαg(p)=p^{\alpha}) and the missing Shannon entropy (g(p)=plogpg(p)=-p\log p) include estimating distance from uniformity of the missing distribution and its partial estimation. Minimax estimation is studied for order-α\alpha missing mass for integer values of α\alpha and exact minimax convergence rates are obtained. Concentration is studied for a class of functions gg and specific results are derived for order-α\alpha missing mass and missing Shannon entropy. Sub-Gaussian tail bounds with near-optimal worst-case variance factors are derived. Two new notions of concentration, named strongly sub-Gamma and filtered sub-Gaussian concentration, are introduced and shown to result in right tail bounds that are better than those obtained from sub-Gaussian concentration
    corecore