14 research outputs found

    Near-Optimal Density Estimation in Near-Linear Time Using Variable-Width Histograms

    Get PDF
    Let pp be an unknown and arbitrary probability distribution over [0,1)[0,1). We consider the problem of {\em density estimation}, in which a learning algorithm is given i.i.d. draws from pp and must (with high probability) output a hypothesis distribution that is close to pp. The main contribution of this paper is a highly efficient density estimation algorithm for learning using a variable-width histogram, i.e., a hypothesis distribution with a piecewise constant probability density function. In more detail, for any kk and ϵ\epsilon, we give an algorithm that makes O~(k/ϵ2)\tilde{O}(k/\epsilon^2) draws from pp, runs in O~(k/ϵ2)\tilde{O}(k/\epsilon^2) time, and outputs a hypothesis distribution hh that is piecewise constant with O(klog2(1/ϵ))O(k \log^2(1/\epsilon)) pieces. With high probability the hypothesis hh satisfies dTV(p,h)Coptk(p)+ϵd_{\mathrm{TV}}(p,h) \leq C \cdot \mathrm{opt}_k(p) + \epsilon, where dTVd_{\mathrm{TV}} denotes the total variation distance (statistical distance), CC is a universal constant, and optk(p)\mathrm{opt}_k(p) is the smallest total variation distance between pp and any kk-piecewise constant distribution. The sample size and running time of our algorithm are optimal up to logarithmic factors. The "approximation factor" CC in our result is inherent in the problem, as we prove that no algorithm with sample size bounded in terms of kk and ϵ\epsilon can achieve C<2C<2 regardless of what kind of hypothesis distribution it uses.Comment: conference version appears in NIPS 201

    Robust Learning of Fixed-Structure Bayesian Networks

    Full text link
    We investigate the problem of learning Bayesian networks in a robust model where an ϵ\epsilon-fraction of the samples are adversarially corrupted. In this work, we study the fully observable discrete case where the structure of the network is given. Even in this basic setting, previous learning algorithms either run in exponential time or lose dimension-dependent factors in their error guarantees. We provide the first computationally efficient robust learning algorithm for this problem with dimension-independent error guarantees. Our algorithm has near-optimal sample complexity, runs in polynomial time, and achieves error that scales nearly-linearly with the fraction of adversarially corrupted samples. Finally, we show on both synthetic and semi-synthetic data that our algorithm performs well in practice
    corecore