2 research outputs found

    Denoising as well as the best of any two denoisers

    Full text link
    Given two arbitrary sequences of denoisers for block lengths tending to infinity we ask if it is possible to construct a third sequence of denoisers with an asymptotically vanishing (in block length) excess expected loss relative to the best expected loss of the two given denoisers for all clean channel input sequences. As in the setting of DUDE [1], which solves this problem when the given denoisers are sliding block denoisers, the construction is allowed to depend on the two given denoisers and the channel transition probabilities. We show that under certain restrictions on the two given denoisers the problem can be solved using a straightforward application of a known loss estimation paradigm. We then show by way of a counter-example that the loss estimation approach fails in the general case. Finally, we show that for the binary symmetric channel, combining the loss estimation with a randomization step leads to a solution to the stated problem under no restrictions on the given denoisers.Comment: 19 pages. Appeared, in part, in Proceedings of 2013 IEEE Intl. Symp. on Info. Theory. This version has full proofs (e.g., of Proposition 2

    A Denoising Loss Bound for Neural Network based Universal Discrete Denoisers

    Full text link
    We obtain a denoising loss bound of the recently proposed neural network based universal discrete denoiser, Neural DUDE, which can adaptively learn its parameters solely from the noise-corrupted data, by minimizing the \emph{empirical estimated loss}. The resulting bound resembles the generalization error bound of the standard empirical risk minimizers (ERM) in supervised learning, and we show that the well-known bias-variance tradeoff also exists in our loss bound. The key tool we develop is the concentration of the unbiased estimated loss on the true denoising loss, which is shown to hold \emph{uniformly} for \emph{all} bounded network parameters and \emph{all} underlying clean sequences. For proving our main results, we make a novel application of the tools from the statistical learning theory. Finally, we show that the hyperparameters of Neural DUDE can be chosen from a small validation set to significantly improve the denoising performance, as predicted by the theoretical result of this paper.Comment: submitted to ICML 201
    corecore