10 research outputs found

    Black-box Generalization of Machine Teaching

    Full text link
    Hypothesis-pruning maximizes the hypothesis updates for active learning to find those desired unlabeled data. An inherent assumption is that this learning manner can derive those updates into the optimal hypothesis. However, its convergence may not be guaranteed well if those incremental updates are negative and disordered. In this paper, we introduce a black-box teaching hypothesis hTh^\mathcal{T} employing a tighter slack term (1+FT(h^t))Δt\left(1+\mathcal{F}^{\mathcal{T}}(\widehat{h}_t)\right)\Delta_t to replace the typical 2Δt2\Delta_t for pruning. Theoretically, we prove that, under the guidance of this teaching hypothesis, the learner can converge into a tighter generalization error and label complexity bound than those non-educated learners who do not receive any guidance from a teacher:1) the generalization error upper bound can be reduced from R(h∗)+4ΔT−1R(h^*)+4\Delta_{T-1} to approximately R(hT)+2ΔT−1R(h^{\mathcal{T}})+2\Delta_{T-1}, and 2) the label complexity upper bound can be decreased from 4θ(TR(h∗)+2O(T))4 \theta\left(TR(h^{*})+2O(\sqrt{T})\right) to approximately 2θ(2TR(hT)+3O(T))2\theta\left(2TR(h^{\mathcal{T}})+3 O(\sqrt{T})\right). To be strict with our assumption, self-improvement of teaching is firstly proposed when hTh^\mathcal{T} loosely approximates h∗h^*. Against learning, we further consider two teaching scenarios: teaching a white-box and black-box learner. Experiments verify this idea and show better generalization performance than the fundamental active learning strategies, such as IWAL, IWAL-D, etc

    Adaptation Algorithms for Neural Network-Based Speech Recognition: An Overview

    Get PDF
    We present a structured overview of adaptation algorithms for neural network-based speech recognition, considering both hybrid hidden Markov model / neural network systems and end-to-end neural network systems, with a focus on speaker adaptation, domain adaptation, and accent adaptation. The overview characterizes adaptation algorithms as based on embeddings, model parameter adaptation, or data augmentation. We present a meta-analysis of the performance of speech recognition adaptation algorithms, based on relative error rate reductions as reported in the literature.Comment: Submitted to IEEE Open Journal of Signal Processing. 30 pages, 27 figure
    corecore