20 research outputs found

    Learning with a Drifting Target Concept

    Full text link
    We study the problem of learning in the presence of a drifting target concept. Specifically, we provide bounds on the error rate at a given time, given a learner with access to a history of independent samples labeled according to a target concept that can change on each round. One of our main contributions is a refinement of the best previous results for polynomial-time algorithms for the space of linear separators under a uniform distribution. We also provide general results for an algorithm capable of adapting to a variable rate of drift of the target concept. Some of the results also describe an active learning variant of this setting, and provide bounds on the number of queries for the labels of points in the sequence sufficient to obtain the stated bounds on the error rates

    Black-box Generalization of Machine Teaching

    Full text link
    Hypothesis-pruning maximizes the hypothesis updates for active learning to find those desired unlabeled data. An inherent assumption is that this learning manner can derive those updates into the optimal hypothesis. However, its convergence may not be guaranteed well if those incremental updates are negative and disordered. In this paper, we introduce a black-box teaching hypothesis hTh^\mathcal{T} employing a tighter slack term (1+FT(h^t))Ξ”t\left(1+\mathcal{F}^{\mathcal{T}}(\widehat{h}_t)\right)\Delta_t to replace the typical 2Ξ”t2\Delta_t for pruning. Theoretically, we prove that, under the guidance of this teaching hypothesis, the learner can converge into a tighter generalization error and label complexity bound than those non-educated learners who do not receive any guidance from a teacher:1) the generalization error upper bound can be reduced from R(hβˆ—)+4Ξ”Tβˆ’1R(h^*)+4\Delta_{T-1} to approximately R(hT)+2Ξ”Tβˆ’1R(h^{\mathcal{T}})+2\Delta_{T-1}, and 2) the label complexity upper bound can be decreased from 4ΞΈ(TR(hβˆ—)+2O(T))4 \theta\left(TR(h^{*})+2O(\sqrt{T})\right) to approximately 2ΞΈ(2TR(hT)+3O(T))2\theta\left(2TR(h^{\mathcal{T}})+3 O(\sqrt{T})\right). To be strict with our assumption, self-improvement of teaching is firstly proposed when hTh^\mathcal{T} loosely approximates hβˆ—h^*. Against learning, we further consider two teaching scenarios: teaching a white-box and black-box learner. Experiments verify this idea and show better generalization performance than the fundamental active learning strategies, such as IWAL, IWAL-D, etc
    corecore