114,937 research outputs found

    Online Learning with Feedback Graphs: Beyond Bandits

    Get PDF
    We study a general class of online learning problems where the feedback is specified by a graph. This class includes online prediction with expert advice and the multi-armed bandit problem, but also several learning problems where the online player does not necessarily observe his own loss. We analyze how the structure of the feedback graph controls the inherent difficulty of the induced TT-round learning problem. Specifically, we show that any feedback graph belongs to one of three classes: strongly observable graphs, weakly observable graphs, and unobservable graphs. We prove that the first class induces learning problems with Θ~(α1/2T1/2)\widetilde\Theta(\alpha^{1/2} T^{1/2}) minimax regret, where α\alpha is the independence number of the underlying graph; the second class induces problems with Θ~(δ1/3T2/3)\widetilde\Theta(\delta^{1/3}T^{2/3}) minimax regret, where δ\delta is the domination number of a certain portion of the graph; and the third class induces problems with linear minimax regret. Our results subsume much of the previous work on learning with feedback graphs and reveal new connections to partial monitoring games. We also show how the regret is affected if the graphs are allowed to vary with time

    On the Minimax Regret for Online Learning with Feedback Graphs

    Full text link
    In this work, we improve on the upper and lower bounds for the regret of online learning with strongly observable undirected feedback graphs. The best known upper bound for this problem is O(αTlnK)\mathcal{O}\bigl(\sqrt{\alpha T\ln K}\bigr), where KK is the number of actions, α\alpha is the independence number of the graph, and TT is the time horizon. The lnK\sqrt{\ln K} factor is known to be necessary when α=1\alpha = 1 (the experts case). On the other hand, when α=K\alpha = K (the bandits case), the minimax rate is known to be Θ(KT)\Theta\bigl(\sqrt{KT}\bigr), and a lower bound Ω(αT)\Omega\bigl(\sqrt{\alpha T}\bigr) is known to hold for any α\alpha. Our improved upper bound O(αT(1+ln(K/α)))\mathcal{O}\bigl(\sqrt{\alpha T(1+\ln(K/\alpha))}\bigr) holds for any α\alpha and matches the lower bounds for bandits and experts, while interpolating intermediate cases. To prove this result, we use FTRL with qq-Tsallis entropy for a carefully chosen value of q[1/2,1)q \in [1/2, 1) that varies with α\alpha. The analysis of this algorithm requires a new bound on the variance term in the regret. We also show how to extend our techniques to time-varying graphs, without requiring prior knowledge of their independence numbers. Our upper bound is complemented by an improved Ω(αT(lnK)/(lnα))\Omega\bigl(\sqrt{\alpha T(\ln K)/(\ln\alpha)}\bigr) lower bound for all α>1\alpha > 1, whose analysis relies on a novel reduction to multitask learning. This shows that a logarithmic factor is necessary as soon as α<K\alpha < K

    Stochastic Online Learning with Probabilistic Graph Feedback

    Full text link
    We consider a problem of stochastic online learning with general probabilistic graph feedback, where each directed edge in the feedback graph has probability pijp_{ij}. Two cases are covered. (a) The one-step case, where after playing arm ii the learner observes a sample reward feedback of arm jj with independent probability pijp_{ij}. (b) The cascade case where after playing arm ii the learner observes feedback of all arms jj in a probabilistic cascade starting from ii -- for each (i,j)(i,j) with probability pijp_{ij}, if arm ii is played or observed, then a reward sample of arm jj would be observed with independent probability pijp_{ij}. Previous works mainly focus on deterministic graphs which corresponds to one-step case with pij{0,1}p_{ij} \in \{0,1\}, an adversarial sequence of graphs with certain topology guarantees, or a specific type of random graphs. We analyze the asymptotic lower bounds and design algorithms in both cases. The regret upper bounds of the algorithms match the lower bounds with high probability
    corecore