13 research outputs found

    Ranking a set of objects: a graph based least-square approach

    Get PDF
    We consider the problem of ranking NN objects starting from a set of noisy pairwise comparisons provided by a crowd of equal workers. We assume that objects are endowed with intrinsic qualities and that the probability with which an object is preferred to another depends only on the difference between the qualities of the two competitors. We propose a class of non-adaptive ranking algorithms that rely on a least-squares optimization criterion for the estimation of qualities. Such algorithms are shown to be asymptotically optimal (i.e., they require O(Nϵ2logNδ)O(\frac{N}{\epsilon^2}\log \frac{N}{\delta}) comparisons to be (ϵ,δ)(\epsilon, \delta)-PAC). Numerical results show that our schemes are very efficient also in many non-asymptotic scenarios exhibiting a performance similar to the maximum-likelihood algorithm. Moreover, we show how they can be extended to adaptive schemes and test them on real-world datasets

    Borda Regret Minimization for Generalized Linear Dueling Bandits

    Full text link
    Dueling bandits are widely used to model preferential feedback prevalent in many applications such as recommendation systems and ranking. In this paper, we study the Borda regret minimization problem for dueling bandits, which aims to identify the item with the highest Borda score while minimizing the cumulative regret. We propose a rich class of generalized linear dueling bandit models, which cover many existing models. We first prove a regret lower bound of order Ω(d2/3T2/3)\Omega(d^{2/3} T^{2/3}) for the Borda regret minimization problem, where dd is the dimension of contextual vectors and TT is the time horizon. To attain this lower bound, we propose an explore-then-commit type algorithm for the stochastic setting, which has a nearly matching regret upper bound O~(d2/3T2/3)\tilde{O}(d^{2/3} T^{2/3}). We also propose an EXP3-type algorithm for the adversarial linear setting, where the underlying model parameter can change at each round. Our algorithm achieves an O~(d2/3T2/3)\tilde{O}(d^{2/3} T^{2/3}) regret, which is also optimal. Empirical evaluations on both synthetic data and a simulated real-world environment are conducted to corroborate our theoretical analysis.Comment: 33 pages, 5 figure. This version includes new results for dueling bandits in the adversarial settin

    Versatile Dueling Bandits: Best-of-both-World Analyses for Online Learning from Preferences

    Get PDF
    International audienceWe study the problem of KK-armed dueling bandit for both stochastic and adversarial environments, where the goal of the learner is to aggregate information through relative preferences of pair of decisions points queried in an online sequential manner. We first propose a novel reduction from any (general) dueling bandits to multi-armed bandits and despite the simplicity, it allows us to improve many existing results in dueling bandits. In particular, \emph{we give the first best-of-both world result for the dueling bandits regret minimization problem} -- a unified framework that is guaranteed to perform optimally for both stochastic and adversarial preferences simultaneously. Moreover, our algorithm is also the first to achieve an optimal O(i=1KlogTΔi)O(\sum_{i = 1}^K \frac{\log T}{\Delta_i}) regret bound against the Condorcet-winner benchmark, which scales optimally both in terms of the arm-size KK and the instance-specific suboptimality gaps {Δi}i=1K\{\Delta_i\}_{i = 1}^K. This resolves the long-standing problem of designing an instancewise gap-dependent order optimal regret algorithm for dueling bandits (with matching lower bounds up to small constant factors). We further justify the robustness of our proposed algorithm by proving its optimal regret rate under adversarially corrupted preferences -- this outperforms the existing state-of-the-art corrupted dueling results by a large margin. In summary, we believe our reduction idea will find a broader scope in solving a diverse class of dueling bandits setting, which are otherwise studied separately from multi-armed bandits with often more complex solutions and worse guarantees. The efficacy of our proposed algorithms is empirically corroborated against the existing dueling bandit methods
    corecore