793,893 research outputs found

    Controlling Fairness and Bias in Dynamic Learning-to-Rank

    Full text link
    Rankings are the primary interface through which many online platforms match users to items (e.g. news, products, music, video). In these two-sided markets, not only the users draw utility from the rankings, but the rankings also determine the utility (e.g. exposure, revenue) for the item providers (e.g. publishers, sellers, artists, studios). It has already been noted that myopically optimizing utility to the users, as done by virtually all learning-to-rank algorithms, can be unfair to the item providers. We, therefore, present a learning-to-rank approach for explicitly enforcing merit-based fairness guarantees to groups of items (e.g. articles by the same publisher, tracks by the same artist). In particular, we propose a learning algorithm that ensures notions of amortized group fairness, while simultaneously learning the ranking function from implicit feedback data. The algorithm takes the form of a controller that integrates unbiased estimators for both fairness and utility, dynamically adapting both as more data becomes available. In addition to its rigorous theoretical foundation and convergence guarantees, we find empirically that the algorithm is highly practical and robust.Comment: First two authors contributed equally. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval 202

    Adaptive Guidance: Effects On Self-Regulated Learning In Technology-Based Training

    Get PDF
    Guidance provides trainees with the information necessary to make effective use of the learner control inherent in technology-based training, but also allows them to retain a sense of control over their learning (Bell & Kozlowski, 2002). One challenge, however, is determining how much learner control, or autonomy, to build into the guidance strategy. We examined the effects of alternative forms of guidance (autonomy supportive vs. controlling) on trainees’ learning and performance, and examined trainees’ cognitive ability and motivation to learn as potential moderators of these effects. Consistent with our hypotheses, trainees receiving adaptive guidance had higher levels of knowledge and performance than trainees in a learner control guidance. Controlling guidance had the most consistent positive impact on the learning outcomes, while autonomy supportive guidance demonstrated utility for more strategic outcomes. In addition, guidance was generally more effective for trainees with higher levels of cognitive ability and autonomy guidance served to enhance the positive effects of motivation to learn on the training outcomes

    Learning obstacle avoidance with an operant behavioral model

    Get PDF
    Artificial intelligence researchers have been attracted by the idea of having robots learn how to accomplish a task, rather than being told explicitly. Reinforcement learning has been proposed as an appealing framework to be used in controlling mobile agents. Robot learning research, as well as research in biological systems, face many similar problems in order to display high flexibility in performing a variety of tasks. In this work, the controlling of a vehicle in an avoidance task by a previously developed operant learning model (a form of animal learning) is studied. An environment in which a mobile robot with proximity sensors has to minimize the punishment for colliding against obstacles is simulated. The results were compared with the Q-Learning algorithm, and the proposed model had better performance. In this way a new artificial intelligence agent inspired by neurobiology, psychology, and ethology research is proposed.Fil: Gutnisky, D. A.. Universidad de Buenos Aires. Facultad de Ingeniería.Instituto de Ingeniería Biomédica; ArgentinaFil: Zanutto, Bonifacio Silvano. Consejo Nacional de Investigaciones Científicas y Técnicas. Instituto de Biología y Medicina Experimental. Fundación de Instituto de Biología y Medicina Experimental. Instituto de Biología y Medicina Experimental; Argentina. Universidad de Buenos Aires. Facultad de Ingeniería.Instituto de Ingeniería Biomédica; Argentin

    Using assessment to drive retention and achievement upwards

    Get PDF
    Identifying what the student really wants from their course provides a method of controlling the learning process. We contend that the prime student driver is the achievement of their qualification by gaining marks and not learning itself. We examine the effect of tying attendance directly to the achievement of marks. It is found that attendance improves from around 60% to 98.5%. Further methods are discussed which might have a direct affect on retention and achievement based on the identification of student wants

    Seeing is as Good as Doing

    Get PDF
    Given the privileged status claimed for active learning in a variety of domains (visuomotor learning, causal induction, problem solving, education, skill learning), the present study examines whether action-based learning is a necessary, or a suffi cient, means of acquiring the relevant skills needed to perform a task typically described as requiring active learning. To achieve this, the present study compares the effects of action-based and observation-based learning when controlling a complex dynamic task environment (N = 96). Both action- and observation-based individuals learn either by describing the changes in the environment in the form of a conditional statement, or not. The study reveals that for both active and observational learners, advantages in performance (p < .05), accuracy in knowledge of the task (p < .05), and self-insight (p < .05) are found when learning is based on inducing rules from the task environment. Moreover, the study provides evidence suggesting that, given task instructions that encourage rule-based knowledge, both active and observation-based learning can lead to high levels of problem solving skills in a complex dynamic environment

    Learning curves for Soft Margin Classifiers

    Full text link
    Typical learning curves for Soft Margin Classifiers (SMCs) learning both realizable and unrealizable tasks are determined using the tools of Statistical Mechanics. We derive the analytical behaviour of the learning curves in the regimes of small and large training sets. The generalization errors present different decay laws towards the asymptotic values as a function of the training set size, depending on general geometrical characteristics of the rule to be learned. Optimal generalization curves are deduced through a fine tuning of the hyperparameter controlling the trade-off between the error and the regularization terms in the cost function. Even if the task is realizable, the optimal performance of the SMC is better than that of a hard margin Support Vector Machine (SVM) learning the same rule, and is very close to that of the Bayesian classifier.Comment: 26 pages, 10 figure
    • …
    corecore