793,893 research outputs found
Controlling Fairness and Bias in Dynamic Learning-to-Rank
Rankings are the primary interface through which many online platforms match
users to items (e.g. news, products, music, video). In these two-sided markets,
not only the users draw utility from the rankings, but the rankings also
determine the utility (e.g. exposure, revenue) for the item providers (e.g.
publishers, sellers, artists, studios). It has already been noted that
myopically optimizing utility to the users, as done by virtually all
learning-to-rank algorithms, can be unfair to the item providers. We,
therefore, present a learning-to-rank approach for explicitly enforcing
merit-based fairness guarantees to groups of items (e.g. articles by the same
publisher, tracks by the same artist). In particular, we propose a learning
algorithm that ensures notions of amortized group fairness, while
simultaneously learning the ranking function from implicit feedback data. The
algorithm takes the form of a controller that integrates unbiased estimators
for both fairness and utility, dynamically adapting both as more data becomes
available. In addition to its rigorous theoretical foundation and convergence
guarantees, we find empirically that the algorithm is highly practical and
robust.Comment: First two authors contributed equally. In Proceedings of the 43rd
International ACM SIGIR Conference on Research and Development in Information
Retrieval 202
Adaptive Guidance: Effects On Self-Regulated Learning In Technology-Based Training
Guidance provides trainees with the information necessary to make effective use of the learner control inherent in technology-based training, but also allows them to retain a sense of control over their learning (Bell & Kozlowski, 2002). One challenge, however, is determining how much learner control, or autonomy, to build into the guidance strategy. We examined the effects of alternative forms of guidance (autonomy supportive vs. controlling) on trainees’ learning and performance, and examined trainees’ cognitive ability and motivation to learn as potential moderators of these effects. Consistent with our hypotheses, trainees receiving adaptive guidance had higher levels of knowledge and performance than trainees in a learner control guidance. Controlling guidance had the most consistent positive impact on the learning outcomes, while autonomy supportive guidance demonstrated utility for more strategic outcomes. In addition, guidance was generally more effective for trainees with higher levels of cognitive ability and autonomy guidance served to enhance the positive effects of motivation to learn on the training outcomes
Learning obstacle avoidance with an operant behavioral model
Artificial intelligence researchers have been attracted by the idea of having robots learn how to accomplish a task, rather than being told explicitly. Reinforcement learning has been proposed as an appealing framework to be used in controlling mobile agents. Robot learning research, as well as research in biological systems, face many similar problems in order to display high flexibility in performing a variety of tasks. In this work, the controlling of a vehicle in an avoidance task by a previously developed operant learning model (a form of animal learning) is studied. An environment in which a mobile robot with proximity sensors has to minimize the punishment for colliding against obstacles is simulated. The results were compared with the Q-Learning algorithm, and the proposed model had better performance. In this way a new artificial intelligence agent inspired by neurobiology, psychology, and ethology research is proposed.Fil: Gutnisky, D. A.. Universidad de Buenos Aires. Facultad de IngenierĂa.Instituto de IngenierĂa BiomĂ©dica; ArgentinaFil: Zanutto, Bonifacio Silvano. Consejo Nacional de Investigaciones CientĂficas y TĂ©cnicas. Instituto de BiologĂa y Medicina Experimental. FundaciĂłn de Instituto de BiologĂa y Medicina Experimental. Instituto de BiologĂa y Medicina Experimental; Argentina. Universidad de Buenos Aires. Facultad de IngenierĂa.Instituto de IngenierĂa BiomĂ©dica; Argentin
Using assessment to drive retention and achievement upwards
Identifying what the student really wants from their course provides a method of controlling the learning process. We contend that the prime student driver is the achievement of their qualification by gaining marks and not learning itself. We examine the effect of tying attendance directly to the achievement of marks. It is found that attendance improves from around 60% to 98.5%. Further methods are discussed which might have a direct affect on retention and achievement based on the identification of student wants
Seeing is as Good as Doing
Given the privileged status claimed for active learning in a variety of domains (visuomotor learning, causal induction, problem solving, education, skill learning), the present study examines whether action-based learning is a necessary, or a suffi cient, means of acquiring the relevant skills needed to perform a task typically described as requiring active learning. To achieve this, the present study compares the effects of action-based and observation-based learning when controlling a complex dynamic task environment (N = 96). Both action- and observation-based individuals learn either by describing the changes in the environment in the form of a conditional statement, or not. The study reveals that for both active and observational learners, advantages in performance (p < .05), accuracy in knowledge of the task (p < .05), and self-insight (p < .05) are found when learning is based on inducing rules from the task environment. Moreover, the study provides evidence suggesting that, given task instructions that encourage rule-based knowledge, both active and observation-based learning can lead to high levels of problem solving skills in a complex dynamic environment
Learning curves for Soft Margin Classifiers
Typical learning curves for Soft Margin Classifiers (SMCs) learning both
realizable and unrealizable tasks are determined using the tools of Statistical
Mechanics. We derive the analytical behaviour of the learning curves in the
regimes of small and large training sets. The generalization errors present
different decay laws towards the asymptotic values as a function of the
training set size, depending on general geometrical characteristics of the rule
to be learned. Optimal generalization curves are deduced through a fine tuning
of the hyperparameter controlling the trade-off between the error and the
regularization terms in the cost function. Even if the task is realizable, the
optimal performance of the SMC is better than that of a hard margin Support
Vector Machine (SVM) learning the same rule, and is very close to that of the
Bayesian classifier.Comment: 26 pages, 10 figure
- …