316,422 research outputs found

    A Tutorial on Bayesian Optimization of Expensive Cost Functions, with Application to Active User Modeling and Hierarchical Reinforcement Learning

    Full text link
    We present a tutorial on Bayesian optimization, a method of finding the maximum of expensive cost functions. Bayesian optimization employs the Bayesian technique of setting a prior over the objective function and combining it with evidence to get a posterior function. This permits a utility-based selection of the next observation to make on the objective function, which must take into account both exploration (sampling from areas of high uncertainty) and exploitation (sampling areas likely to offer improvement over the current best observation). We also present two detailed extensions of Bayesian optimization, with experiments---active user modelling with preferences, and hierarchical reinforcement learning---and a discussion of the pros and cons of Bayesian optimization based on our experiences

    Scenario selection optimization in system engineering projects under uncertainty: a multi-objective ant colony method based on a learning mechanism

    Get PDF
    This paper presents a multi-objective Ant Colony Optimization (MOACO) algorithm based on a learning mechanism (named MOACO-L) for the optimization of project scenario selection under uncertainty in a system engineering (SE) process. The objectives to minimize are the total cost of the project, its total duration and the global risk. Risk is considered as an uncertainty about task costs and task durations in the project graph. The learning mechanism aims to improve the MOACO algorithm for the selection of optimal project scenarios in aSE project by considering the uncertainties on the project objectives. The MOACO-L algorithm is then developed by taking into account ants’ past experiences. The learning mechanism allows a better exploration of the search space and an improvement of the MOACO algorithm performance. To validate our approach, some experimental results are presented

    Transcranial Direct Corrent stimulation (tDCS) of the anterior prefrontal cortex (aPFC) modulates reinforcement learning and decision-making under uncertainty: A doubleblind crossover study

    Get PDF
    Reinforcement learning refers to the ability to acquire information from the outcomes of prior choices (i.e. positive and negative) in order to make predictions on the effect of future decision and adapt the behaviour basing on past experiences. The anterior prefrontal cortex (aPFC) is considered to play a key role in the representation of event value, reinforcement learning and decision-making. However, a causal evidence of the involvement of this area in these processes has not been provided yet. The aim of the study was to test the role of the orbitofrontal cortex in feedback processing, reinforcement learning and decision-making under uncertainly. Eighteen healthy individuals underwent three sessions of tDCS over the prefrontal pole (anodal, cathodal, sham) during a probabilistic learning (PL) task. In the PL task, participants were invited to learn the covert probabilistic stimulusoutcome association from positive and negative feedbacks in order to choose the best option. Afterwards, a probabilistic selection (PS) task was delivered to assess decisions based on the stimulus-reward associations acquired in the PL task. During cathodal tDCS, accuracy in the PL task was reduced and participants were less prone to maintain their choice after positive feedback or to change it after a negative one (i.e., winstay and lose-shift behavior). In addition, anodal tDCS affected the subsequent PS task by reducing the ability to choose the best alternative during hard probabilistic decisions. In conclusion, the present study suggests a causal role of aPFC in feedback trial-by-trial behavioral adaptation and decision-making under uncertainty

    Bootstrapping trust evaluations through stereotypes

    Get PDF
    Publisher PD
    • …
    corecore