29 research outputs found

    Schematic example of independent and joint clustering agents.

    No full text
    <p><i>Top Left</i>: The independent clustering agent groups each context into two clusters, associated with a reward (<i>R</i>) and mapping (<i>Ď•</i>) function, respectively. Planning involves combining these functions to generate a policy. The clustering prior induces a parsimony bias such that new contexts are more likely to be assigned to more popular clusters. Arrows denote assignment of context into clusters and creation of policies from component functions. <i>Top Right</i>: The joint clustering agent assigns each context into a cluster linked to both functions (i.e., assumes a holistic task structure), and hence the policy is determined by this cluster assignment. In this example, both agents generate the same two policies for the three contexts but the independent clustering agent generalizes the reward function across all three contexts. <i>Bottom</i>: An example mapping (<i>left</i>) and reward function (<i>right</i>) for a gridworld task.</p

    Performance of independent vs. joint clustering in predicting a sequence <i>X</i><sup><i>R</i></sup>, measured in bits of information gained by observation of each item.

    No full text
    <p><i>Left</i>: Relative performance of independent clustering over joint clustering as a function of mutual information between the rewards and transitions. <i>Right</i>: Noise in observation of <i>X</i><sup><i>T</i></sup> sequences parametrically increases advantage for independent clustering. Green line shows relative performance in sequences with no residual uncertainty in <i>R</i> given <i>T</i> (perfect correspondence), orange line shows relative performance for a sequence with residual uncertainty <i>H</i>(<i>R</i>|<i>T</i>) > 0bits.</p

    Simulation 1.

    No full text
    <p><i>A</i>: Schematic representation of the task domain. Four contexts (blue circles) were simulated, each paired with a unique combination of one of two goal locations (reward functions) and one of two mappings. <i>B</i>: Number of steps taken by each agent shown across trials within a single context (left) and over all trials (right). Fewer steps reflect better performance. <i>C</i>: KL-divergence of the models’ estimates of the reward (left) and mapping (right) functions as a function of time. Lower KL-divergence represents better function estimates. Time shown as the number of trials in a context (left) and the number of steps in a context collapsed across trials (right) for clarity.</p

    Meta-agent.

    No full text
    <p><i>A</i>: On each trial, the meta-agent samples the policy of joint or independent actor based on model evidence for each strategy. Both agents, and their model evidences, are updated at each time step. <i>B</i>: Overall performance of independent, joint and meta agents on simulation 1 <i>C</i>: Overall performance of independent, joint and meta agents on simulation 2 <i>D,E</i>: Probability of the selecting the policy joint clustering over time in simulation 1 (<i>D</i>) and simulation 2 (<i>E</i>).</p

    Simulation 2.

    No full text
    <p>A: Schematic representation of the second task domain. Eight contexts (blue circles) were simulated, each paired with a combination of one of four orthogonal reward functions and one of four mappings, such that each pairing was repeated across two contexts, providing a discoverable relationship. B: Number of steps taken by each agent shown across trials within a single context (left) and over all trials (right). C: KL-divergence of the models’ estimates of the reward (left) and mapping (right) functions as a function of time.</p

    “Diabolic rooms problem”.

    No full text
    <p><i>A</i>: Schamatic diagram of rooms problem. Agents enter a room and choose a door to navigate to the next room. Choosing the correct door (green) leads to the next room while choosing the other two doors leads to the start of the task. The agent learns three mappings across rooms <i>B</i>: Distribution of steps taken to solve the task by the three agents (left) and median of the distributions (right). <i>C,D</i>: Regression of the number of steps to complete the task as a function of grid area (C) and the number of rooms in the task (D) for the joint and independent clustering agents.</p

    Hiragana stimuli used in the probabilistic learning task in experiments 1a and 1b and experiment 3.

    No full text
    <p>Each pair of stimuli was randomly presented in separate trials. During each trial participants chose one of the stimuli of the pair. Feedback following participant’s choice was determined probabilistically. Reward probability (indicated below each stimulus) differed between characters. Task version 1 and 2 differed with respect to the characters associated with more probable positive and more probable negative feedback, respectively. Specifically, in task version 2, Hiragana stimuli were switched within the AB, CD and EF pair. Half of the participants were subjected to version 1 and the other half to task version 2. In experiment 3, a task version was used for which only the Hiragana stimuli within the AB pair were switched.</p
    corecore