6 research outputs found

    Insights into Model-Agnostic Meta-Learning on Reinforcement Learning Tasks

    No full text
    Meta-learning has been gaining traction in the Deep Learning field as an approach to build models that are able to efficiently adapt to new tasks after deployment. Contrary to conventional Machine Learning approaches, which are trained on a specific task (e.g image classification on a set of labels), meta-learning methods are meta-trained across multiple tasks (e.g image classification across multiple sets of labels). Their end objective is to learn how to solve unseen tasks with just a few samples. One of the most renowned methods of the field is Model-Agnostic Meta-Learning (MAML). The objective of this thesis is to supplement the latest relevant research with novel observations regarding the capabilities, limitations and network dynamics of MAML. For this end, experiments were performed on the meta-reinforcement learning benchmark Meta-World. Additionally, a comparison with a recent variation of MAML, called Almost No Inner Loop (ANIL) was conducted, providing insights on the changes of the network’s representation during adaptation (meta-testing). The results of this study indicate that MAML is able to outperform the baselines on the challenging Meta-World benchmark but shows little signs actual ”rapid learning” during meta-testing thus supporting the hypothesis that it reuses features learnt during meta-training. Meta-Learning har fĂ„tt dragkraft inom Deep Learning fĂ€ltet som ett tillvĂ€gagĂ„ngssĂ€tt för att bygga modeller som effektivt kan anpassa sig till nya uppgifter efter distribution. I motsats till konventionella maskininlĂ€rnings metoder som Ă€r trĂ€nade för en specifik uppgift (t.ex. bild klassificering pĂ„ en uppsĂ€ttning klasser), sĂ„ metatrĂ€nas meta-learning metoder över flera uppgifter (t.ex. bild klassificering över flera uppsĂ€ttningar av klasser). Deras slutmĂ„l Ă€r att lĂ€ra sig att lösa osedda uppgifter med bara nĂ„gra fĂ„ prover. En av de mest kĂ€nda metoderna inom omrĂ„det Ă€r Model-Agnostic Meta-Learning (MAML). Syftet med denna avhandling Ă€r att komplettera den senaste relevanta forskningen med nya observationer avseende MAML: s kapacitet, begrĂ€nsningar och nĂ€tverksdynamik. För detta Ă€ndamĂ„l utfördes experiment pĂ„ metaförstĂ€rkningslĂ€rande riktmĂ€rke Meta-World. Dessutom gjordes en jĂ€mförelse med en ny variant av MAML, kallad Almost No Inner Loop (ANIL), som gav insikter om förĂ€ndringarna i nĂ€tverkets representation under anpassning (metatestning). Resultaten av denna studie indikerar att MAML kan övertrĂ€ffa baslinjerna för det utmanande Meta-WorldriktmĂ€rket men visar smĂ„ tecken pĂ„ faktisk ”snabb inlĂ€rning” under metatestning, vilket stödjer hypotesen att den Ă„teranvĂ€nder funktioner som den lĂ€rt sig under metatrĂ€ning
    corecore