202 research outputs found
Relevance Grounding for Planning in Relational Domains
Abstract. Probabilistic relational models are an efficient way to learn and represent the dynamics in realistic environments consisting of many objects. Autonomous intelligent agents that ground this representation for all objects need to plan in exponentially large state spaces and large sets of stochastic actions. A key insight for computational efficiency is that successful planning typically involves only a small subset of relevant objects. In this paper, we introduce a probabilistic model to represent planning with subsets of objects and provide a definition of object relevance. Our definition is sufficient to prove consistency between repeated planning in partially grounded models restricted to relevant objects and planning in the fully grounded model. We propose an algorithm that exploits object relevance to plan efficiently in complex domains. Empirical results in a simulated 3D blocksworld with an articulated manipulator and realistic physics prove the effectiveness of our approach.
Learning action representations using kernel perceptrons
Action representation is fundamental to many aspects of cognition, including language.
Theories of situated cognition suggest that the form of such representation is distinctively
determined by grounding in the real world. This thesis tackles the question of
how to ground action representations, and proposes an approach for learning action
models in noisy, partially observable domains, using deictic representations and kernel
perceptrons.
Agents operating in real-world settings often require domain models to support
planning and decision-making. To operate effectively in the world, an agent must be
able to accurately predict when its actions will be successful, and what the effects of its
actions will be. Only when a reliable action model is acquired can the agent usefully
combine sequences of actions into plans, in order to achieve wider goals. However,
learning the dynamics of a domain can be a challenging problem: agents’ observations
may be noisy, or incomplete; actions may be non-deterministic; the world itself may
be noisy; or the world may contain many objects and relations which are irrelevant.
In this thesis, I first show that voted perceptrons, equipped with the DNF family
of kernels, easily learn action models in STRIPS domains, even when subject to noise
and partial observability. Key to the learning process is, firstly, the implicit exploration
of the space of conjunctions of possible fluents (the space of potential action preconditions)
enabled by the DNF kernels; secondly, the identification of objects playing
similar roles in different states, enabled by a simple deictic representation; and lastly,
the use of an attribute-value representation for world states.
Next, I extend the model to more complex domains by generalising both the kernel
and the deictic representation to a relational setting, where world states are represented
as graphs. Finally, I propose a method to extract STRIPS-like rules from the learnt
models. I give preliminary results for STRIPS domains and discuss how the method
can be extended to more complex domains. As such, the model is both appropriate for
learning data generated by robot explorations as well as suitable for use by automated
planning systems. This combination is essential for the development of autonomous
agents which can learn action models from their environment and use them to generate
successful plans
Decision-Making with Multi-Step Expert Advice on the Web
This thesis deals with solving multi-step tasks by using advice from experts, which are algorithms to solve individual steps of such tasks. We contribute with methods for maximizing the number of correct task solutions by selecting and combining experts for individual task instances and methods for automating the process of solving tasks on the Web, where experts are available as Web services.
Multi-step tasks frequently occur in Natural Language Processing (NLP) or Computer Vision, and as research progresses an increasing amount of exchangeable experts for the same steps are available on the Web. Service provider platforms such as Algorithmia monetize expert access by making expert services available via their platform and having customers pay for single executions.
Such experts can be used to solve diverse tasks, which often consist of multiple steps and thus require pipelines of experts to generate hypotheses.
We perceive two distinct problems for solving multi-step tasks with expert services: (1) Given that the task is sufficiently complex, no single pipeline generates correct solutions for all possible task instances. One thus must learn how to construct individual expert pipelines for individual task instances in order to maximize the number of correct solutions, while also taking into account the costs adhered to executing an expert. (2) To automatically solve multi-step tasks with expert services, we need to discover, execute and compose expert pipelines. With mostly textual descriptions of complex functionalities and input parameters, Web automation entails to integrate available expert services and data, interpreting user-specified task goals or efficiently finding correct service configurations.
In this thesis, we present solutions to both problems: (1) We enable to learn well-performing expert pipelines assuming available reference data sets (comprising a number of task instances and solutions), where we distinguish between centralized and decentralized decision-making. We formalize the problem as specialization of a Markov Decision Process (MDP), which we refer to as Expert Process (EP) and integrate techniques from Statistical Relational Learning (SRL) or Multiagent coordination. (2) We develop a framework for automatically discovering, executing and composing expert pipelines by exploiting methods developed for the Semantic Web. We lift the representations of experts with structured vocabularies modeled with the Resource Description Framework (RDF) and extend EPs to Semantic Expert Processes (SEPs) to enable the data-driven execution of experts in Web-based architectures.
We evaluate our methods in different domains, namely Medical Assistance with tasks in Image Processing and Surgical Phase Recognition, and NLP for textual data on the Web, where we deal with the task of Named Entity Recognition and Disambiguation (NERD)
- …