305,750 research outputs found
Expectation-Aware Planning: A Unifying Framework for Synthesizing and Executing Self-Explaining Plans for Human-Aware Planning
In this work, we present a new planning formalism called Expectation-Aware
planning for decision making with humans in the loop where the human's
expectations about an agent may differ from the agent's own model. We show how
this formulation allows agents to not only leverage existing strategies for
handling model differences but can also exhibit novel behaviors that are
generated through the combination of these different strategies. Our
formulation also reveals a deep connection to existing approaches in epistemic
planning. Specifically, we show how we can leverage classical planning
compilations for epistemic planning to solve Expectation-Aware planning
problems. To the best of our knowledge, the proposed formulation is the first
complete solution to decision-making in the presence of diverging user
expectations that is amenable to a classical planning compilation while
successfully combining previous works on explanation and explicability. We
empirically show how our approach provides a computational advantage over
existing approximate approaches that unnecessarily try to search in the space
of models while also failing to facilitate the full gamut of behaviors enabled
by our framework
KR: An Architecture for Knowledge Representation and Reasoning in Robotics
This paper describes an architecture that combines the complementary
strengths of declarative programming and probabilistic graphical models to
enable robots to represent, reason with, and learn from, qualitative and
quantitative descriptions of uncertainty and knowledge. An action language is
used for the low-level (LL) and high-level (HL) system descriptions in the
architecture, and the definition of recorded histories in the HL is expanded to
allow prioritized defaults. For any given goal, tentative plans created in the
HL using default knowledge and commonsense reasoning are implemented in the LL
using probabilistic algorithms, with the corresponding observations used to
update the HL history. Tight coupling between the two levels enables automatic
selection of relevant variables and generation of suitable action policies in
the LL for each HL action, and supports reasoning with violation of defaults,
noisy observations and unreliable actions in large and complex domains. The
architecture is evaluated in simulation and on physical robots transporting
objects in indoor domains; the benefit on robots is a reduction in task
execution time of 39% compared with a purely probabilistic, but still
hierarchical, approach.Comment: The paper appears in the Proceedings of the 15th International
Workshop on Non-Monotonic Reasoning (NMR 2014
- …