6,657 research outputs found
MaestROB: A Robotics Framework for Integrated Orchestration of Low-Level Control and High-Level Reasoning
This paper describes a framework called MaestROB. It is designed to make the
robots perform complex tasks with high precision by simple high-level
instructions given by natural language or demonstration. To realize this, it
handles a hierarchical structure by using the knowledge stored in the forms of
ontology and rules for bridging among different levels of instructions.
Accordingly, the framework has multiple layers of processing components;
perception and actuation control at the low level, symbolic planner and Watson
APIs for cognitive capabilities and semantic understanding, and orchestration
of these components by a new open source robot middleware called Project Intu
at its core. We show how this framework can be used in a complex scenario where
multiple actors (human, a communication robot, and an industrial robot)
collaborate to perform a common industrial task. Human teaches an assembly task
to Pepper (a humanoid robot from SoftBank Robotics) using natural language
conversation and demonstration. Our framework helps Pepper perceive the human
demonstration and generate a sequence of actions for UR5 (collaborative robot
arm from Universal Robots), which ultimately performs the assembly (e.g.
insertion) task.Comment: IEEE International Conference on Robotics and Automation (ICRA) 2018.
Video: https://www.youtube.com/watch?v=19JsdZi0TW
A Proposal for Semantic Map Representation and Evaluation
Semantic mapping is the incremental process of “mapping” relevant information of the world (i.e., spatial information, temporal events, agents and actions) to a formal description supported by a reasoning engine. Current research focuses on learning the semantic of environments based on their spatial location, geometry and appearance. Many methods to tackle this problem have been proposed, but the lack of a uniform representation, as well as standard benchmarking suites, prevents their direct comparison. In this paper, we propose a standardization in the representation of semantic maps, by defining an easily extensible formalism to be used on top of metric maps of the environments. Based on this, we describe the procedure to build a dataset (based on real sensor data) for benchmarking semantic mapping techniques, also hypothesizing some possible evaluation metrics. Nevertheless, by providing a tool for the construction of a semantic map ground truth, we aim at the contribution of the scientific community in acquiring data for populating the dataset
Task planning using physics-based heuristics on manipulation actions
© 2016 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.In order to solve mobile manipulation problems, the efficient combination of task and motion planning is usually required. Moreover, the incorporation of physics-based information has recently been taken into account in order to plan the tasks in a more realistic way. In the present paper, a task and motion planning framework is proposed based on a modified version of the Fast-Forward task planner that is guided by physics-based knowledge.
The proposal uses manipulation knowledge for reasoning on symbolic literals (both in offline and online modes) taking into account geometric information in order to evaluate the applicability as well as feasibility of actions while evaluating the heuristic cost. It results in an efficient search of the state space and in the obtention of low-cost physically-feasible plans. The proposal has been implemented and is illustrated with a manipulation problem consisting of a mobile robot and some fixed and manipulatable objects.Peer ReviewedPostprint (author's final draft
Physics-based Motion Planning with Temporal Logic Specifications
One of the main foci of robotics is nowadays centered in providing a great
degree of autonomy to robots. A fundamental step in this direction is to give
them the ability to plan in discrete and continuous spaces to find the required
motions to complete a complex task. In this line, some recent approaches
describe tasks with Linear Temporal Logic (LTL) and reason on discrete actions
to guide sampling-based motion planning, with the aim of finding
dynamically-feasible motions that satisfy the temporal-logic task
specifications. The present paper proposes an LTL planning approach enhanced
with the use of ontologies to describe and reason about the task, on the one
hand, and that includes physics-based motion planning to allow the purposeful
manipulation of objects, on the other hand. The proposal has been implemented
and is illustrated with didactic examples with a mobile robot in simple
scenarios where some of the goals are occupied with objects that must be
removed in order to fulfill the task.Comment: The 20th World Congress of the International Federation of Automatic
Control, 9-14 July 201
Ontology-based collaborative framework for disaster recovery scenarios
This paper aims at designing of adaptive framework for supporting
collaborative work of different actors in public safety and disaster recovery
missions. In such scenarios, firemen and robots interact to each other to reach
a common goal; firemen team is equipped with smart devices and robots team is
supplied with communication technologies, and should carry on specific tasks.
Here, reliable connection is mandatory to ensure the interaction between
actors. But wireless access network and communication resources are vulnerable
in the event of a sudden unexpected change in the environment. Also, the
continuous change in the mission requirements such as inclusion/exclusion of
new actor, changing the actor's priority and the limitations of smart devices
need to be monitored. To perform dynamically in such case, the presented
framework is based on a generic multi-level modeling approach that ensures
adaptation handled by semantic modeling. Automated self-configuration is driven
by rule-based reconfiguration policies through ontology
- …