3,829 research outputs found
Implicit Shape Model Trees: Recognition of 3-D Indoor Scenes and Prediction of Object Poses for Mobile Robots
For a mobile robot, we present an approach to recognize scenes in
arrangements of objects distributed over cluttered environments. Recognition is
made possible by letting the robot alternately search for objects and assign
found objects to scenes. Our scene model "Implicit Shape Model (ISM) trees"
allows us to solve these two tasks together. For the ISM trees, this article
presents novel algorithms for recognizing scenes and predicting the poses of
searched objects. We define scenes as sets of objects, where some objects are
connected by 3-D spatial relations. In previous work, we recognized scenes
using single ISMs. However, these ISMs were prone to false positives. To
address this problem, we introduced ISM trees, a hierarchical model that
includes multiple ISMs. Through the recognition algorithm it contributes, this
article ultimately enables the use of ISM trees in scene recognition. We intend
to enable users to generate ISM trees from object arrangements demonstrated by
humans. The lack of a suitable algorithm is overcome by the introduction of an
ISM tree generation algorithm. In scene recognition, it is usually assumed that
image data is already available. However, this is not always the case for
robots. For this reason, we combined scene recognition and object search in
previous work. However, we did not provide an efficient algorithm to link the
two tasks. This article introduces such an algorithm that predicts the poses of
searched objects with relations. Experiments show that our overall approach
enables robots to find and recognize object arrangements that cannot be
perceived from a single viewpoint.Comment: 22 pages, 24 figures; For associated video clips, see
https://www.youtube.com/playlist?list=PL3RZ_UQY_uOIfuIJNqdS8wDMjTjOAeOm
Towards gestural understanding for intelligent robots
Fritsch JN. Towards gestural understanding for intelligent robots. Bielefeld: Universität Bielefeld; 2012.A strong driving force of scientific progress in the technical sciences is the quest for systems that assist humans in their daily life and make their life easier and more enjoyable. Nowadays smartphones are probably the most typical instances of such systems. Another class of systems that is getting increasing attention are intelligent robots. Instead of offering a smartphone touch screen to select actions, these systems are intended to offer a more natural human-machine interface to their users. Out of the large range of actions performed by humans, gestures performed with the hands play a very important role especially when humans interact with their direct surrounding like, e.g., pointing to an object or manipulating it. Consequently, a robot has to understand such gestures to offer an intuitive interface. Gestural understanding is, therefore, a key capability on the way to intelligent robots.
This book deals with vision-based approaches for gestural understanding. Over the past two decades, this has been an intensive field of research which has resulted in a variety of algorithms to analyze human hand motions. Following a categorization of different gesture types and a review of other sensing techniques, the design of vision systems that achieve hand gesture understanding for intelligent robots is analyzed. For each of the individual algorithmic steps – hand detection, hand tracking, and trajectory-based gesture recognition – a separate Chapter introduces common techniques and algorithms and provides example methods. The resulting recognition algorithms are considering gestures in isolation and are often not sufficient for interacting with a robot who can only understand such gestures when incorporating the context like, e.g., what object was pointed at or manipulated.
Going beyond a purely trajectory-based gesture recognition by incorporating context is an important prerequisite to achieve gesture understanding and is addressed explicitly in a separate Chapter of this book. Two types of context, user-provided context and situational context, are reviewed and existing approaches to incorporate context for gestural understanding are reviewed. Example approaches for both context types provide a deeper algorithmic insight into this field of research. An overview of recent robots capable of gesture recognition and understanding summarizes the currently realized human-robot interaction quality.
The approaches for gesture understanding covered in this book are manually designed while humans learn to recognize gestures automatically during growing up. Promising research targeted at analyzing developmental learning in children in order to mimic this capability in technical systems is highlighted in the last Chapter completing this book as this research direction may be highly influential for creating future gesture understanding systems
NASA space station automation: AI-based technology review
Research and Development projects in automation for the Space Station are discussed. Artificial Intelligence (AI) based automation technologies are planned to enhance crew safety through reduced need for EVA, increase crew productivity through the reduction of routine operations, increase space station autonomy, and augment space station capability through the use of teleoperation and robotics. AI technology will also be developed for the servicing of satellites at the Space Station, system monitoring and diagnosis, space manufacturing, and the assembly of large space structures
From surfaces to objects : Recognizing objects using surface information and object models.
This thesis describes research on recognizing partially obscured objects using
surface information like Marr's 2D sketch ([MAR82]) and surface-based geometrical
object models. The goal of the recognition process is to produce a fully
instantiated object hypotheses, with either image evidence for each feature or
explanations for their absence, in terms of self or external occlusion.
The central point of the thesis is that using surface information should be
an important part of the image understanding process. This is because surfaces
are the features that directly link perception to the objects perceived (for
normal "camera-like" sensing) and because surfaces make explicit information
needed to understand and cope with some visual problems (e.g. obscured features).
Further, because surfaces are both the data and model primitive, detailed
recognition can be made both simpler and more complete.
Recognition input is a surface image, which represents surface orientation and
absolute depth. Segmentation criteria are proposed for forming surface patches
with constant curvature character, based on surface shape discontinuities which
become labeled segmentation- boundaries.
Partially obscured object surfaces are reconstructed using stronger surface based
constraints. Surfaces are grouped to form surface clusters, which are 3D
identity-independent solids that often correspond to model primitives. These are
used here as a context within which to select models and find all object features.
True three-dimensional properties of image boundaries, surfaces and surface
clusters are directly estimated using the surface data.
Models are invoked using a network formulation, where individual nodes
represent potential identities for image structures. The links between nodes are
defined by generic and structural relationships. They define indirect evidence relationships
for an identity. Direct evidence for the identities comes from the data
properties. A plausibility computation is defined according to the constraints inherent
in the evidence types. When a node acquires sufficient plausibility, the
model is invoked for the corresponding image structure.Objects are primarily represented using a surface-based geometrical model.
Assemblies are formed from subassemblies and surface primitives, which are
defined using surface shape and boundaries. Variable affixments between assemblies
allow flexibly connected objects.
The initial object reference frame is estimated from model-data surface relationships,
using correspondences suggested by invocation. With the reference
frame, back-facing, tangential, partially self-obscured, totally self-obscured and
fully visible image features are deduced. From these, the oriented model is used
for finding evidence for missing visible model features. IT no evidence is found,
the program attempts to find evidence to justify the features obscured by an unrelated
object. Structured objects are constructed using a hierarchical synthesis
process.
Fully completed hypotheses are verified using both existence and identity
constraints based on surface evidence.
Each of these processes is defined by its computational constraints and are
demonstrated on two test images. These test scenes are interesting because they
contain partially and fully obscured object features, a variety of surface and solid
types and flexibly connected objects. All modeled objects were fully identified
and analyzed to the level represented in their models and were also acceptably
spatially located.
Portions of this work have been reported elsewhere ([FIS83], [FIS85a], [FIS85b],
[FIS86]) by the author
Robot skill learning through human demonstration and interaction
Nowadays robots are increasingly involved in more complex and less structured tasks. Therefore, it is highly desirable to develop new approaches to fast robot skill acquisition. This research is aimed to develop an overall framework for robot skill learning through human demonstration and interaction. Through low-level demonstration and interaction with humans, the robot can learn basic skills. These basic skills are treated as primitive actions. In high-level learning, the complex skills demonstrated by the human can be automatically translated into skill scripts which are executed by the robot. This dissertation summarizes my major research activities in robot skill learning. First, a framework for Programming by Demonstration (PbD) with reinforcement learning for human-robot collaborative manipulation tasks is described. With this framework, the robot can learn low level skills such as collaborating with a human to lift a table successfully and efficiently. Second, to develop a high-level skill acquisition system, we explore the use of a 3D sensor to recognize human actions. A Kinect based action recognition system is implemented which considers both object/action dependencies and the sequential constraints. Third, we extend the action recognition framework by fusing information from multimodal sensors which can recognize fine assembly actions. Fourth, a Portable Assembly Demonstration (PAD) system is built which can automatically generate skill scripts from human demonstration. The skill script includes the object type, the tool, the action used, and the assembly state. Finally, the generated skill scripts are implemented by a dual-arm robot. The proposed framework was experimentally evaluated
- …