Skip to main content
Article thumbnail
Location of Repository

A Human-Assisted Approach for a Mobile Robot to Learn 3D Object Models using Active Vision

By Matthijs Zwinderman, Paul E. Rybski and Gert Kootstra

Abstract

Abstract — In this paper we present an algorithm that allows a human to naturally and easily teach a mobile robot how to recognize objects in its environment. The human selects the object by pointing at it using a laser pointer. The robot recognizes the laser reflections with its cameras and uses this data to generate an initial 2D segmentation of the object. The 3D position of SURF feature points are extracted from the designated area using stereo vision. As the robot moves around the object, new views of the object are obtained from which feature points are extracted. These features are filtered using active vision. The complete object representation consists of feature points registered with 3D pose data. We describe the method and show that it works well by performing experiments on real world data collected with our robot. We use an extensive dataset of 21 objects, differing in size, shape and texture. I

Year: 2010
OAI identifier: oai:CiteSeerX.psu:10.1.1.170.4428
Provided by: CiteSeerX
Download PDF:
Sorry, we are unable to provide the full text but you may find it at the following location(s):
  • http://citeseerx.ist.psu.edu/v... (external link)
  • http://www.ri.cmu.edu/pub_file... (external link)
  • http://www.ri.cmu.edu/pub_file... (external link)
  • Suggested articles


    To submit an update or takedown request for this paper, please submit an Update/Correction/Removal Request.