Skip to main content
Article thumbnail
Location of Repository

Large-Scale Dense 3D Reconstruction from Stereo Imagery

By Pablo F. Alcantarilla, Chris Beall and Frank Dellaert


©2013 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other users, including reprinting/ republishing this material for advertising or promotional purposes, creating new collective works for resale or redistribution to servers or lists, or reuse of any copyrighted components of this work in other works.In this paper we propose a novel method for large-scale dense 3D reconstruction from stereo imagery. Assuming that stereo camera calibration and camera motion are known, our method is able to reconstruct accurately dense 3D models of urban environments in the form of point clouds. We take advantage of recent stereo matching techniques that are able to build dense and accurate disparity maps from two rectified images. Then, we fuse the information from multiple disparity maps into a global model by using an efficient data association technique that takes into account stereo uncertainty and performs geometric and photometric consistency validation in a multi-view setup. Finally, we use efficient voxel grid filtering techniques to deal with storage requirements in large-scale environments. In addition, our method automatically discards possible moving obstacles in the scene. We show experimental results on real video large-scale sequences and compare our approach with respect to other state-of-the-art methods such as PMVS and StereoScan

Topics: Computational complexity, Disparity maps, Egomotion, Image plane, Obstacles, Point clouds, PMVS, Robotics, Stereo imagery, StereoScan, 3D reconstruction, Voxel grid filtering
Publisher: Institute of Electrical and Electronics Engineers
Year: 2013
OAI identifier:
Download PDF:
Sorry, we are unable to provide the full text but you may find it at the following location(s):
  • (external link)
  • Suggested articles

    To submit an update or takedown request for this paper, please submit an Update/Correction/Removal Request.