900 research outputs found
Integration of optical and acoustic sensors for D underwater scene reconstruction.
Combination of optical and acoustic sensors to overcome the shortcomings
presented by optical systems in underwater 3D acquisition is an emerging field
of research. In this work, an opti-acoustic system composed by a single camera and
a multibeam sonar is proposed, providing a simulation environment to validate its
potential use in 3D reconstruction. Since extrinsic calibration is a prerequisite for this
kind of feature-level sensor fusion, an effective approach to address the calibration
problem between a multibeam and a camera system is presented.Peer Reviewe
Underwater reconstruction using depth sensors
In this paper we describe experiments in which we acquire range images of underwater surfaces with four types of depth sensors and attempt to reconstruct underwater surfaces. Two conditions are tested: acquiring range images by submersing the sensors and by holding the sensors over the water line and recording through water. We found out that only the Kinect sensor is able to acquire depth images of submersed surfaces by holding the sensor above water. We compare the reconstructed underwater geometry with meshes obtained when the surfaces were not submersed. These findings show that 3D underwater reconstruction using depth sensors is possible, despite the high water absorption of the near infrared spectrum in which these sensors operate
3D reconstruction and motion estimation using forward looking sonar
Autonomous Underwater Vehicles (AUVs) are increasingly used in different domains
including archaeology, oil and gas industry, coral reef monitoring, harbourâs security,
and mine countermeasure missions. As electromagnetic signals do not penetrate
underwater environment, GPS signals cannot be used for AUV navigation, and optical
cameras have very short range underwater which limits their use in most underwater
environments.
Motion estimation for AUVs is a critical requirement for successful vehicle recovery
and meaningful data collection. Classical inertial sensors, usually used for AUV motion
estimation, suffer from large drift error. On the other hand, accurate inertial sensors are
very expensive which limits their deployment to costly AUVs. Furthermore, acoustic
positioning systems (APS) used for AUV navigation require costly installation and
calibration. Moreover, they have poor performance in terms of the inferred resolution.
Underwater 3D imaging is another challenge in AUV industry as 3D information is
increasingly demanded to accomplish different AUV missions. Different systems have
been proposed for underwater 3D imaging, such as planar-array sonar and T-configured
3D sonar. While the former features good resolution in general, it is very expensive and
requires huge computational power, the later is cheaper implementation but requires
long time for full 3D scan even in short ranges.
In this thesis, we aim to tackle AUV motion estimation and underwater 3D imaging by
proposing relatively affordable methodologies and study different parameters affecting
their performance. We introduce a new motion estimation framework for AUVs which
relies on the successive acoustic images to infer AUV ego-motion. Also, we propose an
Acoustic Stereo Imaging (ASI) system for underwater 3D reconstruction based on
forward looking sonars; the proposed system features cheaper implementation than
planar array sonars and solves the delay problem in T configured 3D sonars
Map building fusing acoustic and visual information using autonomous underwater vehicles
Author Posting. © The Author(s), 2012. This is the author's version of the work. It is posted here by permission of John Wiley & Sons for personal use, not for redistribution. The definitive version was published in Journal of Field Robotics 30 (2013): 763â783, doi:10.1002/rob.21473.We present a system for automatically building 3-D maps of underwater terrain fusing
visual data from a single camera with range data from multibeam sonar. The six-degree
of freedom location of the camera relative to the navigation frame is derived as part of the
mapping process, as are the attitude offsets of the multibeam head and the on-board velocity
sensor. The system uses pose graph optimization and the square root information smoothing
and mapping framework to simultaneously solve for the robotâs trajectory, the map, and
the camera location in the robotâs frame. Matched visual features are treated within the
pose graph as images of 3-D landmarks, while multibeam bathymetry submap matches are
used to impose relative pose constraints linking robot poses from distinct tracklines of the
dive trajectory. The navigation and mapping system presented works under a variety of
deployment scenarios, on robots with diverse sensor suites. Results of using the system to
map the structure and appearance of a section of coral reef are presented using data acquired
by the Seabed autonomous underwater vehicle.The work described herein was funded by the National Science Foundation Censsis ERC under grant number
EEC-9986821, and by the National Oceanic and Atmospheric Administration under grant number
NA090AR4320129
High-resolution underwater robotic vision-based mapping and three-dimensional reconstruction for archaeology
Documenting underwater archaeological sites is an extremely challenging problem. Sites covering large areas are particularly daunting for traditional techniques. In this paper, we present a novel approach to this problem using both an autonomous underwater vehicle (AUV) and a diver-controlled stereo imaging platform to document the submerged Bronze Age city at Pavlopetri, Greece. The result is a three-dimensional (3D) reconstruction covering 26,600 m2 at a resolution of 2 mm/pixel, the largest-scale underwater optical 3D map, at such a resolution, in the world to date. We discuss the advances necessary to achieve this result, including i) an approach to color correct large numbers of images at varying altitudes and over varying bottom types; ii) a large-scale bundle adjustment framework that is capable of handling upward of 400,000 stereo images; and iii) a novel approach to the registration and rapid documentation of an underwater excavations area that can quickly produce maps of site change. We present visual and quantitative comparisons to the authors' previous underwater mapping approaches
The Geometry and Usage of the Supplementary Fisheye Lenses in Smartphones
Nowadays, mobile phones are more than a device that can only satisfy the communication need between people. Since fisheye lenses integrated with mobile phones are lightweight and easy to use, they are advantageous. In addition to this advantage, it is experimented whether fisheye lens and mobile phone combination can be used in a photogrammetric way, and if so, what will be the result. Fisheye lens equipment used with mobile phones was tested in this study. For this, standard calibration of âOlloclip 3 in oneâ fisheye lens used with iPhone 4S mobile phone and âNikon FCâE9â fisheye lens used with Nikon Coolpix8700 are compared based on equidistant model. This experimental study shows that Olloclip 3 in one fisheye lens developed for mobile phones has at least the similar characteristics with classic fisheye lenses. The dimensions of fisheye lenses used with smart phones are getting smaller and the prices are reducing. Moreover, as verified in this study, the accuracy of fisheye lenses used in smartphones is better than conventional fisheye lenses. The use of smartphones with fisheye lenses will give the possibility of practical applications to ordinary users in the near future
- âŠ