5,520 research outputs found
Camera Calibration from Dynamic Silhouettes Using Motion Barcodes
Computing the epipolar geometry between cameras with very different
viewpoints is often problematic as matching points are hard to find. In these
cases, it has been proposed to use information from dynamic objects in the
scene for suggesting point and line correspondences.
We propose a speed up of about two orders of magnitude, as well as an
increase in robustness and accuracy, to methods computing epipolar geometry
from dynamic silhouettes. This improvement is based on a new temporal
signature: motion barcode for lines. Motion barcode is a binary temporal
sequence for lines, indicating for each frame the existence of at least one
foreground pixel on that line. The motion barcodes of two corresponding
epipolar lines are very similar, so the search for corresponding epipolar lines
can be limited only to lines having similar barcodes. The use of motion
barcodes leads to increased speed, accuracy, and robustness in computing the
epipolar geometry.Comment: Update metadat
Capturing natural-colour 3D models of insects for species discovery
Collections of biological specimens are fundamental to scientific
understanding and characterization of natural diversity. This paper presents a
system for liberating useful information from physical collections by bringing
specimens into the digital domain so they can be more readily shared, analyzed,
annotated and compared. It focuses on insects and is strongly motivated by the
desire to accelerate and augment current practices in insect taxonomy which
predominantly use text, 2D diagrams and images to describe and characterize
species. While these traditional kinds of descriptions are informative and
useful, they cannot cover insect specimens "from all angles" and precious
specimens are still exchanged between researchers and collections for this
reason. Furthermore, insects can be complex in structure and pose many
challenges to computer vision systems. We present a new prototype for a
practical, cost-effective system of off-the-shelf components to acquire
natural-colour 3D models of insects from around 3mm to 30mm in length. Colour
images are captured from different angles and focal depths using a digital
single lens reflex (DSLR) camera rig and two-axis turntable. These 2D images
are processed into 3D reconstructions using software based on a visual hull
algorithm. The resulting models are compact (around 10 megabytes), afford
excellent optical resolution, and can be readily embedded into documents and
web pages, as well as viewed on mobile devices. The system is portable, safe,
relatively affordable, and complements the sort of volumetric data that can be
acquired by computed tomography. This system provides a new way to augment the
description and documentation of insect species holotypes, reducing the need to
handle or ship specimens. It opens up new opportunities to collect data for
research, education, art, entertainment, biodiversity assessment and
biosecurity control.Comment: 24 pages, 17 figures, PLOS ONE journa
Recommended from our members
Image-Based 3D Photography Using Opacity Hulls
We have built a system for acquiring and displaying high quality graphical models of objects that are impossible to scan with traditional scanners. Our system can acquire highly specular and fuzzy materials, such as fur and feathers. The hardware set-up consists of a turntable, two plasma displays, an array of cameras, and a rotating array of directional lights. We use multi-background matting techniques to acquire alpha mattes of the object from multiple viewpoints. The alpha mattes are used to construct an opacity hull. The opacity hull is a new shape representation, defined as the visual hull of the object with view-dependent opacity. It enables visualization of complex object silhouettes and seamless blending of objects into new environments. Our system also supports relighting of objects with arbitrary appearance using surface reflectance fields, a purely image-based appearance representation. Our system is the first to acquire and render surface reflectance fields under varying illumination from arbitrary viewpoints. We have built three generations of digitizers with increasing sophistication. In this paper, we present our results from digitizing hundreds of models.Engineering and Applied Science
3D Object Reconstruction using Multi-View Calibrated Images
In this study, two models are proposed, one is a visual hull model and another one is a 3D object reconstruction model. The proposed visual hull model, which is based on bounding edge representation, obtains high time performance which makes it to be one of the best methods. The main contribution of the proposed visual hull model is to provide bounding surfaces over the bounding edges, which results a complete triangular surface mesh. Moreover, the proposed visual hull model can be computed over the camera networks distributedly. The second model is a depth map based 3D object reconstruction model which results a watertight triangular surface mesh. The proposed model produces the result with acceptable accuracy as well as high completeness, only using stereo matching and triangulation. The contribution of this model is to playing with the 3D points to find the best reliable ones and fitting a surface over them
Recommended from our members
Opacity Light Fields: Interactive Rendering of Surface Light Fields with View-Dependent Opacity
We present new hardware-accelerated techniques for rendering surface light fields with opacity hulls that allow for interactive visualization of objects that have complex reflectance properties and elaborate geometrical details. The opacity hull is a shape enclosing the object with view-dependent opacity parameterized onto that shape. We call the combination of opacity hulls and surface light fields the opacity light field. Opacity light fields are ideally suited for rendering of the visually complex objects and scenes obtained with 3D photography. We show how to implement opacity light fields in the framework of three surface light field rendering methods: view-dependent texture mapping, unstructured lumigraph rendering, and light field mapping. The modified algorithms can be effectively supported on modern graphics hardware. Our results show that all three implementations are able to achieve interactive or real-time frame rates.Engineering and Applied Science
OSGAR: a scene graph with uncertain transformations
An important problem for augmented reality is registration error. No system can be perfectly tracked, calibrated or modeled. As a result, the overlaid graphics are not aligned perfectly with objects in the physical world. This can be distracting, annoying or confusing. In this paper, we propose a method for mitigating the effects of registration errors that enables application developers to build dynamically adaptive AR displays. Our solution is implemented in a programming toolkit called OSGAR. Built upon OpenSceneGraph (OSG), OSGAR statistically characterizes registration errors, monitors those errors and, when a set of criteria are met, dynamically adapts the display to mitigate the effects of the errors. Because the architecture is based on a scene graph, it provides a simple, familiar and intuitive environment for application developers. We describe the components of OSGAR, discuss how several proposed methods for error registration can be implemented, and illustrate its use through a set of examples
- …