4,110 research outputs found
The Spine of the Cosmic Web
We present the SpineWeb framework for the topological analysis of the Cosmic
Web and the identification of its walls, filaments and cluster nodes. Based on
the watershed segmentation of the cosmic density field, the SpineWeb method
invokes the local adjacency properties of the boundaries between the watershed
basins to trace the critical points in the density field and the separatrices
defined by them. The separatrices are classified into walls and the spine, the
network of filaments and nodes in the matter distribution. Testing the method
with a heuristic Voronoi model yields outstanding results. Following the
discussion of the test results, we apply the SpineWeb method to a set of
cosmological N-body simulations. The latter illustrates the potential for
studying the structure and dynamics of the Cosmic Web.Comment: Accepted for publication HIGH-RES version:
http://skysrv.pha.jhu.edu/~miguel/SpineWeb
Chain Homotopies for Object Topological Representations
This paper presents a set of tools to compute topological information of
simplicial complexes, tools that are applicable to extract topological
information from digital pictures. A simplicial complex is encoded in a
(non-unique) algebraic-topological format called AM-model. An AM-model for a
given object K is determined by a concrete chain homotopy and it provides, in
particular, integer (co)homology generators of K and representative (co)cycles
of these generators. An algorithm for computing an AM-model and the
cohomological invariant HB1 (derived from the rank of the cohomology ring) with
integer coefficients for a finite simplicial complex in any dimension is
designed here. A concept of generators which are "nicely" representative cycles
is also presented. Moreover, we extend the definition of AM-models to 3D binary
digital images and we design algorithms to update the AM-model information
after voxel set operations (union, intersection, difference and inverse)
ScanComplete: Large-Scale Scene Completion and Semantic Segmentation for 3D Scans
We introduce ScanComplete, a novel data-driven approach for taking an
incomplete 3D scan of a scene as input and predicting a complete 3D model along
with per-voxel semantic labels. The key contribution of our method is its
ability to handle large scenes with varying spatial extent, managing the cubic
growth in data size as scene size increases. To this end, we devise a
fully-convolutional generative 3D CNN model whose filter kernels are invariant
to the overall scene size. The model can be trained on scene subvolumes but
deployed on arbitrarily large scenes at test time. In addition, we propose a
coarse-to-fine inference strategy in order to produce high-resolution output
while also leveraging large input context sizes. In an extensive series of
experiments, we carefully evaluate different model design choices, considering
both deterministic and probabilistic models for completion and semantic
inference. Our results show that we outperform other methods not only in the
size of the environments handled and processing efficiency, but also with
regard to completion quality and semantic segmentation performance by a
significant margin.Comment: Video: https://youtu.be/5s5s8iH0NF
STV-based Video Feature Processing for Action Recognition
In comparison to still image-based processes, video features can provide rich and intuitive information about dynamic events occurred over a period of time, such as human actions, crowd behaviours, and other subject pattern changes. Although substantial progresses have been made in the last decade on image processing and seen its successful applications in face matching and object recognition, video-based event detection still remains one of the most difficult challenges in computer vision research due to its complex continuous or discrete input signals, arbitrary dynamic feature definitions, and the often ambiguous analytical methods. In this paper, a Spatio-Temporal Volume (STV) and region intersection (RI) based 3D shape-matching method has been proposed to facilitate the definition and recognition of human actions recorded in videos. The distinctive characteristics and the performance gain of the devised approach stemmed from a coefficient factor-boosted 3D region intersection and matching mechanism developed in this research. This paper also reported the investigation into techniques for efficient STV data filtering to reduce the amount of voxels (volumetric-pixels) that need to be processed in each operational cycle in the implemented system. The encouraging features and improvements on the operational performance registered in the experiments have been discussed at the end
Matterport3D: Learning from RGB-D Data in Indoor Environments
Access to large, diverse RGB-D datasets is critical for training RGB-D scene
understanding algorithms. However, existing datasets still cover only a limited
number of views or a restricted scale of spaces. In this paper, we introduce
Matterport3D, a large-scale RGB-D dataset containing 10,800 panoramic views
from 194,400 RGB-D images of 90 building-scale scenes. Annotations are provided
with surface reconstructions, camera poses, and 2D and 3D semantic
segmentations. The precise global alignment and comprehensive, diverse
panoramic set of views over entire buildings enable a variety of supervised and
self-supervised computer vision tasks, including keypoint matching, view
overlap prediction, normal prediction from color, semantic segmentation, and
region classification
A Survey of Methods for Volumetric Scene Reconstruction from Photographs
Scene reconstruction, the task of generating a 3D model of a scene given multiple 2D photographs taken of the scene, is an old and difficult problem in computer vision. Since its introduction, scene reconstruction has found application in many fields, including robotics, virtual reality, and entertainment. Volumetric models are a natural choice for scene reconstruction. Three broad classes of volumetric reconstruction techniques have been developed based on geometric intersections, color consistency, and pair-wise matching. Some of these techniques have spawned a number of variations and undergone considerable refinement. This paper is a survey of techniques for volumetric scene reconstruction
- …