14,949 research outputs found
The 12 prophets dataset
The "Ajeijadinho 3D" project is an initiative supported by the University of
S\~ao Paulo (Museum of Science and Dean of Culture and Extension), which
involves the 3D digitization of art works of Brazilian sculptor Antonio
Francisco Lisboa, better known as Aleijadinho. The project made use of advanced
acquisition and processing of 3D meshes for preservation and dissemination of
the cultural heritage. The dissemination occurs through a Web portal, so that
the population has the opportunity to meet the art works in detail using 3D
visualization and interaction. The portal address is
http://www.aleijadinho3d.icmc.usp.br. The 3D acquisitions were conducted over a
week at the end of July 2013 in the cities of Ouro Preto, MG, Brazil and
Congonhas do Campo, MG, Brazil. The scanning was done with a special equipment
supplied by company Leica Geosystems, which allowed the work to take place at
distances between 10 and 30 meters, defining a non-invasive procedure,
simplified logistics, and without the need for preparation or isolation of the
sites. In Ouro Preto, we digitized the churches of Francisco of Assis, Our Lady
of Carmo, and Our Lady of Mercy; in Congonhas do Campo we scanned the entire
Sanctuary of Bom Jesus de Matosinhos and his 12 prophets. Once scanned, the art
works went through a long process of preparation, which required careful
handling of meshes done by experts from the University of S\~ao Paulo in
partnership with company Imprimate.Comment: Full dataset online at http://aleijadinho3d.icmc.usp.br/data.htm
Prototyping Information Visualization in 3D City Models: a Model-based Approach
When creating 3D city models, selecting relevant visualization techniques is
a particularly difficult user interface design task. A first obstacle is that
current geodata-oriented tools, e.g. ArcGIS, have limited 3D capabilities and
limited sets of visualization techniques. Another important obstacle is the
lack of unified description of information visualization techniques for 3D city
models. If many techniques have been devised for different types of data or
information (wind flows, air quality fields, historic or legal texts, etc.)
they are generally described in articles, and not really formalized. In this
paper we address the problem of visualizing information in (rich) 3D city
models by presenting a model-based approach for the rapid prototyping of
visualization techniques. We propose to represent visualization techniques as
the composition of graph transformations. We show that these transformations
can be specified with SPARQL construction operations over RDF graphs. These
specifications can then be used in a prototype generator to produce 3D scenes
that contain the 3D city model augmented with data represented using the
desired technique.Comment: Proc. of 3DGeoInfo 2014 Conference, Dubai, November 201
Advanced Mid-Water Tools for 4D Marine Data Fusion and Visualization
Mapping and charting of the seafloor underwent a revolution approximately 20 years ago with the introduction of multibeam sonars -- sonars that provided complete, high-resolution coverage of the seafloor rather than sparse measurements. The initial focus of these sonar systems was the charting of depths in support of safety of navigation and offshore exploration; more recently innovations in processing software have led to approaches to characterize seafloor type and for mapping seafloor habitat in support of fisheries research. In recent years, a new generation of multibeam sonars has been developed that, for the first time, have the ability to map the water column along with the seafloor. This ability will potentially allow multibeam sonars to address a number of critical ocean problems including the direct mapping of fish and marine mammals, the location of mid-water targets and, if water column properties are appropriate, a wide range of physical oceanographic processes. This potential relies on suitable software to make use of all of the new available data. Currently, the users of these sonars have a limited view of the mid-water data in real-time and limited capacity to store it, replay it, or run further analysis. The data also needs to be integrated with other sensor assets such as bathymetry, backscatter, sub-bottom, seafloor characterizations and other assets so that a “complete” picture of the marine environment under analysis can be realized. Software tools developed for this type of data integration should support a wide range of sonars with a unified format for the wide variety of mid-water sonar types. This paper describes the evolution and result of an effort to create a software tool that meets these needs, and details case studies using the new tools in the areas of fisheries research, static target search, wreck surveys and physical oceanographic processes
Matterport3D: Learning from RGB-D Data in Indoor Environments
Access to large, diverse RGB-D datasets is critical for training RGB-D scene
understanding algorithms. However, existing datasets still cover only a limited
number of views or a restricted scale of spaces. In this paper, we introduce
Matterport3D, a large-scale RGB-D dataset containing 10,800 panoramic views
from 194,400 RGB-D images of 90 building-scale scenes. Annotations are provided
with surface reconstructions, camera poses, and 2D and 3D semantic
segmentations. The precise global alignment and comprehensive, diverse
panoramic set of views over entire buildings enable a variety of supervised and
self-supervised computer vision tasks, including keypoint matching, view
overlap prediction, normal prediction from color, semantic segmentation, and
region classification
Integrating data from 3D CAD and 3D cameras for Real-Time Modeling
In a reversal of historic trends, the capital facilities industry is expressing an increasing desire for automation of equipment and construction processes. Simultaneously, the industry has become conscious that higher levels of interoperability are a key towards higher productivity and safer projects. In complex, dynamic, and rapidly changing three-dimensional (3D) environments such as facilities sites, cutting-edge 3D sensing technologies and processing algorithms are one area of development that can dramatically impact those projects factors. New 3D technologies are now being developed, with among them 3D camera. The main focus here is an investigation of the feasibility of rapidly combining and comparing – integrating – 3D sensed data (from a 3D camera) and 3D CAD data. Such a capability could improve construction quality assessment, facility aging assessment, as well as rapid environment reconstruction and construction automation. Some preliminary results are presented here. They deal with the challenge of fusing sensed and CAD data that are completely different in nature
Integrating 2D Mouse Emulation with 3D Manipulation for Visualizations on a Multi-Touch Table
We present the Rizzo, a multi-touch virtual mouse that has been designed to provide the fine grained interaction for information visualization on a multi-touch table. Our solution enables touch interaction for existing mouse-based visualizations. Previously, this transition to a multi-touch environment was difficult because the mouse emulation of touch surfaces is often insufficient to provide full information visualization functionality. We present a unified design, combining many Rizzos that have been designed not only to provide mouse capabilities but also to act as zoomable lenses that make precise information access feasible. The Rizzos and the information visualizations all exist within a touch-enabled 3D window management system. Our approach permits touch interaction with both the 3D windowing environment as well as with the contents of the individual windows contained therein. We describe an implementation of our technique that augments the VisLink 3D visualization environment to demonstrate how to enable multi-touch capabilities on all visualizations written with the popular prefuse visualization toolkit.
- …