9 research outputs found
NeRFtrinsic Four: An End-To-End Trainable NeRF Jointly Optimizing Diverse Intrinsic and Extrinsic Camera Parameters
Novel view synthesis using neural radiance fields (NeRF) is the
state-of-the-art technique for generating high-quality images from novel
viewpoints. Existing methods require a priori knowledge about extrinsic and
intrinsic camera parameters. This limits their applicability to synthetic
scenes, or real-world scenarios with the necessity of a preprocessing step.
Current research on the joint optimization of camera parameters and NeRF
focuses on refining noisy extrinsic camera parameters and often relies on the
preprocessing of intrinsic camera parameters. Further approaches are limited to
cover only one single camera intrinsic. To address these limitations, we
propose a novel end-to-end trainable approach called NeRFtrinsic Four. We
utilize Gaussian Fourier features to estimate extrinsic camera parameters and
dynamically predict varying intrinsic camera parameters through the supervision
of the projection error. Our approach outperforms existing joint optimization
methods on LLFF and BLEFF. In addition to these existing datasets, we introduce
a new dataset called iFF with varying intrinsic camera parameters. NeRFtrinsic
Four is a step forward in joint optimization NeRF-based view synthesis and
enables more realistic and flexible rendering in real-world scenarios with
varying camera parameters
DynaMoN: Motion-Aware Fast And Robust Camera Localization for Dynamic NeRF
Dynamic reconstruction with neural radiance fields (NeRF) requires accurate
camera poses. These are often hard to retrieve with existing
structure-from-motion (SfM) pipelines as both camera and scene content can
change. We propose DynaMoN that leverages simultaneous localization and mapping
(SLAM) jointly with motion masking to handle dynamic scene content. Our robust
SLAM-based tracking module significantly accelerates the training process of
the dynamic NeRF while improving the quality of synthesized views at the same
time. Extensive experimental validation on TUM RGB-D, BONN RGB-D Dynamic and
the DyCheck's iPhone dataset, three real-world datasets, shows the advantages
of DynaMoN both for camera pose estimation and novel view synthesis.Comment: 6 pages, 4 figure
HouseCat6D -- A Large-Scale Multi-Modal Category Level 6D Object Pose Dataset with Household Objects in Realistic Scenarios
Estimating the 6D pose of objects is a major 3D computer vision problem.
Since the promising outcomes from instance-level approaches, research heads
also move towards category-level pose estimation for more practical application
scenarios. However, unlike well-established instance-level pose datasets,
available category-level datasets lack annotation quality and provided pose
quantity. We propose the new category-level 6D pose dataset HouseCat6D
featuring 1) Multi-modality of Polarimetric RGB and Depth (RGBD+P), 2) Highly
diverse 194 objects of 10 household object categories including 2
photometrically challenging categories, 3) High-quality pose annotation with an
error range of only 1.35 mm to 1.74 mm, 4) 41 large-scale scenes with extensive
viewpoint coverage and occlusions, 5) Checkerboard-free environment throughout
the entire scene, and 6) Additionally annotated dense 6D parallel-jaw grasps.
Furthermore, we also provide benchmark results of state-of-the-art
category-level pose estimation networks
Deep Sensor Fusion with Pyramid Fusion Networks for 3D Semantic Segmentation
Robust environment perception for autonomous vehicles is a tremendous
challenge, which makes a diverse sensor set with e.g. camera, lidar and radar
crucial. In the process of understanding the recorded sensor data, 3D semantic
segmentation plays an important role. Therefore, this work presents a
pyramid-based deep fusion architecture for lidar and camera to improve 3D
semantic segmentation of traffic scenes. Individual sensor backbones extract
feature maps of camera images and lidar point clouds. A novel Pyramid Fusion
Backbone fuses these feature maps at different scales and combines the
multimodal features in a feature pyramid to compute valuable multimodal,
multi-scale features. The Pyramid Fusion Head aggregates these pyramid features
and further refines them in a late fusion step, incorporating the final
features of the sensor backbones. The approach is evaluated on two challenging
outdoor datasets and different fusion strategies and setups are investigated.
It outperforms recent range view based lidar approaches as well as all so far
proposed fusion strategies and architectures.Comment: conditionally accepted at IEEE IV 2022, 7 pages, 4 figures, 5 table
A case study of microbial mat-related features in coastal epeiric sandstones from the Paleoproterozoic Pretoria Group (Transvaal Supergroup, Kaapvaal craton, South Africa); The effect of preservation (reflecting sequence stratigraphic models) on the relationship between mat features and inferred paleoenvironment
Isotope ratios of H, C, and O in CO2 and H2O of the Martian atmosphere
Stable isotope ratios of H, C, and O are powerful indicators of a wide variety of planetary geophysical processes, and for Mars they reveal the record of loss of its atmosphere and subsequent interactions with its surface such as carbonate formation. We report in situ measurements of the isotopic ratios of D/H and O-18/O-16 in water and C-13/C-12, O-18/O-16, O-17/O-16, and (CO)-C-13-O-18/(CO)-C-12-O-16 in carbon dioxide, made in the martian atmosphere at Gale Crater from the Curiosity rover using the Sample Analysis at Mars (SAM)'s tunable laser spectrometer (TLS). Comparison between our measurements in the modern atmosphere and those of martian meteorites such as ALH 84001 implies that the martian reservoirs of CO2 and H2O were largely established similar to 4 billion years ago, but that atmospheric loss or surface interaction may be still ongoing