10 research outputs found
Event Fusion Photometric Stereo Network
We present a novel method to estimate the surface normal of an object in an
ambient light environment using RGB and event cameras. Modern photometric
stereo methods rely on an RGB camera, mainly in a dark room, to avoid ambient
illumination. To alleviate the limitations of the darkroom environment and to
use essential light information, we employ an event camera with a high dynamic
range and low latency. This is the first study that uses an event camera for
the photometric stereo task, which works on continuous light sources and
ambient light environment. In this work, we also curate a novel photometric
stereo dataset that is constructed by capturing objects with event and RGB
cameras under numerous ambient lights environment. Additionally, we propose a
novel framework named Event Fusion Photometric Stereo Network~(EFPS-Net), which
estimates the surface normals of an object using both RGB frames and event
signals. Our proposed method interpolates event observation maps that generate
light information with sparse event signals to acquire fluent light
information. Subsequently, the event-interpolated observation maps are fused
with the RGB observation maps. Our numerous experiments showed that EFPS-Net
outperforms state-of-the-art methods on a dataset captured in the real world
where ambient lights exist. Consequently, we demonstrate that incorporating
additional modalities with EFPS-Net alleviates the limitations that occurred
from ambient illumination.Comment: 33 pages, 11 figure
Differentiable Display Photometric Stereo
Photometric stereo leverages variations in illumination conditions to
reconstruct per-pixel surface normals. The concept of display photometric
stereo, which employs a conventional monitor as an illumination source, has the
potential to overcome limitations often encountered in bulky and
difficult-to-use conventional setups. In this paper, we introduce
Differentiable Display Photometric Stereo (DDPS), a method designed to achieve
high-fidelity normal reconstruction using an off-the-shelf monitor and camera.
DDPS addresses a critical yet often neglected challenge in photometric stereo:
the optimization of display patterns for enhanced normal reconstruction. We
present a differentiable framework that couples basis-illumination image
formation with a photometric-stereo reconstruction method. This facilitates the
learning of display patterns that leads to high-quality normal reconstruction
through automatic differentiation. Addressing the synthetic-real domain gap
inherent in end-to-end optimization, we propose the use of a real-world
photometric-stereo training dataset composed of 3D-printed objects. Moreover,
to reduce the ill-posed nature of photometric stereo, we exploit the linearly
polarized light emitted from the monitor to optically separate diffuse and
specular reflections in the captured images. We demonstrate that DDPS allows
for learning display patterns optimized for a target configuration and is
robust to initialization. We assess DDPS on 3D-printed objects with
ground-truth normals and diverse real-world objects, validating that DDPS
enables effective photometric-stereo reconstruction
ORA3D: Overlap Region Aware Multi-view 3D Object Detection
Current multi-view 3D object detection methods often fail to detect objects
in the overlap region properly, and the networks' understanding of the scene is
often limited to that of a monocular detection network. Moreover, objects in
the overlap region are often largely occluded or suffer from deformation due to
camera distortion, causing a domain shift. To mitigate this issue, we propose
using the following two main modules: (1) Stereo Disparity Estimation for Weak
Depth Supervision and (2) Adversarial Overlap Region Discriminator. The former
utilizes the traditional stereo disparity estimation method to obtain reliable
disparity information from the overlap region. Given the disparity estimates as
supervision, we propose regularizing the network to fully utilize the geometric
potential of binocular images and improve the overall detection accuracy
accordingly. Further, the latter module minimizes the representational gap
between non-overlap and overlapping regions. We demonstrate the effectiveness
of the proposed method with the nuScenes large-scale multi-view 3D object
detection data. Our experiments show that our proposed method outperforms
current state-of-the-art models, i.e., DETR3D and BEVDet.Comment: BMVC202
InterHandGen: Two-Hand Interaction Generation via Cascaded Reverse Diffusion
We present InterHandGen, a novel framework that learns the generative prior
of two-hand interaction. Sampling from our model yields plausible and diverse
two-hand shapes in close interaction with or without an object. Our prior can
be incorporated into any optimization or learning methods to reduce ambiguity
in an ill-posed setup. Our key observation is that directly modeling the joint
distribution of multiple instances imposes high learning complexity due to its
combinatorial nature. Thus, we propose to decompose the modeling of joint
distribution into the modeling of factored unconditional and conditional single
instance distribution. In particular, we introduce a diffusion model that
learns the single-hand distribution unconditional and conditional to another
hand via conditioning dropout. For sampling, we combine anti-penetration and
classifier-free guidance to enable plausible generation. Furthermore, we
establish the rigorous evaluation protocol of two-hand synthesis, where our
method significantly outperforms baseline generative models in terms of
plausibility and diversity. We also demonstrate that our diffusion prior can
boost the performance of two-hand reconstruction from monocular in-the-wild
images, achieving new state-of-the-art accuracy.Comment: Accepted to CVPR 2024, project page:
https://jyunlee.github.io/projects/interhandgen
Progressive Acquisition of SVBRDF and Shape in Motion
To estimate appearance parameters, traditional SVBRDF acquisition methods require multiple input images to be captured with various angles of light and camera, followed by a post-processing step. For this reason, subjects have been limited to static scenes, or a multiview system is required to capture dynamic objects. In this paper, we propose a simultaneous acquisition method of SVBRDF and shape allowing us to capture the material appearance of deformable objects in motion using a single RGBD camera. To do so, we progressively integrate photometric samples of surfaces in motion in a volumetric data structure with a deformation graph. Then, building upon recent advances of fusion-based methods, we estimate SVBRDF parameters in motion. We make use of a conventional RGBD camera that consists of the colour and infrared cameras with active infrared illumination. The colour camera is used for capturing diffuse properties, and the infrared camera-illumination module is employed for estimating specular properties by means of active illumination. Our joint optimization yields complete material appearance parameters. We demonstrate the effectiveness of our method with extensive evaluation on both synthetic and real data that include various deformable objects of specular and diffuse appearance.11Nsciescopu
Image-Based Acquisition and Modeling of Polarimetric Reflectance
Realistic modeling of the bidirectional reflectance distribution function (BRDF) of scene objects is a vital prerequisite for any type of physically based rendering. In the last decades, the availability of databases containing real-world material measurements has fueled considerable innovation in the development of such models. However, previous work in this area was mainly focused on increasing the visual realism of images, and hence ignored the effect of scattering on the polarization state of light, which is normally imperceptible to the human eye. Existing databases thus only capture scattered flux, or polarimetric BRDF datasets are too directionally sparse (e.g., in-plane) to be usable for simulation. While subtle to human observers, polarization is easily perceived by any optical sensor (e.g., using polarizing filters), providing a wealth of additional information about shape and material properties of the object under observation. Given the increasing application of rendering in the solution of inverse problems via analysis-by-synthesis and differentiation, the ability to realistically model polarized radiative transport is thus highly desirable. Polarization depends on the wavelength of the spectrum, and thus we provide the first polarimetric BRDF (pBRDF) dataset that captures the polarimetric properties of real-world materials over the full angular domain, and at multiple wavelengths. Acquisition of such reflectance data is challenging due to the extremely large space of angular, spectral, and polarimetric configurations that must be observed, and we propose a scheme combining image-based acquisition with spectroscopic ellipsometry to perform measurements in a realistic amount of time. This process yields raw Mueller matrices, which we subsequently transform into Rusinkiewicz-parameterized pBRDFs that can be used for rendering. Our dataset provides 25 isotropic pBRDFs spanning a wide range of appearances: diffuse/specular, metallic/dielectric, rough/smooth, and different color albedos, captured in five wavelength ranges covering the visible spectrum. We demonstrate usage of our data-driven pBRDF model in a physically based renderer that accounts for polarized interreflection, and we investigate the relationship of polarization and material appearance, providing insights into the behavior of characteristic real-world pBRDFs.11Nsciescopu