10,419 research outputs found
EgoFace: Egocentric Face Performance Capture and Videorealistic Reenactment
Face performance capture and reenactment techniques use multiple cameras and sensors, positioned at a distance from the face or mounted on heavy wearable devices. This limits their applications in mobile and outdoor environments. We present EgoFace, a radically new lightweight setup for face performance capture and front-view videorealistic reenactment using a single egocentric RGB camera. Our lightweight setup allows operations in uncontrolled environments, and lends itself to telepresence applications such as video-conferencing from dynamic environments. The input image is projected into a low dimensional latent space of the facial expression parameters. Through careful adversarial training of the parameter-space synthetic rendering, a videorealistic animation is produced. Our problem is challenging as the human visual system is sensitive to the smallest face irregularities that could occur in the final results. This sensitivity is even stronger for video results. Our solution is trained in a pre-processing stage, through a supervised manner without manual annotations. EgoFace captures a wide variety of facial expressions, including mouth movements and asymmetrical expressions. It works under varying illuminations, background, movements, handles people from different ethnicities and can operate in real time
InLoc: Indoor Visual Localization with Dense Matching and View Synthesis
We seek to predict the 6 degree-of-freedom (6DoF) pose of a query photograph
with respect to a large indoor 3D map. The contributions of this work are
three-fold. First, we develop a new large-scale visual localization method
targeted for indoor environments. The method proceeds along three steps: (i)
efficient retrieval of candidate poses that ensures scalability to large-scale
environments, (ii) pose estimation using dense matching rather than local
features to deal with textureless indoor scenes, and (iii) pose verification by
virtual view synthesis to cope with significant changes in viewpoint, scene
layout, and occluders. Second, we collect a new dataset with reference 6DoF
poses for large-scale indoor localization. Query photographs are captured by
mobile phones at a different time than the reference 3D map, thus presenting a
realistic indoor localization scenario. Third, we demonstrate that our method
significantly outperforms current state-of-the-art indoor localization
approaches on this new challenging data
COMPARISON OF LOW COST PHOTOGRAMMETRIC SURVEY WITH TLS AND LEICA PEGASUS BACKPACK 3D MODELS
This paper considers Leica backpack and photogrammetric surveys of a mediaeval bastion in Padua, Italy. Furhtermore, terrestrial
laser scanning (TLS) survey is considered in order to provide a state of the art reconstruction of the bastion. Despite control points
are typically used to avoid deformations in photogrammetric surveys and ensure correct scaling of the reconstruction, in this paper
a different approach is considered: this work is part of a project aiming at the development of a system exploiting ultra-wide band
(UWB) devices to provide correct scaling of the reconstruction. In particular, low cost Pozyx UWB devices are used to estimate
camera positions during image acquisitions. Then, in order to obtain a metric reconstruction, scale factor in the photogrammetric
survey is estimated by comparing camera positions obtained from UWB measurements with those obtained from photogrammetric
reconstruction. Compared with the TLS survey, the considered photogrammetric model of the bastion results in a RMSE of 21.9cm, average error 13.4cm, and standard deviation 13.5cm. Excluding the final part of the bastion left wing, where the presence of several poles make reconstruction more difficult, (RMSE) fitting error is 17.3cm, average error 11.5cm, and standard deviation 9.5cm. Instead, comparison of Leica backpack and TLS surveys leads to an average error of 4.7cm and standard deviation 0.6cm (4.2 cm and 0.3 cm, respectively, by excluding the final part of the left wing)
Comparative Analysis of Mobile 3D Scanning Technologies for Design, Manufacture of Interior and Exterior Tensile Material Structures and Canvasman Ltd. Case Study
This report aimed to investigate mobile 3D Scanning technologies to improve the 3D data capture and efficiency into Canvasman’s CAD design and manufacturing processes with focus on accurate resolution. The Santander funded Collaborative Venture Fund (CVF) project has provided research, survey data, evaluation and analysis for Canvasman Ltd. on 3D portable scanning hardware and software. The project solutions recommended in this report offers impartial product information on the current appropriate 3D scanning technology that potentially could improve efficiency of data capturing, design and manufacture of interior and exterior spaces, boats, vehicles and other similar constructions for creating and installing flexible coverings and indoor and outdoor structures
- …