55 research outputs found

    Procedures for condition mapping using 360° images

    Get PDF
    The identification of deterioration mechanisms and their monitoring over time is an essential phase for conservation. This work aimed at developing a novel approach for deterioration mapping and monitoring based on 360° images, which allows for simple and rapid data collection. The opportunity to capture the whole scene around a 360° camera reduces the number of images needed in a condition mapping project, resulting in a powerful solution to document small and narrow spaces. The paper will describe the implemented workflow for deterioration mapping based on 360° images, which highlights pathologies on surfaces and quantitatively measures their extension. Such a result will be available as standard outputs as well as an innovative virtual environment for immersive visualization. The case of multi-temporal data acquisition will be considered and discussed as well. Multiple 360° images acquired at different epochs from slightly different points are co-registered to obtain pixel-to-pixel correspondence, providing a solution to quantify and track deterioration effects

    Simplicial Complex based Point Correspondence between Images warped onto Manifolds

    Full text link
    Recent increase in the availability of warped images projected onto a manifold (e.g., omnidirectional spherical images), coupled with the success of higher-order assignment methods, has sparked an interest in the search for improved higher-order matching algorithms on warped images due to projection. Although currently, several existing methods "flatten" such 3D images to use planar graph / hypergraph matching methods, they still suffer from severe distortions and other undesired artifacts, which result in inaccurate matching. Alternatively, current planar methods cannot be trivially extended to effectively match points on images warped onto manifolds. Hence, matching on these warped images persists as a formidable challenge. In this paper, we pose the assignment problem as finding a bijective map between two graph induced simplicial complexes, which are higher-order analogues of graphs. We propose a constrained quadratic assignment problem (QAP) that matches each p-skeleton of the simplicial complexes, iterating from the highest to the lowest dimension. The accuracy and robustness of our approach are illustrated on both synthetic and real-world spherical / warped (projected) images with known ground-truth correspondences. We significantly outperform existing state-of-the-art spherical matching methods on a diverse set of datasets.Comment: Accepted at ECCV 202

    OmniDepth: Dense Depth Estimation for Indoors Spherical Panoramas.

    Get PDF
    Recent work on depth estimation up to now has only focused on projective images ignoring 360o content which is now increasingly and more easily produced. We show that monocular depth estimation models trained on traditional images produce sub-optimal results on omnidirectional images, showcasing the need for training directly on 360o datasets, which however, are hard to acquire. In this work, we circumvent the challenges associated with acquiring high quality 360o datasets with ground truth depth annotations, by re-using recently released large scale 3D datasets and re-purposing them to 360o via rendering. This dataset, which is considerably larger than similar projective datasets, is publicly offered to the community to enable future research in this direction. We use this dataset to learn in an end-to-end fashion the task of depth estimation from 360o images. We show promising results in our synthesized data as well as in unseen realistic images

    A dataset of annotated omnidirectional videos for distancing applications

    Get PDF
    Omnidirectional (or 360◦ ) cameras are acquisition devices that, in the next few years, could have a big impact on video surveillance applications, research, and industry, as they can record a spherical view of a whole environment from every perspective. This paper presents two new contributions to the research community: the CVIP360 dataset, an annotated dataset of 360◦ videos for distancing applications, and a new method to estimate the distances of objects in a scene from a single 360◦ image. The CVIP360 dataset includes 16 videos acquired outdoors and indoors, annotated by adding information about the pedestrians in the scene (bounding boxes) and the distances to the camera of some points in the 3D world by using markers at fixed and known intervals. The proposed distance estimation algorithm is based on geometry facts regarding the acquisition process of the omnidirectional device, and is uncalibrated in practice: the only required parameter is the camera height. The proposed algorithm was tested on the CVIP360 dataset, and empirical results demonstrate that the estimation error is negligible for distancing applications
    corecore