349 research outputs found
Prior-based facade rectification for AR in urban environment
International audienceWe present a method for automatic facade rectification and detection in the Manhattan world scenario. A Bayesian inference approach is proposed to recover the Manhattan directions in camera coordinate system, based on a prior we derived from the analysis of urban datasets. In addition, a SVM-based procedure is used to identify right-angle corners in the rectified images. These corners are clustered in facade regions using a greedy rectangular min-cut technique. Experiments on a standard dataset show that our algorithm performs better or as well as state-of-the-art techniques while being much faster
Design and Deployment of Photo2Building: A Cloud-based Procedural Modeling Tool as a Service
We present a Photo2Building tool to create a plausible 3D model of a building
from only a single photograph. Our tool is based on a prior desktop version
which, as described in this paper, is converted into a client-server model,
with job queuing, web-page support, and support of concurrent usage. The
reported cloud-based web-accessible tool can reconstruct a building in 40
seconds on average and costing only 0.60 USD with current pricing. This
provides for an extremely scalable and possibly widespread tool for creating
building models for use in urban design and planning applications. With the
growing impact of rapid urbanization on weather and climate and resource
availability, access to such a service is expected to help a wide variety of
users such as city planners, urban meteorologists worldwide in the quest to
improved prediction of urban weather and designing climate-resilient cities of
the future.Comment: 7 pages, 7 figures, PEARC '20: Practice and Experience in Advanced
Research Computing, July 26--30, 2020, Portland, OR, US
FacadeNet: Conditional Facade Synthesis via Selective Editing
We introduce FacadeNet, a deep learning approach for synthesizing building
facade images from diverse viewpoints. Our method employs a conditional GAN,
taking a single view of a facade along with the desired viewpoint information
and generates an image of the facade from the distinct viewpoint. To precisely
modify view-dependent elements like windows and doors while preserving the
structure of view-independent components such as walls, we introduce a
selective editing module. This module leverages image embeddings extracted from
a pre-trained vision transformer. Our experiments demonstrated state-of-the-art
performance on building facade generation, surpassing alternative methods
Facade Proposals for Urban Augmented Reality
International audienceWe introduce a novel object proposals method specific to building facades. We define new image cues that measure typical facadecharacteristics such as semantic, symmetry and repetitions. They are combined to generate a few facade candidates in urban environments fast. We show that our method outperforms state-of-the-art object proposals techniques for this task on the 1000 images of the Zurich Building Database. We demonstrate the interest of this procedure for augmented reality through facade recognition and camera pose initialization. In a very time-efficient pipeline we classify the candidates and match them to a facade references database using CNN-based descriptors. We prove that this approach is more robust to severe changes of viewpoint and occlusions than standard object recognition methods
Holistic Multi-View Building Analysis in the Wild with Projection Pooling
We address six different classification tasks related to fine-grained
building attributes: construction type, number of floors, pitch and geometry of
the roof, facade material, and occupancy class. Tackling such a remote building
analysis problem became possible only recently due to growing large-scale
datasets of urban scenes. To this end, we introduce a new benchmarking dataset,
consisting of 49426 images (top-view and street-view) of 9674 buildings. These
photos are further assembled, together with the geometric metadata. The dataset
showcases various real-world challenges, such as occlusions, blur, partially
visible objects, and a broad spectrum of buildings. We propose a new projection
pooling layer, creating a unified, top-view representation of the top-view and
the side views in a high-dimensional space. It allows us to utilize the
building and imagery metadata seamlessly. Introducing this layer improves
classification accuracy -- compared to highly tuned baseline models --
indicating its suitability for building analysis.Comment: Accepted for publication at the 35th AAAI Conference on Artificial
Intelligence (AAAI 2021
Single-Image Depth Prediction Makes Feature Matching Easier
Good local features improve the robustness of many 3D re-localization and
multi-view reconstruction pipelines. The problem is that viewing angle and
distance severely impact the recognizability of a local feature. Attempts to
improve appearance invariance by choosing better local feature points or by
leveraging outside information, have come with pre-requisites that made some of
them impractical. In this paper, we propose a surprisingly effective
enhancement to local feature extraction, which improves matching. We show that
CNN-based depths inferred from single RGB images are quite helpful, despite
their flaws. They allow us to pre-warp images and rectify perspective
distortions, to significantly enhance SIFT and BRISK features, enabling more
good matches, even when cameras are looking at the same scene but in opposite
directions.Comment: 14 pages, 7 figures, accepted for publication at the European
conference on computer vision (ECCV) 202
- …