398 research outputs found
Structured Indoor Modeling
In this dissertation, we propose data-driven approaches to reconstruct 3D models for indoor scenes which are represented in a structured way (e.g., a wall is represented by a planar surface and two rooms are connected via the wall). The structured representation of models is more application ready than dense representations (e.g., a point cloud), but poses additional challenges for reconstruction since extracting structures requires high-level understanding about geometries. To address this challenging problem, we explore two common structural regularities of indoor scenes: 1) most indoor structures consist of planar surfaces (planarity), and 2) structural surfaces (e.g., walls and floor) can be represented by a 2D floorplan as a top-down view projection (orthogonality). With breakthroughs in data capturing techniques, we develop automated systems to tackle structured modeling problems, namely piece-wise planar reconstruction and floorplan reconstruction, by learning shape priors (i.e., planarity and orthogonality) from data. With structured representations and production-level quality, the reconstructed models have an immediate impact on many industrial applications
Learning to Construct 3D Building Wireframes from 3D Line Clouds
Line clouds, though under-investigated in the previous work, potentially
encode more compact structural information of buildings than point clouds
extracted from multi-view images. In this work, we propose the first network to
process line clouds for building wireframe abstraction. The network takes a
line cloud as input , i.e., a nonstructural and unordered set of 3D line
segments extracted from multi-view images, and outputs a 3D wireframe of the
underlying building, which consists of a sparse set of 3D junctions connected
by line segments. We observe that a line patch, i.e., a group of neighboring
line segments, encodes sufficient contour information to predict the existence
and even the 3D position of a potential junction, as well as the likelihood of
connectivity between two query junctions. We therefore introduce a two-layer
Line-Patch Transformer to extract junctions and connectivities from sampled
line patches to form a 3D building wireframe model. We also introduce a
synthetic dataset of multi-view images with ground-truth 3D wireframe. We
extensively justify that our reconstructed 3D wireframe models significantly
improve upon multiple baseline building reconstruction methods. The code and
data can be found at https://github.com/Luo1Cheng/LC2WF.Comment: 10 pages, 6 figure
Towards automatic reconstruction of indoor scenes from incomplete point clouds: door and window detection and regularization
In the last years, point clouds have become the main source of information for building modelling. Although a considerable amount of methodologies addressing the automated generation of 3D models from point clouds have been developed, indoor modelling is still a challenging task due to complex building layouts and the high presence of severe clutters and occlusions. Most of methodologies are highly dependent on data quality, often producing irregular and non-consistent models. Although manmade environments generally exhibit some regularities, they are not commonly considered. This paper presents an optimization-based approach for detecting regularities (i.e., same shape, same alignment and same spacing) in building indoor features. The methodology starts from the detection of openings based on a voxel-based visibility analysis to distinguish âoccludedâ from âemptyâ regions in wall surfaces. The extraction of regular patterns in windows is addressed from studying the point cloud from an outdoor perspective. The layout is regularized by minimizing deformations while respecting the detected constraints. The methodology applies for elements placed in the same planeXunta de Galicia | Ref. ED481B 2016/079-
2D+3D Indoor Scene Understanding from a Single Monocular Image
Scene understanding, as a broad field encompassing many
subtopics, has gained great interest in recent years. Among these
subtopics, indoor scene understanding, having its own specific
attributes and challenges compared to outdoor scene under-
standing, has drawn a lot of attention. It has potential
applications in a wide variety of domains, such as robotic
navigation, object grasping for personal robotics, augmented
reality, etc. To our knowledge, existing research for indoor
scenes typically makes use of depth sensors, such as Kinect, that
is however not always available.
In this thesis, we focused on addressing the indoor scene
understanding tasks in a general case, where only a monocular
color image of the scene is available. Specifically, we first
studied the problem of estimating a detailed depth map from a
monocular image. Then, benefiting from deep-learning-based depth
estimation, we tackled the higher-level tasks of 3D box proposal
generation, and scene parsing with instance segmentation,
semantic labeling and support relationship inference from a
monocular image. Our research on indoor scene understanding
provides a comprehensive scene interpretation at various
perspectives and scales.
For monocular image depth estimation, previous approaches are
limited in that they only reason about depth locally on a single
scale, and do not utilize the important information of geometric
scene structures. Here, we developed a novel graphical model,
which reasons about detailed depth while leveraging geometric
scene structures at multiple scales.
For 3D box proposals, to our best knowledge, our approach
constitutes the first attempt to reason about class-independent
3D box proposals from a single monocular image. To this end, we
developed a novel integrated, differentiable framework that
estimates depth, extracts a volumetric scene representation and
generates 3D proposals. At the core of this framework lies a
novel residual, differentiable truncated signed distance function
module, which is able to handle the relatively low accuracy of
the predicted depth map.
For scene parsing, we tackled its three subtasks of instance
segmentation, se- mantic labeling, and the support relationship
inference on instances. Existing work typically reasons about
these individual subtasks independently. Here, we leverage the
fact that they bear strong connections, which can facilitate
addressing these sub- tasks if modeled properly. To this end, we
developed an integrated graphical model that reasons about the
mutual relationships of the above subtasks.
In summary, in this thesis, we introduced novel and effective
methodologies for each of three indoor scene understanding tasks,
i.e., depth estimation, 3D box proposal generation, and scene
parsing, and exploited the dependencies on depth estimates of the
latter two tasks. Evaluation on several benchmark datasets
demonstrated the effectiveness of our algorithms and the benefits
of utilizing depth estimates for higher-level tasks
Quantum Multi-Model Fitting
Geometric model fitting is a challenging but fundamental computer vision
problem. Recently, quantum optimization has been shown to enhance robust
fitting for the case of a single model, while leaving the question of
multi-model fitting open. In response to this challenge, this paper shows that
the latter case can significantly benefit from quantum hardware and proposes
the first quantum approach to multi-model fitting (MMF). We formulate MMF as a
problem that can be efficiently sampled by modern adiabatic quantum computers
without the relaxation of the objective function. We also propose an iterative
and decomposed version of our method, which supports real-world-sized problems.
The experimental evaluation demonstrates promising results on a variety of
datasets. The source code is available at:
https://github.com/FarinaMatteo/qmmf.Comment: In Computer Vision and Pattern Recognition (CVPR) 2023; Highligh
Automatic Reconstruction of Parametric, Volumetric Building Models from 3D Point Clouds
Planning, construction, modification, and analysis of buildings requires means of representing a building's physical structure and related semantics in a meaningful way. With the rise of novel technologies and increasing requirements in the architecture, engineering and construction (AEC) domain, two general concepts for representing buildings have gained particular attention in recent years. First, the concept of Building Information Modeling (BIM) is increasingly used as a modern means for representing and managing a building's as-planned state digitally, including not only a geometric model but also various additional semantic properties. Second, point cloud measurements are now widely used for capturing a building's as-built condition by means of laser scanning techniques. A particular challenge and topic of current research are methods for combining the strengths of both point cloud measurements and Building Information Modeling concepts to quickly obtain accurate building models from measured data. In this thesis, we present our recent approaches to tackle the intermeshed challenges of automated indoor point cloud interpretation using targeted segmentation methods, and the automatic reconstruction of high-level, parametric and volumetric building models as the basis for further usage in BIM scenarios. In contrast to most reconstruction methods available at the time, we fundamentally base our approaches on BIM principles and standards, and overcome critical limitations of previous approaches in order to reconstruct globally plausible, volumetric, and parametric models.Automatische Rekonstruktion von parametrischen, volumetrischen GebĂ€udemodellen aus 3D Punktwolken FĂŒr die Planung, Konstruktion, Modifikation und Analyse von GebĂ€uden werden Möglichkeiten zur sinnvollen ReprĂ€sentation der physischen GebĂ€udestruktur sowie dazugehöriger Semantik benötigt. Mit dem Aufkommen neuer Technologien und steigenden Anforderungen im Bereich von Architecture, Engineering and Construction (AEC) haben zwei Konzepte fĂŒr die ReprĂ€sentation von GebĂ€uden in den letzten Jahren besondere Aufmerksamkeit erlangt. Erstens wird das Konzept des Building Information Modeling (BIM) zunehmend als ein modernes Mittel zur digitalen Abbildung und Verwaltung "As-Planned"-Zustands von GebĂ€uden verwendet, welches nicht nur ein geometrisches Modell sondern auch verschiedene zusĂ€tzliche semantische Eigenschaften beinhaltet. Zweitens werden Punktwolkenmessungen inzwischen hĂ€ufig zur Aufnahme des "As-Built"-Zustands mittels Laser-Scan-Techniken eingesetzt. Eine besondere Herausforderung und Thema aktueller Forschung ist die Entwicklung von Methoden zur Vereinigung der StĂ€rken von Punktwolken und Konzepten des Building Information Modeling um schnell akkurate GebĂ€udemodelle aus den gemessenen Daten zu erzeugen. In dieser Dissertation prĂ€sentieren wir unsere aktuellen AnsĂ€tze um die miteinander verwobenen Herausforderungen anzugehen, Punktwolken mithilfe geeigneter Segmentierungsmethoden automatisiert zu interpretieren, sowie hochwertige, parametrische und volumetrische GebĂ€udemodelle als Basis fĂŒr die Verwendung im BIM-Umfeld zu rekonstruieren. Im Gegensatz zu den meisten derzeit verfĂŒgbaren Rekonstruktionsverfahren basieren unsere AnsĂ€tze grundlegend auf Prinzipien und Standards aus dem BIM-Umfeld und ĂŒberwinden kritische EinschrĂ€nkungen bisheriger AnsĂ€tze um vollstĂ€ndig plausible, volumetrische und parametrische Modelle zu erzeugen.</p
TOWARDS AUTOMATIC RECONSTRUCTION OF INDOOR SCENES FROM INCOMPLETE POINT CLOUDS: DOOR AND WINDOW DETECTION AND REGULARIZATION
In the last years, point clouds have become the main source of information for building modelling. Although a considerable amount of methodologies addressing the automated generation of 3D models from point clouds have been developed, indoor modelling is still a challenging task due to complex building layouts and the high presence of severe clutters and occlusions. Most of methodologies are highly dependent on data quality, often producing irregular and non-consistent models. Although manmade environments generally exhibit some regularities, they are not commonly considered. This paper presents an optimization-based approach for detecting regularities (i.e., same shape, same alignment and same spacing) in building indoor features. The methodology starts from the detection of openings based on a voxel-based visibility analysis to distinguish âoccludedâ from âemptyâ regions in wall surfaces. The extraction of regular patterns in windows is addressed from studying the point cloud from an outdoor perspective. The layout is regularized by minimizing deformations while respecting the detected constraints. The methodology applies for elements placed in the same plane
- âŠ