165,831 research outputs found
3D Point Cloud Denoising via Deep Neural Network based Local Surface Estimation
We present a neural-network-based architecture for 3D point cloud denoising
called neural projection denoising (NPD). In our previous work, we proposed a
two-stage denoising algorithm, which first estimates reference planes and
follows by projecting noisy points to estimated reference planes. Since the
estimated reference planes are inevitably noisy, multi-projection is applied to
stabilize the denoising performance. NPD algorithm uses a neural network to
estimate reference planes for points in noisy point clouds. With more accurate
estimations of reference planes, we are able to achieve better denoising
performances with only one-time projection. To the best of our knowledge, NPD
is the first work to denoise 3D point clouds with deep learning techniques. To
conduct the experiments, we sample 40000 point clouds from the 3D data in
ShapeNet to train a network and sample 350 point clouds from the 3D data in
ModelNet10 to test. Experimental results show that our algorithm can estimate
normal vectors of points in noisy point clouds. Comparing to five competitive
methods, the proposed algorithm achieves better denoising performance and
produces much smaller variances
Learning to See the Wood for the Trees: Deep Laser Localization in Urban and Natural Environments on a CPU
Localization in challenging, natural environments such as forests or
woodlands is an important capability for many applications from guiding a robot
navigating along a forest trail to monitoring vegetation growth with handheld
sensors. In this work we explore laser-based localization in both urban and
natural environments, which is suitable for online applications. We propose a
deep learning approach capable of learning meaningful descriptors directly from
3D point clouds by comparing triplets (anchor, positive and negative examples).
The approach learns a feature space representation for a set of segmented point
clouds that are matched between a current and previous observations. Our
learning method is tailored towards loop closure detection resulting in a small
model which can be deployed using only a CPU. The proposed learning method
would allow the full pipeline to run on robots with limited computational
payload such as drones, quadrupeds or UGVs.Comment: Accepted for publication at RA-L/ICRA 2019. More info:
https://ori.ox.ac.uk/esm-localizatio
Pose-Graph Attentional Graph Neural Network for Lidar Place Recognition
This paper proposes a lidar place recognition approach, called P-GAT, to
increase the receptive field between point clouds captured over time. Instead
of comparing pairs of point clouds, we compare the similarity between sets of
point clouds to use the maximum spatial and temporal information between
neighbour clouds utilising the concept of pose-graph SLAM. Leveraging intra-
and inter-attention and graph neural network, P-GAT relates point clouds
captured in nearby locations in Euclidean space and their embeddings in feature
space. Experimental results on the large-scale publically available datasets
demonstrate the effectiveness of our approach in recognising scenes lacking
distinct features and when training and testing environments have different
distributions (domain adaptation). Further, an exhaustive comparison with the
state-of-the-art shows improvements in performance gains. Code will be
available upon acceptance.Comment: 8 pages, 3 figures, 5 table
Prior-less 3D Human Shape Reconstruction with an Earth Mover’s Distance Informed CNN
We propose a novel end-to-end deep learning framework, capable of 3D human shape reconstruction from a 2D image without the need of a 3D prior parametric model. We employ a “prior-less” representation of the human shape using unordered point clouds. Due to the lack of prior information, comparing the generated and ground truth point clouds to evaluate the reconstruction error is challenging. We solve this problem by proposing an Earth Mover’s Distance (EMD) function to find the optimal mapping between point clouds. Our experimental results show that we are able to obtain a visually accurate estimation of the 3D human shape from a single 2D image, with some inaccuracy for heavily occluded parts
Santiago urban dataset SUD: Combination of Handheld and Mobile Laser Scanning point clouds
Santiago Urban Dataset SUD is a real dataset that combines Mobile Laser Scanning (MLS) and Handheld Mobile Laser Scanning (HMLS) point clouds. The data is composed by 2 km of streets, sited in Santiago de Compostela (Spain). Point clouds undergo a manual labelling process supported by both heuristic and Deep Learning methods, resulting in the classification of eight specific classes: road, sidewalk, curb, buildings, vehicles, vegetation, poles, and others. Three PointNet++ models were trained; the first one using MLS point clouds, the second one with HMLS point clouds and the third one with both H&MLS point clouds. In order to ascertain the quality and efficacy of each Deep Learning model, various metrics were employed, including confusion matrices, precision, recall, F1-score, and IoU. The results are consistent with other state-of-the-art works and indicate that SUD is valid for comparing point cloud semantic segmentation works. Furthermore, the survey's extensive coverage and the limited occlusions indicate the potential utility of SUD in urban mobility research.Agencia Estatal de InvestigaciĂłn | Ref. PID2019-105221RB-C43Xunta de Galicia | Ref. ED481B-2019-061Xunta de Galicia | Ref. ED431C 2020/01Universidade de Vigo/CISU
SalsaNet: Fast Road and Vehicle Segmentation in LiDAR Point Clouds for Autonomous Driving
In this paper, we introduce a deep encoder-decoder network, named SalsaNet,
for efficient semantic segmentation of 3D LiDAR point clouds. SalsaNet segments
the road, i.e. drivable free-space, and vehicles in the scene by employing the
Bird-Eye-View (BEV) image projection of the point cloud. To overcome the lack
of annotated point cloud data, in particular for the road segments, we
introduce an auto-labeling process which transfers automatically generated
labels from the camera to LiDAR. We also explore the role of imagelike
projection of LiDAR data in semantic segmentation by comparing BEV with
spherical-front-view projection and show that SalsaNet is projection-agnostic.
We perform quantitative and qualitative evaluations on the KITTI dataset, which
demonstrate that the proposed SalsaNet outperforms other state-of-the-art
semantic segmentation networks in terms of accuracy and computation time. Our
code and data are publicly available at
https://gitlab.com/aksoyeren/salsanet.git
COMPARING ACCURACY OF ULTRA-DENSE LASER SCANNER AND PHOTOGRAMMETRY POINT CLOUDS
Abstract. Massive point clouds have now become a common product from surveys using passive (photogrammetry) or active (laser scanning) technologies. A common question is what is the difference in terms of accuracy and precision of different technologies and processing options. In this work four ultra-dense point-clouds (PCs) from drone surveys are compared. Two PCs were created from imagery using a photogrammetric workflow, with and without ground control points. The laser scanning PCs were created with two drone flights with Riegl MiniVUX-3 lidar sensor, resulting in a point cloud with ~300 million points, and Riegl VUX-120 lidar sensor, leading to a point cloud with ~1 billion points. Relative differences between pairs from permutations of the four PCs are analysed calculating point-to-point distances over nearest neighbours. Eleven clipped PC subsets are used for this task. Ground control points (GCPs) are also used to assess residuals in the two photogrammetric point clouds in order to quantify the improvement from using GCPs vs not using GCPs when processing the images.Results related to comparing the two photogrammetric point clouds with and without GCPs show an improvement of average absolute position error from 0.12 m to 0.05 m and RMSE from 0.03 m to 0.01 m. Point-to-point distances over the PC pairs show that the closest point clouds are the two lidar clouds, with mean absolute distance (MAD), median absolute distance (MdAD) and standard deviation of distances (RMSE) respectively of 0.031 m, 0.025 m, 0.019 m; largest difference is between photogrammetric PC with GCPs, with 0.208 m, 0.206 m and 0.116 m, with the Z component providing most of the difference. Photogrammetry without GCP was more consistent with the lidar point clouds, with MAD of 0.064 m, MdAD of 0.048 m and RMSE value of 0.114 m
- …