165,831 research outputs found

    3D Point Cloud Denoising via Deep Neural Network based Local Surface Estimation

    Full text link
    We present a neural-network-based architecture for 3D point cloud denoising called neural projection denoising (NPD). In our previous work, we proposed a two-stage denoising algorithm, which first estimates reference planes and follows by projecting noisy points to estimated reference planes. Since the estimated reference planes are inevitably noisy, multi-projection is applied to stabilize the denoising performance. NPD algorithm uses a neural network to estimate reference planes for points in noisy point clouds. With more accurate estimations of reference planes, we are able to achieve better denoising performances with only one-time projection. To the best of our knowledge, NPD is the first work to denoise 3D point clouds with deep learning techniques. To conduct the experiments, we sample 40000 point clouds from the 3D data in ShapeNet to train a network and sample 350 point clouds from the 3D data in ModelNet10 to test. Experimental results show that our algorithm can estimate normal vectors of points in noisy point clouds. Comparing to five competitive methods, the proposed algorithm achieves better denoising performance and produces much smaller variances

    Learning to See the Wood for the Trees: Deep Laser Localization in Urban and Natural Environments on a CPU

    Full text link
    Localization in challenging, natural environments such as forests or woodlands is an important capability for many applications from guiding a robot navigating along a forest trail to monitoring vegetation growth with handheld sensors. In this work we explore laser-based localization in both urban and natural environments, which is suitable for online applications. We propose a deep learning approach capable of learning meaningful descriptors directly from 3D point clouds by comparing triplets (anchor, positive and negative examples). The approach learns a feature space representation for a set of segmented point clouds that are matched between a current and previous observations. Our learning method is tailored towards loop closure detection resulting in a small model which can be deployed using only a CPU. The proposed learning method would allow the full pipeline to run on robots with limited computational payload such as drones, quadrupeds or UGVs.Comment: Accepted for publication at RA-L/ICRA 2019. More info: https://ori.ox.ac.uk/esm-localizatio

    Pose-Graph Attentional Graph Neural Network for Lidar Place Recognition

    Full text link
    This paper proposes a lidar place recognition approach, called P-GAT, to increase the receptive field between point clouds captured over time. Instead of comparing pairs of point clouds, we compare the similarity between sets of point clouds to use the maximum spatial and temporal information between neighbour clouds utilising the concept of pose-graph SLAM. Leveraging intra- and inter-attention and graph neural network, P-GAT relates point clouds captured in nearby locations in Euclidean space and their embeddings in feature space. Experimental results on the large-scale publically available datasets demonstrate the effectiveness of our approach in recognising scenes lacking distinct features and when training and testing environments have different distributions (domain adaptation). Further, an exhaustive comparison with the state-of-the-art shows improvements in performance gains. Code will be available upon acceptance.Comment: 8 pages, 3 figures, 5 table

    Prior-less 3D Human Shape Reconstruction with an Earth Mover’s Distance Informed CNN

    Get PDF
    We propose a novel end-to-end deep learning framework, capable of 3D human shape reconstruction from a 2D image without the need of a 3D prior parametric model. We employ a “prior-less” representation of the human shape using unordered point clouds. Due to the lack of prior information, comparing the generated and ground truth point clouds to evaluate the reconstruction error is challenging. We solve this problem by proposing an Earth Mover’s Distance (EMD) function to find the optimal mapping between point clouds. Our experimental results show that we are able to obtain a visually accurate estimation of the 3D human shape from a single 2D image, with some inaccuracy for heavily occluded parts

    Santiago urban dataset SUD: Combination of Handheld and Mobile Laser Scanning point clouds

    Get PDF
    Santiago Urban Dataset SUD is a real dataset that combines Mobile Laser Scanning (MLS) and Handheld Mobile Laser Scanning (HMLS) point clouds. The data is composed by 2 km of streets, sited in Santiago de Compostela (Spain). Point clouds undergo a manual labelling process supported by both heuristic and Deep Learning methods, resulting in the classification of eight specific classes: road, sidewalk, curb, buildings, vehicles, vegetation, poles, and others. Three PointNet++ models were trained; the first one using MLS point clouds, the second one with HMLS point clouds and the third one with both H&MLS point clouds. In order to ascertain the quality and efficacy of each Deep Learning model, various metrics were employed, including confusion matrices, precision, recall, F1-score, and IoU. The results are consistent with other state-of-the-art works and indicate that SUD is valid for comparing point cloud semantic segmentation works. Furthermore, the survey's extensive coverage and the limited occlusions indicate the potential utility of SUD in urban mobility research.Agencia Estatal de InvestigaciĂłn | Ref. PID2019-105221RB-C43Xunta de Galicia | Ref. ED481B-2019-061Xunta de Galicia | Ref. ED431C 2020/01Universidade de Vigo/CISU

    SalsaNet: Fast Road and Vehicle Segmentation in LiDAR Point Clouds for Autonomous Driving

    Full text link
    In this paper, we introduce a deep encoder-decoder network, named SalsaNet, for efficient semantic segmentation of 3D LiDAR point clouds. SalsaNet segments the road, i.e. drivable free-space, and vehicles in the scene by employing the Bird-Eye-View (BEV) image projection of the point cloud. To overcome the lack of annotated point cloud data, in particular for the road segments, we introduce an auto-labeling process which transfers automatically generated labels from the camera to LiDAR. We also explore the role of imagelike projection of LiDAR data in semantic segmentation by comparing BEV with spherical-front-view projection and show that SalsaNet is projection-agnostic. We perform quantitative and qualitative evaluations on the KITTI dataset, which demonstrate that the proposed SalsaNet outperforms other state-of-the-art semantic segmentation networks in terms of accuracy and computation time. Our code and data are publicly available at https://gitlab.com/aksoyeren/salsanet.git

    COMPARING ACCURACY OF ULTRA-DENSE LASER SCANNER AND PHOTOGRAMMETRY POINT CLOUDS

    Get PDF
    Abstract. Massive point clouds have now become a common product from surveys using passive (photogrammetry) or active (laser scanning) technologies. A common question is what is the difference in terms of accuracy and precision of different technologies and processing options. In this work four ultra-dense point-clouds (PCs) from drone surveys are compared. Two PCs were created from imagery using a photogrammetric workflow, with and without ground control points. The laser scanning PCs were created with two drone flights with Riegl MiniVUX-3 lidar sensor, resulting in a point cloud with ~300 million points, and Riegl VUX-120 lidar sensor, leading to a point cloud with ~1 billion points. Relative differences between pairs from permutations of the four PCs are analysed calculating point-to-point distances over nearest neighbours. Eleven clipped PC subsets are used for this task. Ground control points (GCPs) are also used to assess residuals in the two photogrammetric point clouds in order to quantify the improvement from using GCPs vs not using GCPs when processing the images.Results related to comparing the two photogrammetric point clouds with and without GCPs show an improvement of average absolute position error from 0.12 m to 0.05 m and RMSE from 0.03 m to 0.01 m. Point-to-point distances over the PC pairs show that the closest point clouds are the two lidar clouds, with mean absolute distance (MAD), median absolute distance (MdAD) and standard deviation of distances (RMSE) respectively of 0.031 m, 0.025 m, 0.019 m; largest difference is between photogrammetric PC with GCPs, with 0.208 m, 0.206 m and 0.116 m, with the Z component providing most of the difference. Photogrammetry without GCP was more consistent with the lidar point clouds, with MAD of 0.064 m, MdAD of 0.048 m and RMSE value of 0.114 m
    • …
    corecore