74 research outputs found

    FusionRCNN: LiDAR-Camera Fusion for Two-stage 3D Object Detection

    Full text link
    3D object detection with multi-sensors is essential for an accurate and reliable perception system of autonomous driving and robotics. Existing 3D detectors significantly improve the accuracy by adopting a two-stage paradigm which merely relies on LiDAR point clouds for 3D proposal refinement. Though impressive, the sparsity of point clouds, especially for the points far away, making it difficult for the LiDAR-only refinement module to accurately recognize and locate objects.To address this problem, we propose a novel multi-modality two-stage approach named FusionRCNN, which effectively and efficiently fuses point clouds and camera images in the Regions of Interest(RoI). FusionRCNN adaptively integrates both sparse geometry information from LiDAR and dense texture information from camera in a unified attention mechanism. Specifically, it first utilizes RoIPooling to obtain an image set with a unified size and gets the point set by sampling raw points within proposals in the RoI extraction step; then leverages an intra-modality self-attention to enhance the domain-specific features, following by a well-designed cross-attention to fuse the information from two modalities.FusionRCNN is fundamentally plug-and-play and supports different one-stage methods with almost no architectural changes. Extensive experiments on KITTI and Waymo benchmarks demonstrate that our method significantly boosts the performances of popular detectors.Remarkably, FusionRCNN significantly improves the strong SECOND baseline by 6.14% mAP on Waymo, and outperforms competing two-stage approaches. Code will be released soon at https://github.com/xxlbigbrother/Fusion-RCNN.Comment: 7 pages, 3 figure

    Deep Semantic Graph Matching for Large-scale Outdoor Point Clouds Registration

    Full text link
    Current point cloud registration methods are mainly based on local geometric information and usually ignore the semantic information contained in the scenes. In this paper, we treat the point cloud registration problem as a semantic instance matching and registration task, and propose a deep semantic graph matching method (DeepSGM) for large-scale outdoor point cloud registration. Firstly, the semantic categorical labels of 3D points are obtained using a semantic segmentation network. The adjacent points with the same category labels are then clustered together using the Euclidean clustering algorithm to obtain the semantic instances, which are represented by three kinds of attributes including spatial location information, semantic categorical information, and global geometric shape information. Secondly, the semantic adjacency graph is constructed based on the spatial adjacency relations of semantic instances. To fully explore the topological structures between semantic instances in the same scene and across different scenes, the spatial distribution features and the semantic categorical features are learned with graph convolutional networks, and the global geometric shape features are learned with a PointNet-like network. These three kinds of features are further enhanced with the self-attention and cross-attention mechanisms. Thirdly, the semantic instance matching is formulated as an optimal transport problem, and solved through an optimal matching layer. Finally, the geometric transformation matrix between two point clouds is first estimated by the SVD algorithm and then refined by the ICP algorithm. Experimental results conducted on the KITTI Odometry dataset demonstrate that the proposed method improves the registration performance and outperforms various state-of-the-art methods.Comment: 12 pages, 6 figure

    CAGroup3D: Class-Aware Grouping for 3D Object Detection on Point Clouds

    Full text link
    We present a novel two-stage fully sparse convolutional 3D object detection framework, named CAGroup3D. Our proposed method first generates some high-quality 3D proposals by leveraging the class-aware local group strategy on the object surface voxels with the same semantic predictions, which considers semantic consistency and diverse locality abandoned in previous bottom-up approaches. Then, to recover the features of missed voxels due to incorrect voxel-wise segmentation, we build a fully sparse convolutional RoI pooling module to directly aggregate fine-grained spatial information from backbone for further proposal refinement. It is memory-and-computation efficient and can better encode the geometry-specific features of each 3D proposal. Our model achieves state-of-the-art 3D detection performance with remarkable gains of +\textit{3.6\%} on ScanNet V2 and +\textit{2.6}\% on SUN RGB-D in term of [email protected]. Code will be available at https://github.com/Haiyang-W/CAGroup3D.Comment: Accept by NeurIPS202

    Sample-adaptive Augmentation for Point Cloud Recognition Against Real-world Corruptions

    Full text link
    Robust 3D perception under corruption has become an essential task for the realm of 3D vision. While current data augmentation techniques usually perform random transformations on all point cloud objects in an offline way and ignore the structure of the samples, resulting in over-or-under enhancement. In this work, we propose an alternative to make sample-adaptive transformations based on the structure of the sample to cope with potential corruption via an auto-augmentation framework, named as AdaptPoint. Specially, we leverage a imitator, consisting of a Deformation Controller and a Mask Controller, respectively in charge of predicting deformation parameters and producing a per-point mask, based on the intrinsic structural information of the input point cloud, and then conduct corruption simulations on top. Then a discriminator is utilized to prevent the generation of excessive corruption that deviates from the original data distribution. In addition, a perception-guidance feedback mechanism is incorporated to guide the generation of samples with appropriate difficulty level. Furthermore, to address the paucity of real-world corrupted point cloud, we also introduce a new dataset ScanObjectNN-C, that exhibits greater similarity to actual data in real-world environments, especially when contrasted with preceding CAD datasets. Experiments show that our method achieves state-of-the-art results on multiple corruption benchmarks, including ModelNet-C, our ScanObjectNN-C, and ShapeNet-C.Comment: Accepted by ICCV2023; code: https://github.com/Roywangj/AdaptPoin

    Case report: Sex-specific characteristics of epilepsy phenotypes associated with Xp22.31 deletion: A case report and review

    Get PDF
    Deletion in the Xp22.31 region is increasingly suggested to be involved in the etiology of epilepsy. Little is known regarding the genomic and clinical delineations of X-linked epilepsy in the Chinese population or the sex-stratified difference in epilepsy characteristics associated with deletions in the Xp22.31 region. In this study, we reported two siblings with a 1.69 Mb maternally inherited microdeletion at Xp22.31 involving the genes VCX3A, HDHD1, STS, VCX, VCX2, and PNPLA4 presenting with easily controlled focal epilepsy and language delay with mild ichthyosis in a Chinese family with a traceable 4-generation history of skin ichthyosis. Both brain magnetic resonance imaging results were normal, while EEG revealed epileptic abnormalities. We further performed an exhaustive literature search, documenting 25 patients with epilepsy with gene defects in Xp22.31, and summarized the epilepsy heterogeneities between sexes. Males harboring the Xp22.31 deletion mainly manifested with child-onset, easily controlled focal epilepsy accompanied by X-linked ichthyosis; the deletions were mostly X-linked recessive, with copy number variants (CNVs) in the classic region of deletion (863.38 kb–2 Mb). In contrast, epilepsy in females tended to be earlier-onset, and relatively refractory, with pathogenic CNV sizes varying over a larger range (859 kb–56.36 Mb); the alterations were infrequently inherited and almost combined with additional CNVs. A candidate region encompassing STS, HDHD1, and MIR4767 was the likely pathogenic epilepsy-associated region. This study filled in the knowledge gap regarding the genomic and clinical delineations of X-linked recessive epilepsy in the Chinese population and extends the understanding of the sex-specific characteristics of Xp22.31 deletion in regard to epilepsy

    Chromosomal aberrations in pediatric patients with moderate/severe developmental delay/intellectual disability with abundant phenotypic heterogeneities: A single-center study

    Get PDF
    Background: This study aimed to examine the clinical usefulness of chromosome microarray (CMA) for selective implementation in patients with unexplained moderate or severe developmental delay/intellectual disability (DD/ID) and/or combined with different dysphonic features in the Han Chinese population. Methods: We retrospectively analyzed data on 122 pediatric patients with unexplained isolated moderate/severe DD/ID with or without autism spectrum disorders, epilepsy, dystonia, and congenital abnormalities from a single-center neurorehabilitation clinic in southern China. Results: A total of 46 probands (37.7%) had abnormal CMA results among the 122 study patients. With the exclusion of aneuploidies, uniparental disomies, and multiple homozygotes, 37 patients harbored 39 pathogenic copy number variations (pCNVs) (median [interquartile range] size: 3.57 [1.6 to 7.1] Mb; 33 deletions and 6 duplications), enriched in chromosomes 5, 7, 15, 17, and 22, with a markedly high prevalence of Angelman/Prader-Willi syndrome (24.3% [nine of 37]). Three rare deletions in the regions 5q33.2q34, 17p13.2, and 13q33.2 were reported, with specific delineation of clinical phenotypes. The frequencies of pCNVs were 18%, 33.3%, 38.89%, 41.67%, and 100% for patients with 1, 2, 3, 4, and 5 study phenotypes, respectively; patients with more concomitant abnormalities in the heart, brain, craniofacial region, and/or other organs had a higher CMA diagnostic yield and pCNV prevalence (P \u3c 0.05). Conclusions: Clinical application of CMA as a first-tier test among patients with moderate/severe DD/ID combined with congenital structural anomalies improved diagnostic yields and the quality of clinical management in this series of patients

    Cre-loxP-mediated genetic lineage tracing: Unraveling cell fate and origin in the developing heart

    Get PDF
    The Cre-loxP-mediated genetic lineage tracing system is essential for constructing the fate mapping of single-cell progeny or cell populations. Understanding the structural hierarchy of cardiac progenitor cells facilitates unraveling cell fate and origin issues in cardiac development. Several prospective Cre-loxP-based lineage-tracing systems have been used to analyze precisely the fate determination and developmental characteristics of endocardial cells (ECs), epicardial cells, and cardiomyocytes. Therefore, emerging lineage-tracing techniques advance the study of cardiovascular-related cellular plasticity. In this review, we illustrate the principles and methods of the emerging Cre-loxP-based genetic lineage tracing technology for trajectory monitoring of distinct cell lineages in the heart. The comprehensive demonstration of the differentiation process of single-cell progeny using genetic lineage tracing technology has made outstanding contributions to cardiac development and homeostasis, providing new therapeutic strategies for tissue regeneration in congenital and cardiovascular diseases (CVDs)

    Learning Point-wise Abstaining Penalty for Point Cloud Anomaly Detection

    Full text link
    LiDAR-based semantic scene understanding is an important module in the modern autonomous driving perception stack. However, identifying Out-Of-Distribution (OOD) points in a LiDAR point cloud is challenging as point clouds lack semantically rich features when compared with RGB images. We revisit this problem from the perspective of selective classification, which introduces a selective function into the standard closed-set classification setup. Our solution is built upon the basic idea of abstaining from choosing any known categories but learns a point-wise abstaining penalty with a marginbased loss. Synthesizing outliers to approximate unlimited OOD samples is also critical to this idea, so we propose a strong synthesis pipeline that generates outliers originated from various factors: unrealistic object categories, sampling patterns and sizes. We demonstrate that learning different abstaining penalties, apart from point-wise penalty, for different types of (synthesized) outliers can further improve the performance. We benchmark our method on SemanticKITTI and nuScenes and achieve state-of-the-art results. Risk-coverage analysis further reveals intrinsic properties of different methods. Codes and models will be publicly available.Comment: codes is available at https://github.com/Daniellli/PAD.gi
    • …
    corecore