239 research outputs found

    SupFusion: Supervised LiDAR-Camera Fusion for 3D Object Detection

    Full text link
    In this paper, we propose a novel training strategy called SupFusion, which provides an auxiliary feature level supervision for effective LiDAR-Camera fusion and significantly boosts detection performance. Our strategy involves a data enhancement method named Polar Sampling, which densifies sparse objects and trains an assistant model to generate high-quality features as the supervision. These features are then used to train the LiDAR-Camera fusion model, where the fusion feature is optimized to simulate the generated high-quality features. Furthermore, we propose a simple yet effective deep fusion module, which contiguously gains superior performance compared with previous fusion methods with SupFusion strategy. In such a manner, our proposal shares the following advantages. Firstly, SupFusion introduces auxiliary feature-level supervision which could boost LiDAR-Camera detection performance without introducing extra inference costs. Secondly, the proposed deep fusion could continuously improve the detector's abilities. Our proposed SupFusion and deep fusion module is plug-and-play, we make extensive experiments to demonstrate its effectiveness. Specifically, we gain around 2% 3D mAP improvements on KITTI benchmark based on multiple LiDAR-Camera 3D detectors.Comment: Accepted to ICCV202

    Polyhistor: Parameter-Efficient Multi-Task Adaptation for Dense Vision Tasks

    Full text link
    Adapting large-scale pretrained models to various downstream tasks via fine-tuning is a standard method in machine learning. Recently, parameter-efficient fine-tuning methods show promise in adapting a pretrained model to different tasks while training only a few parameters. Despite their success, most existing methods are proposed in Natural Language Processing tasks with language Transformers, and adaptation to Computer Vision tasks with Vision Transformers remains under-explored, especially for dense vision tasks. Further, in multi-task settings, individually fine-tuning and storing separate models for different tasks is inefficient. In this work, we provide an extensive multi-task parameter-efficient benchmark and examine existing parameter-efficient fine-tuning NLP methods for vision tasks. Our results on four different dense vision tasks showed that existing methods cannot be efficiently integrated due to the hierarchical nature of the Hierarchical Vision Transformers. To overcome this issue, we propose Polyhistor and Polyhistor-Lite, consisting of Decomposed HyperNetworks and Layer-wise Scaling Kernels, to share information across different tasks with a few trainable parameters. This leads to favorable performance improvements against existing parameter-efficient methods while using fewer trainable parameters. Specifically, Polyhistor achieves competitive accuracy compared to the state-of-the-art while only using ~10% of their trainable parameters. Furthermore, our methods show larger performance gains when large networks and more pretraining data are used.Comment: Accepted to NeurIPS 2022; Project Page is at https://ycliu93.github.io/projects/polyhistor.htm

    Context-aware Event Forecasting via Graph Disentanglement

    Full text link
    Event forecasting has been a demanding and challenging task throughout the entire human history. It plays a pivotal role in crisis alarming and disaster prevention in various aspects of the whole society. The task of event forecasting aims to model the relational and temporal patterns based on historical events and makes forecasting to what will happen in the future. Most existing studies on event forecasting formulate it as a problem of link prediction on temporal event graphs. However, such pure structured formulation suffers from two main limitations: 1) most events fall into general and high-level types in the event ontology, and therefore they tend to be coarse-grained and offers little utility which inevitably harms the forecasting accuracy; and 2) the events defined by a fixed ontology are unable to retain the out-of-ontology contextual information. To address these limitations, we propose a novel task of context-aware event forecasting which incorporates auxiliary contextual information. First, the categorical context provides supplementary fine-grained information to the coarse-grained events. Second and more importantly, the context provides additional information towards specific situation and condition, which is crucial or even determinant to what will happen next. However, it is challenging to properly integrate context into the event forecasting framework, considering the complex patterns in the multi-context scenario. Towards this end, we design a novel framework named Separation and Collaboration Graph Disentanglement (short as SeCoGD) for context-aware event forecasting. Since there is no available dataset for this novel task, we construct three large-scale datasets based on GDELT. Experimental results demonstrate that our model outperforms a list of SOTA methods.Comment: KDD 2023, 9 pages, 7 figures, 4 table

    Transformation Decoupling Strategy based on Screw Theory for Deterministic Point Cloud Registration with Gravity Prior

    Full text link
    Point cloud registration is challenging in the presence of heavy outlier correspondences. This paper focuses on addressing the robust correspondence-based registration problem with gravity prior that often arises in practice. The gravity directions are typically obtained by inertial measurement units (IMUs) and can reduce the degree of freedom (DOF) of rotation from 3 to 1. We propose a novel transformation decoupling strategy by leveraging screw theory. This strategy decomposes the original 4-DOF problem into three sub-problems with 1-DOF, 2-DOF, and 1-DOF, respectively, thereby enhancing the computation efficiency. Specifically, the first 1-DOF represents the translation along the rotation axis and we propose an interval stabbing-based method to solve it. The second 2-DOF represents the pole which is an auxiliary variable in screw theory and we utilize a branch-and-bound method to solve it. The last 1-DOF represents the rotation angle and we propose a global voting method for its estimation. The proposed method sequentially solves three consensus maximization sub-problems, leading to efficient and deterministic registration. In particular, it can even handle the correspondence-free registration problem due to its significant robustness. Extensive experiments on both synthetic and real-world datasets demonstrate that our method is more efficient and robust than state-of-the-art methods, even when dealing with outlier rates exceeding 99%

    Distortion Reduction in Fractional Delay Filters

    Get PDF
    As the digital version of a continuous-time delay, the concept of fractional delay (FD) is exploited to approximate a desired delay that is not a multiple of the sampling interval. However, in FD filters, there is always a severe distortion at the beginning of delayed signals, referred to as head distortion. This letter identifies the cause of head distortion and proposes a solution to this problem for reducing the overall distortion in FD filters. For the purpose of performance evaluation, relative root-mean-square (RMS) error is formulated as a metric to quantify the overall difference between the frequency-domain response of an FD filter and the ideal one. Moreover, illustrative numerical results on the proposed scheme applied in FD filters with classical sinc, Farrow and Lagrange interpolation substantiate the validity and feasibility of our solution

    Rethinking Sensors Modeling: Hierarchical Information Enhanced Traffic Forecasting

    Full text link
    With the acceleration of urbanization, traffic forecasting has become an essential role in smart city construction. In the context of spatio-temporal prediction, the key lies in how to model the dependencies of sensors. However, existing works basically only consider the micro relationships between sensors, where the sensors are treated equally, and their macroscopic dependencies are neglected. In this paper, we argue to rethink the sensor's dependency modeling from two hierarchies: regional and global perspectives. Particularly, we merge original sensors with high intra-region correlation as a region node to preserve the inter-region dependency. Then, we generate representative and common spatio-temporal patterns as global nodes to reflect a global dependency between sensors and provide auxiliary information for spatio-temporal dependency learning. In pursuit of the generality and reality of node representations, we incorporate a Meta GCN to calibrate the regional and global nodes in the physical data space. Furthermore, we devise the cross-hierarchy graph convolution to propagate information from different hierarchies. In a nutshell, we propose a Hierarchical Information Enhanced Spatio-Temporal prediction method, HIEST, to create and utilize the regional dependency and common spatio-temporal patterns. Extensive experiments have verified the leading performance of our HIEST against state-of-the-art baselines. We publicize the code to ease reproducibility.Comment: 9 pages, accepted by CIKM'2

    Survival in Patients With Metastatic Prostate Cancer Undergoing Radiotherapy: The Importance of Prostate-Specific Antigen-Based Stratification

    Get PDF
    ObjectivesTo explore the effectiveness of radiotherapy in mPCa patients with different PSA stratifications based on the cancer database of a large population.BackgroundScreening criteria for patients with metastatic prostate cancer, who are candidates for radiotherapy, are rarely reported.Patients and MethodsWe identified 22,604 patients with metastatic prostate cancer in the Surveillance, Epidemiology, and End Results database and divided them into a radiotherapy group and a control group. Patients with metastatic prostate cancer were divided into subgroups according to their levels of prostate-specific antigen to evaluate the efficacy of radiotherapy. They were also divided into six subgroups according to their prostate-specific antigen levels. We used multivariate Cox analysis to evaluate overall survival and cancer-specific survival. After 1:1 propensity score matching, Kaplan-Meier analysis was used to explore the difference in overall survival and cancer-specific survival in the radiotherapy and control group.ResultsIn all, 5,505 patients received radiotherapy, compared to 17,099 in the control group. In the multivariate Cox analysis, radiotherapy improved overall survival (hazard ratio [HR]: 0.730, 95% confidence interval [CI]: 0.636–0.838; P<0.001) and cancer-specific survival (HR: 0.764, 95% CI: 0.647–0.903; P=0.002) in patients with a PSA level of 4–10 ng/mL. Similar results were obtained by Kaplan-Meier analysis after 1:1 propensity score matching. In patients with prostate-specific antigen levels between 4–10 ng/mL, the overall survival (P<0.001) and cancer-specific survival (P<0.05) in the radiotherapy group was significantly better than those in the control group.ConclusionThe result of this large population-based study shows that rigorous selection of appropriate metastatic prostate cancer patients for radiotherapy can benefit prognosis significantly. This can be the basis for future prospective trials

    ContraGen: Effective Contrastive Learning For Causal Language Model

    Full text link
    Despite exciting progress in large-scale language generation, the expressiveness of its representations is severely limited by the \textit{anisotropy} issue where the hidden representations are distributed into a narrow cone in the vector space. To address this issue, we present ContraGen, a novel contrastive learning framework to improve the representation with better uniformity and discrimination. We assess ContraGen on a wide range of downstream tasks in natural and programming languages. We show that ContraGen can effectively enhance both uniformity and discrimination of the representations and lead to the desired improvement on various language understanding tasks where discriminative representations are crucial for attaining good performance. Specifically, we attain 44%44\% relative improvement on the Semantic Textual Similarity tasks and 34%34\% on Code-to-Code Search tasks. Furthermore, by improving the expressiveness of the representations, ContraGen also boosts the source code generation capability with 9%9\% relative improvement on execution accuracy on the HumanEval benchmark.Comment: 10 page
    • …
    corecore