239 research outputs found
SupFusion: Supervised LiDAR-Camera Fusion for 3D Object Detection
In this paper, we propose a novel training strategy called SupFusion, which
provides an auxiliary feature level supervision for effective LiDAR-Camera
fusion and significantly boosts detection performance. Our strategy involves a
data enhancement method named Polar Sampling, which densifies sparse objects
and trains an assistant model to generate high-quality features as the
supervision. These features are then used to train the LiDAR-Camera fusion
model, where the fusion feature is optimized to simulate the generated
high-quality features. Furthermore, we propose a simple yet effective deep
fusion module, which contiguously gains superior performance compared with
previous fusion methods with SupFusion strategy. In such a manner, our proposal
shares the following advantages. Firstly, SupFusion introduces auxiliary
feature-level supervision which could boost LiDAR-Camera detection performance
without introducing extra inference costs. Secondly, the proposed deep fusion
could continuously improve the detector's abilities. Our proposed SupFusion and
deep fusion module is plug-and-play, we make extensive experiments to
demonstrate its effectiveness. Specifically, we gain around 2% 3D mAP
improvements on KITTI benchmark based on multiple LiDAR-Camera 3D detectors.Comment: Accepted to ICCV202
Polyhistor: Parameter-Efficient Multi-Task Adaptation for Dense Vision Tasks
Adapting large-scale pretrained models to various downstream tasks via
fine-tuning is a standard method in machine learning. Recently,
parameter-efficient fine-tuning methods show promise in adapting a pretrained
model to different tasks while training only a few parameters. Despite their
success, most existing methods are proposed in Natural Language Processing
tasks with language Transformers, and adaptation to Computer Vision tasks with
Vision Transformers remains under-explored, especially for dense vision tasks.
Further, in multi-task settings, individually fine-tuning and storing separate
models for different tasks is inefficient. In this work, we provide an
extensive multi-task parameter-efficient benchmark and examine existing
parameter-efficient fine-tuning NLP methods for vision tasks. Our results on
four different dense vision tasks showed that existing methods cannot be
efficiently integrated due to the hierarchical nature of the Hierarchical
Vision Transformers. To overcome this issue, we propose Polyhistor and
Polyhistor-Lite, consisting of Decomposed HyperNetworks and Layer-wise Scaling
Kernels, to share information across different tasks with a few trainable
parameters. This leads to favorable performance improvements against existing
parameter-efficient methods while using fewer trainable parameters.
Specifically, Polyhistor achieves competitive accuracy compared to the
state-of-the-art while only using ~10% of their trainable parameters.
Furthermore, our methods show larger performance gains when large networks and
more pretraining data are used.Comment: Accepted to NeurIPS 2022; Project Page is at
https://ycliu93.github.io/projects/polyhistor.htm
Context-aware Event Forecasting via Graph Disentanglement
Event forecasting has been a demanding and challenging task throughout the
entire human history. It plays a pivotal role in crisis alarming and disaster
prevention in various aspects of the whole society. The task of event
forecasting aims to model the relational and temporal patterns based on
historical events and makes forecasting to what will happen in the future. Most
existing studies on event forecasting formulate it as a problem of link
prediction on temporal event graphs. However, such pure structured formulation
suffers from two main limitations: 1) most events fall into general and
high-level types in the event ontology, and therefore they tend to be
coarse-grained and offers little utility which inevitably harms the forecasting
accuracy; and 2) the events defined by a fixed ontology are unable to retain
the out-of-ontology contextual information. To address these limitations, we
propose a novel task of context-aware event forecasting which incorporates
auxiliary contextual information. First, the categorical context provides
supplementary fine-grained information to the coarse-grained events. Second and
more importantly, the context provides additional information towards specific
situation and condition, which is crucial or even determinant to what will
happen next. However, it is challenging to properly integrate context into the
event forecasting framework, considering the complex patterns in the
multi-context scenario. Towards this end, we design a novel framework named
Separation and Collaboration Graph Disentanglement (short as SeCoGD) for
context-aware event forecasting. Since there is no available dataset for this
novel task, we construct three large-scale datasets based on GDELT.
Experimental results demonstrate that our model outperforms a list of SOTA
methods.Comment: KDD 2023, 9 pages, 7 figures, 4 table
Transformation Decoupling Strategy based on Screw Theory for Deterministic Point Cloud Registration with Gravity Prior
Point cloud registration is challenging in the presence of heavy outlier
correspondences. This paper focuses on addressing the robust
correspondence-based registration problem with gravity prior that often arises
in practice. The gravity directions are typically obtained by inertial
measurement units (IMUs) and can reduce the degree of freedom (DOF) of rotation
from 3 to 1. We propose a novel transformation decoupling strategy by
leveraging screw theory. This strategy decomposes the original 4-DOF problem
into three sub-problems with 1-DOF, 2-DOF, and 1-DOF, respectively, thereby
enhancing the computation efficiency. Specifically, the first 1-DOF represents
the translation along the rotation axis and we propose an interval
stabbing-based method to solve it. The second 2-DOF represents the pole which
is an auxiliary variable in screw theory and we utilize a branch-and-bound
method to solve it. The last 1-DOF represents the rotation angle and we propose
a global voting method for its estimation. The proposed method sequentially
solves three consensus maximization sub-problems, leading to efficient and
deterministic registration. In particular, it can even handle the
correspondence-free registration problem due to its significant robustness.
Extensive experiments on both synthetic and real-world datasets demonstrate
that our method is more efficient and robust than state-of-the-art methods,
even when dealing with outlier rates exceeding 99%
Distortion Reduction in Fractional Delay Filters
As the digital version of a continuous-time delay, the concept of fractional delay (FD) is exploited to approximate a desired delay that is not a multiple of the sampling interval. However, in FD filters, there is always a severe distortion at the beginning of delayed signals, referred to as head distortion. This letter identifies the cause of head distortion and proposes a solution to this problem for reducing the overall distortion in FD filters. For the purpose of performance evaluation, relative root-mean-square (RMS) error is formulated as a metric to quantify the overall difference between the frequency-domain response of an FD filter and the ideal one. Moreover, illustrative numerical results on the proposed scheme applied in FD filters with classical sinc, Farrow and Lagrange interpolation substantiate the validity and feasibility of our solution
Rethinking Sensors Modeling: Hierarchical Information Enhanced Traffic Forecasting
With the acceleration of urbanization, traffic forecasting has become an
essential role in smart city construction. In the context of spatio-temporal
prediction, the key lies in how to model the dependencies of sensors. However,
existing works basically only consider the micro relationships between sensors,
where the sensors are treated equally, and their macroscopic dependencies are
neglected. In this paper, we argue to rethink the sensor's dependency modeling
from two hierarchies: regional and global perspectives. Particularly, we merge
original sensors with high intra-region correlation as a region node to
preserve the inter-region dependency. Then, we generate representative and
common spatio-temporal patterns as global nodes to reflect a global dependency
between sensors and provide auxiliary information for spatio-temporal
dependency learning. In pursuit of the generality and reality of node
representations, we incorporate a Meta GCN to calibrate the regional and global
nodes in the physical data space. Furthermore, we devise the cross-hierarchy
graph convolution to propagate information from different hierarchies. In a
nutshell, we propose a Hierarchical Information Enhanced Spatio-Temporal
prediction method, HIEST, to create and utilize the regional dependency and
common spatio-temporal patterns. Extensive experiments have verified the
leading performance of our HIEST against state-of-the-art baselines. We
publicize the code to ease reproducibility.Comment: 9 pages, accepted by CIKM'2
Survival in Patients With Metastatic Prostate Cancer Undergoing Radiotherapy: The Importance of Prostate-Specific Antigen-Based Stratification
ObjectivesTo explore the effectiveness of radiotherapy in mPCa patients with different PSA stratifications based on the cancer database of a large population.BackgroundScreening criteria for patients with metastatic prostate cancer, who are candidates for radiotherapy, are rarely reported.Patients and MethodsWe identified 22,604 patients with metastatic prostate cancer in the Surveillance, Epidemiology, and End Results database and divided them into a radiotherapy group and a control group. Patients with metastatic prostate cancer were divided into subgroups according to their levels of prostate-specific antigen to evaluate the efficacy of radiotherapy. They were also divided into six subgroups according to their prostate-specific antigen levels. We used multivariate Cox analysis to evaluate overall survival and cancer-specific survival. After 1:1 propensity score matching, Kaplan-Meier analysis was used to explore the difference in overall survival and cancer-specific survival in the radiotherapy and control group.ResultsIn all, 5,505 patients received radiotherapy, compared to 17,099 in the control group. In the multivariate Cox analysis, radiotherapy improved overall survival (hazard ratio [HR]: 0.730, 95% confidence interval [CI]: 0.636–0.838; P<0.001) and cancer-specific survival (HR: 0.764, 95% CI: 0.647–0.903; P=0.002) in patients with a PSA level of 4–10 ng/mL. Similar results were obtained by Kaplan-Meier analysis after 1:1 propensity score matching. In patients with prostate-specific antigen levels between 4–10 ng/mL, the overall survival (P<0.001) and cancer-specific survival (P<0.05) in the radiotherapy group was significantly better than those in the control group.ConclusionThe result of this large population-based study shows that rigorous selection of appropriate metastatic prostate cancer patients for radiotherapy can benefit prognosis significantly. This can be the basis for future prospective trials
ContraGen: Effective Contrastive Learning For Causal Language Model
Despite exciting progress in large-scale language generation, the
expressiveness of its representations is severely limited by the
\textit{anisotropy} issue where the hidden representations are distributed into
a narrow cone in the vector space. To address this issue, we present ContraGen,
a novel contrastive learning framework to improve the representation with
better uniformity and discrimination. We assess ContraGen on a wide range of
downstream tasks in natural and programming languages. We show that ContraGen
can effectively enhance both uniformity and discrimination of the
representations and lead to the desired improvement on various language
understanding tasks where discriminative representations are crucial for
attaining good performance. Specifically, we attain relative improvement
on the Semantic Textual Similarity tasks and on Code-to-Code Search
tasks. Furthermore, by improving the expressiveness of the representations,
ContraGen also boosts the source code generation capability with relative
improvement on execution accuracy on the HumanEval benchmark.Comment: 10 page
- …