53 research outputs found
Heuristic algorithms for solving a class of multiobjective zero-one programming problems
Master'sMASTER OF ENGINEERIN
Collaborative Multi-Object Tracking with Conformal Uncertainty Propagation
Object detection and multiple object tracking (MOT) are essential components
of self-driving systems. Accurate detection and uncertainty quantification are
both critical for onboard modules, such as perception, prediction, and
planning, to improve the safety and robustness of autonomous vehicles.
Collaborative object detection (COD) has been proposed to improve detection
accuracy and reduce uncertainty by leveraging the viewpoints of multiple
agents. However, little attention has been paid to how to leverage the
uncertainty quantification from COD to enhance MOT performance. In this paper,
as the first attempt to address this challenge, we design an uncertainty
propagation framework called MOT-CUP. Our framework first quantifies the
uncertainty of COD through direct modeling and conformal prediction, and
propagates this uncertainty information into the motion prediction and
association steps. MOT-CUP is designed to work with different collaborative
object detectors and baseline MOT algorithms. We evaluate MOT-CUP on V2X-Sim, a
comprehensive collaborative perception dataset, and demonstrate a 2%
improvement in accuracy and a 2.67X reduction in uncertainty compared to the
baselines, e.g. SORT and ByteTrack. In scenarios characterized by high
occlusion levels, our MOT-CUP demonstrates a noteworthy improvement in
accuracy. MOT-CUP demonstrates the importance of uncertainty quantification in
both COD and MOT, and provides the first attempt to improve the accuracy and
reduce the uncertainty in MOT based on COD through uncertainty propagation. Our
code is public on https://coperception.github.io/MOT-CUP/.Comment: This paper has been accepted by IEEE Robotics and Automation Letter
Accelerating Dataset Distillation via Model Augmentation
Dataset Distillation (DD), a newly emerging field, aims at generating much
smaller and high-quality synthetic datasets from large ones. Existing DD
methods based on gradient matching achieve leading performance; however, they
are extremely computationally intensive as they require continuously optimizing
a dataset among thousands of randomly initialized models. In this paper, we
assume that training the synthetic data with diverse models leads to better
generalization performance. Thus we propose two \textbf{model augmentation}
techniques, ~\ie using \textbf{early-stage models} and \textbf{weight
perturbation} to learn an informative synthetic set with significantly reduced
training cost. Extensive experiments demonstrate that our method achieves up to
20 speedup and comparable performance on par with state-of-the-art
baseline methods
You Need Multiple Exiting: Dynamic Early Exiting for Accelerating Unified Vision Language Model
Large-scale Transformer models bring significant improvements for various
downstream vision language tasks with a unified architecture. The performance
improvements come with increasing model size, resulting in slow inference speed
and increased cost for severing. While some certain predictions benefit from
the full complexity of the large-scale model, not all of inputs need the same
amount of computation to conduct, potentially leading to computation resource
waste. To handle this challenge, early exiting is proposed to adaptively
allocate computational power in term of input complexity to improve inference
efficiency. The existing early exiting strategies usually adopt output
confidence based on intermediate layers as a proxy of input complexity to incur
the decision of skipping following layers. However, such strategies cannot
apply to encoder in the widely-used unified architecture with both encoder and
decoder due to difficulty of output confidence estimation in the encoder. It is
suboptimal in term of saving computation power to ignore the early exiting in
encoder component. To handle this challenge, we propose a novel early exiting
strategy for unified visual language models, which allows dynamically skip the
layers in encoder and decoder simultaneously in term of input layer-wise
similarities with multiple times of early exiting, namely \textbf{MuE}. By
decomposing the image and text modalities in the encoder, MuE is flexible and
can skip different layers in term of modalities, advancing the inference
efficiency while minimizing performance drop. Experiments on the SNLI-VE and MS
COCO datasets show that the proposed approach MuE can reduce expected inference
time by up to 50\% and 40\% while maintaining 99\% and 96\% performance
respectively
A New Species of the Genus Achalinus (Squamata: Xenodermidae) from the Dabie Mountains, Anhui, China
A new species of Xenodermid snake, Achalinus dabieshanensis sp. nov., was described based on three specimens (two female and one male) collected from the Dabie Mountains of western Anhui Province. It can be distinguished from known congeners by a significant genetic divergence in the mitochondrial gene fragment COI (p-distance ≥ 9.4%) and the following combination of characteristics: (1) length of the suture between the internasals being distinctly shorter than between the prefrontals; (2) a single loreal; (3) dorsal scales strongly keeled, in 23 rows throughout the body; (4) two pairs of prefrontals; (5) six supralabials; (6) five infralabials; (7) temporals 2 + 2 + 3 (or 2 + 2 + 4); (8) 141–155 ventrals; (9) 45–55 subcaudals, unpaired; (10) anal entire; (11) weakly iridescent tinged, uniform, brown to black dorsum with vertebral scales and about three adjacent dorsal scales dark brown forming a longitudinal vertebral line from posterior margin of parietals to tail tip; (12) light brown venter, ventral shields wide, visible on both sides, light brown flanks, giving the appearance of a black subcaudal streak. The recognition of the new species increases the number of described Achalinus species to 22
- …