99 research outputs found
Poster: Self-Supervised Quantization-Aware Knowledge Distillation
Quantization-aware training (QAT) starts with a pre-trained full-precision
model and performs quantization during retraining. However, existing QAT works
require supervision from the labels and they suffer from accuracy loss due to
reduced precision. To address these limitations, this paper proposes a novel
Self-Supervised Quantization-Aware Knowledge Distillation framework (SQAKD).
SQAKD first unifies the forward and backward dynamics of various quantization
functions and then reframes QAT as a co-optimization problem that
simultaneously minimizes the KL-Loss and the discretization error, in a
self-supervised manner. The evaluation shows that SQAKD significantly improves
the performance of various state-of-the-art QAT works. SQAKD establishes
stronger baselines and does not require extensive labeled training data,
potentially making state-of-the-art QAT research more accessible
GlobalTrack: A Simple and Strong Baseline for Long-term Tracking
A key capability of a long-term tracker is to search for targets in very
large areas (typically the entire image) to handle possible target absences or
tracking failures. However, currently there is a lack of such a strong baseline
for global instance search. In this work, we aim to bridge this gap.
Specifically, we propose GlobalTrack, a pure global instance search based
tracker that makes no assumption on the temporal consistency of the target's
positions and scales. GlobalTrack is developed based on two-stage object
detectors, and it is able to perform full-image and multi-scale search of
arbitrary instances with only a single query as the guide. We further propose a
cross-query loss to improve the robustness of our approach against distractors.
With no online learning, no punishment on position or scale changes, no scale
smoothing and no trajectory refinement, our pure global instance search based
tracker achieves comparable, sometimes much better performance on four
large-scale tracking benchmarks (i.e., 52.1% AUC on LaSOT, 63.8% success rate
on TLP, 60.3% MaxGM on OxUvA and 75.4% normalized precision on TrackingNet),
compared to state-of-the-art approaches that typically require complex
post-processing. More importantly, our tracker runs without cumulative errors,
i.e., any type of temporary tracking failures will not affect its performance
on future frames, making it ideal for long-term tracking. We hope this work
will be a strong baseline for long-term tracking and will stimulate future
works in this area. Code is available at
https://github.com/huanglianghua/GlobalTrack.Comment: Accepted in AAAI202
FPGA-based Degradation Evaluation for Traction Power Module with Deep Recurrent Autoencoder
The timely and quantitative evaluation of the degradation is crucial for traction inverter systems in railway applications. The implementation in the industry is impeded by two major challenges including the varying operational profiles and the scalability for system-level applications. This paper proposes a deep recurrent autoencoder-based degradation evaluation method, to assess the degradation level of the traction power module online. The recurrent structure is embedded for processing multivariate time series condition monitoring data stream, in order to exploit the inherent time dependence to improve the accuracy and robustness. The autoencoder-based framework enables the scalability of the proposed method to system-level applications and can be applied under varying operating conditions. The method is experimentally demonstrated on an FPGA-based hardware platform.</p
QueryProp: Object Query Propagation for High-Performance Video Object Detection
Video object detection has been an important yet challenging topic in
computer vision. Traditional methods mainly focus on designing the image-level
or box-level feature propagation strategies to exploit temporal information.
This paper argues that with a more effective and efficient feature propagation
framework, video object detectors can gain improvement in terms of both
accuracy and speed. For this purpose, this paper studies object-level feature
propagation, and proposes an object query propagation (QueryProp) framework for
high-performance video object detection. The proposed QueryProp contains two
propagation strategies: 1) query propagation is performed from sparse key
frames to dense non-key frames to reduce the redundant computation on non-key
frames; 2) query propagation is performed from previous key frames to the
current key frame to improve feature representation by temporal context
modeling. To further facilitate query propagation, an adaptive propagation gate
is designed to achieve flexible key frame selection. We conduct extensive
experiments on the ImageNet VID dataset. QueryProp achieves comparable accuracy
with state-of-the-art methods and strikes a decent accuracy/speed trade-off.
Code is available at https://github.com/hf1995/QueryProp.Comment: This paper is accepted to AAAI202
Nonparametric Classification on Low Dimensional Manifolds using Overparameterized Convolutional Residual Networks
Convolutional residual neural networks (ConvResNets), though
overparameterized, can achieve remarkable prediction performance in practice,
which cannot be well explained by conventional wisdom. To bridge this gap, we
study the performance of ConvResNeXts, which cover ConvResNets as a special
case, trained with weight decay from the perspective of nonparametric
classification. Our analysis allows for infinitely many building blocks in
ConvResNeXts, and shows that weight decay implicitly enforces sparsity on these
blocks. Specifically, we consider a smooth target function supported on a
low-dimensional manifold, then prove that ConvResNeXts can adapt to the
function smoothness and low-dimensional structures and efficiently learn the
function without suffering from the curse of dimensionality. Our findings
partially justify the advantage of overparameterized ConvResNeXts over
conventional machine learning models.Comment: 20 pages, 1 figur
- …