2,070 research outputs found
Angular Gap: Reducing the Uncertainty of Image Difficulty through Model Calibration
Curriculum learning needs example difficulty to proceed from easy to hard. However, the credibility of image difficulty is rarely investigated, which can seriously affect the effectiveness of curricula. In this work, we propose Angular Gap, a measure of difficulty based on the difference in angular distance between feature embeddings and class-weight embeddings built by hyperspherical learning. To ascertain difficulty estimation, we introduce class-wise model calibration, as a post-training technique, to the learnt hyperbolic space. This bridges the gap between probabilistic model calibration and angular distance estimation of hyperspherical learning. We show the superiority of our calibrated Angular Gap over recent difficulty metrics on CIFAR10-H and ImageNetV2. We further propose a curriculum based on Angular Gap for unsupervised domain adaptation that can translate from learning easy samples to mining hard samples. We combine this curriculum with a state-of-the-art self-training method, Cycle Self Training (CST). The proposed Curricular CST learns robust representations and outperforms recent baselines on Office31 and VisDA 2017
Confidence-Aware Paced-Curriculum Learning by Label Smoothing for Surgical Scene Understanding
Curriculum learning and self-paced learning are the training strategies that gradually feed the samples from easy to more complex. They have captivated increasing attention due to their excellent performance in robotic vision. Most recent works focus on designing curricula based on difficulty levels in input samples or smoothing the feature maps. However, smoothing labels to control the learning utility in a curriculum manner is still unexplored. In this work, we design a paced curriculum by label smoothing (P-CBLS) using paced learning with uniform label smoothing (ULS) for classification tasks and fuse uniform and spatially varying label smoothing (SVLS) for semantic segmentation tasks in a curriculum manner. In ULS and SVLS, a bigger smoothing factor value enforces a heavy smoothing penalty in the true label and limits learning less information. Therefore, we design the curriculum by label smoothing (CBLS). We set a bigger smoothing value at the beginning of training and gradually decreased it to zero to control the model learning utility from lower to higher. We also designed a confidence-aware pacing function and combined it with our CBLS to investigate the benefits of various curricula. The proposed techniques are validated on four robotic surgery datasets of multi-class, multi-label classification, captioning, and segmentation tasks. We also investigate the robustness of our method by corrupting validation data into different severity levels. Our extensive analysis shows that the proposed method improves prediction accuracy and robustness. The code is publicly available at https://github.com/XuMengyaAmy/P-CBLS. Note to Practitioners —The motivation of this article is to improve the performance and robustness of deep neural networks in safety-critical applications such as robotic surgery by controlling the learning ability of the model in a curriculum learning manner and allowing the model to imitate the cognitive process of humans and animals. The designed approaches do not add parameters that require additional computational resources
Soft Augmentation for Image Classification
Modern neural networks are over-parameterized and thus rely on strong
regularization such as data augmentation and weight decay to reduce overfitting
and improve generalization. The dominant form of data augmentation applies
invariant transforms, where the learning target of a sample is invariant to the
transform applied to that sample. We draw inspiration from human visual
classification studies and propose generalizing augmentation with invariant
transforms to soft augmentation where the learning target softens non-linearly
as a function of the degree of the transform applied to the sample: e.g., more
aggressive image crop augmentations produce less confident learning targets. We
demonstrate that soft targets allow for more aggressive data augmentation,
offer more robust performance boosts, work with other augmentation policies,
and interestingly, produce better calibrated models (since they are trained to
be less confident on aggressively cropped/occluded examples). Combined with
existing aggressive augmentation strategies, soft target 1) doubles the top-1
accuracy boost across Cifar-10, Cifar-100, ImageNet-1K, and ImageNet-V2, 2)
improves model occlusion performance by up to , and 3) halves the
expected calibration error (ECE). Finally, we show that soft augmentation
generalizes to self-supervised classification tasks
Is Signed Message Essential for Graph Neural Networks?
Message-passing Graph Neural Networks (GNNs), which collect information from
adjacent nodes, achieve satisfying results on homophilic graphs. However, their
performances are dismal in heterophilous graphs, and many researchers have
proposed a plethora of schemes to solve this problem. Especially, flipping the
sign of edges is rooted in a strong theoretical foundation, and attains
significant performance enhancements. Nonetheless, previous analyses assume a
binary class scenario and they may suffer from confined applicability. This
paper extends the prior understandings to multi-class scenarios and points out
two drawbacks: (1) the sign of multi-hop neighbors depends on the message
propagation paths and may incur inconsistency, (2) it also increases the
prediction uncertainty (e.g., conflict evidence) which can impede the stability
of the algorithm. Based on the theoretical understanding, we introduce a novel
strategy that is applicable to multi-class graphs. The proposed scheme combines
confidence calibration to secure robustness while reducing uncertainty. We show
the efficacy of our theorem through extensive experiments on six benchmark
graph datasets
- …