102 research outputs found
Association Graph Learning for Multi-Task Classification with Category Shifts
In this paper, we focus on multi-task classification, where related
classification tasks share the same label space and are learned simultaneously.
In particular, we tackle a new setting, which is more realistic than currently
addressed in the literature, where categories shift from training to test data.
Hence, individual tasks do not contain complete training data for the
categories in the test set. To generalize to such test data, it is crucial for
individual tasks to leverage knowledge from related tasks. To this end, we
propose learning an association graph to transfer knowledge among tasks for
missing classes. We construct the association graph with nodes representing
tasks, classes and instances, and encode the relationships among the nodes in
the edges to guide their mutual knowledge transfer. By message passing on the
association graph, our model enhances the categorical information of each
instance, making it more discriminative. To avoid spurious correlations between
task and class nodes in the graph, we introduce an assignment entropy
maximization that encourages each class node to balance its edge weights. This
enables all tasks to fully utilize the categorical information from related
tasks. An extensive evaluation on three general benchmarks and a medical
dataset for skin lesion classification reveals that our method consistently
performs better than representative baselines
A Bit More Bayesian: Domain-Invariant Learning with Uncertainty
Domain generalization is challenging due to the domain shift and the
uncertainty caused by the inaccessibility of target domain data. In this paper,
we address both challenges with a probabilistic framework based on variational
Bayesian inference, by incorporating uncertainty into neural network weights.
We couple domain invariance in a probabilistic formula with the variational
Bayesian inference. This enables us to explore domain-invariant learning in a
principled way. Specifically, we derive domain-invariant representations and
classifiers, which are jointly established in a two-layer Bayesian neural
network. We empirically demonstrate the effectiveness of our proposal on four
widely used cross-domain visual recognition benchmarks. Ablation studies
validate the synergistic benefits of our Bayesian treatment when jointly
learning domain-invariant representations and classifiers for domain
generalization. Further, our method consistently delivers state-of-the-art mean
accuracy on all benchmarks.Comment: accepted to ICML 202
Learning Variational Neighbor Labels for Test-Time Domain Generalization
This paper strives for domain generalization, where models are trained
exclusively on source domains before being deployed at unseen target domains.
We follow the strict separation of source training and target testing but
exploit the value of the unlabeled target data itself during inference. We make
three contributions. First, we propose probabilistic pseudo-labeling of target
samples to generalize the source-trained model to the target domain at test
time. We formulate the generalization at test time as a variational inference
problem by modeling pseudo labels as distributions to consider the uncertainty
during generalization and alleviate the misleading signal of inaccurate pseudo
labels. Second, we learn variational neighbor labels that incorporate the
information of neighboring target samples to generate more robust pseudo
labels. Third, to learn the ability to incorporate more representative target
information and generate more precise and robust variational neighbor labels,
we introduce a meta-generalization stage during training to simulate the
generalization procedure. Experiments on six widely-used datasets demonstrate
the benefits, abilities, and effectiveness of our proposal.Comment: Under revie
Any-Shift Prompting for Generalization over Distributions
Image-language models with prompt learning have shown remarkable advances in
numerous downstream vision tasks. Nevertheless, conventional prompt learning
methods overfit their training distribution and lose the generalization ability
on test distributions. To improve generalization across various distribution
shifts, we propose any-shift prompting: a general probabilistic inference
framework that considers the relationship between training and test
distributions during prompt learning. We explicitly connect training and test
distributions in the latent space by constructing training and test prompts in
a hierarchical architecture. Within this framework, the test prompt exploits
the distribution relationships to guide the generalization of the CLIP
image-language model from training to any test distribution. To effectively
encode the distribution information and their relationships, we further
introduce a transformer inference network with a pseudo-shift training
mechanism. The network generates the tailored test prompt with both training
and test information in a feedforward pass, avoiding extra training costs at
test time. Extensive experiments on twenty-three datasets demonstrate the
effectiveness of any-shift prompting on the generalization over various
distribution shifts
- …