17,676 research outputs found
Learning to Optimize Domain Specific Normalization for Domain Generalization
We propose a simple but effective multi-source domain generalization
technique based on deep neural networks by incorporating optimized
normalization layers that are specific to individual domains. Our approach
employs multiple normalization methods while learning separate affine
parameters per domain. For each domain, the activations are normalized by a
weighted average of multiple normalization statistics. The normalization
statistics are kept track of separately for each normalization type if
necessary. Specifically, we employ batch and instance normalizations in our
implementation to identify the best combination of these two normalization
methods in each domain. The optimized normalization layers are effective to
enhance the generalizability of the learned model. We demonstrate the
state-of-the-art accuracy of our algorithm in the standard domain
generalization benchmarks, as well as viability to further tasks such as
multi-source domain adaptation and domain generalization in the presence of
label noise
Collaborative Optimization and Aggregation for Decentralized Domain Generalization and Adaptation
Contemporary domain generalization (DG) and multi-source unsupervised domain adaptation (UDA) methods mostly collect data from multiple domains together for joint optimization. However, this centralized training paradigm poses a threat to data privacy and is not applicable when data are non-shared across domains. In this work, we propose a new approach called Collaborative Optimization and Aggregation (COPA), which aims at optimizing a generalized target model for decentralized DG and UDA, where data from different domains are non-shared and private. Our base model consists of a domain-invariant feature extractor and an ensemble of domain-specific classifiers. In an iterative learning process, we optimize a local model for each domain, and then centrally aggregate local feature extractors and assemble domain-specific classifiers to construct a generalized global model, without sharing data from different domains. To improve generalization of feature extractors, we employ hybrid batch-instance normalization and collaboration of frozen classifiers. For better decentralized UDA, we further introduce a prediction agreement mechanism to overcome local disparities towards central model aggregation. Extensive experiments on five DG and UDA benchmark datasets show that COPA is capable of achieving comparable performance against the state-of-the-art DG and UDA methods without the need for centralized data collection in model training
Kitting in the Wild through Online Domain Adaptation
Technological developments call for increasing perception and action capabilities of robots. Among other skills, vision systems that can adapt to any possible change in the working conditions are needed. Since these conditions are unpredictable, we need benchmarks which allow to assess the generalization and robustness capabilities of our visual recognition algorithms. In this work we focus on robotic kitting in unconstrained scenarios. As a first contribution, we present a new visual dataset for the kitting task. Differently from standard object recognition datasets, we provide images of the same objects acquired under various conditions where camera, illumination and background are changed. This novel dataset allows for testing the robustness of robot visual recognition algorithms to a series of different domain shifts both in isolation and unified. Our second contribution is a novel online adaptation algorithm for deep models, based on batch-normalization layers, which allows to continuously adapt a model to the current working conditions. Differently from standard domain adaptation algorithms, it does not require any image from the target domain at training time. We benchmark the performance of the algorithm on the proposed dataset, showing its capability to fill the gap between the performances of a standard architecture and its counterpart adapted offline to the given target domain
Data Dropout: Optimizing Training Data for Convolutional Neural Networks
Deep learning models learn to fit training data while they are highly
expected to generalize well to testing data. Most works aim at finding such
models by creatively designing architectures and fine-tuning parameters. To
adapt to particular tasks, hand-crafted information such as image prior has
also been incorporated into end-to-end learning. However, very little progress
has been made on investigating how an individual training sample will influence
the generalization ability of a model. In other words, to achieve high
generalization accuracy, do we really need all the samples in a training
dataset? In this paper, we demonstrate that deep learning models such as
convolutional neural networks may not favor all training samples, and
generalization accuracy can be further improved by dropping those unfavorable
samples. Specifically, the influence of removing a training sample is
quantifiable, and we propose a Two-Round Training approach, aiming to achieve
higher generalization accuracy. We locate unfavorable samples after the first
round of training, and then retrain the model from scratch with the reduced
training dataset in the second round. Since our approach is essentially
different from fine-tuning or further training, the computational cost should
not be a concern. Our extensive experimental results indicate that, with
identical settings, the proposed approach can boost performance of the
well-known networks on both high-level computer vision problems such as image
classification, and low-level vision problems such as image denoising
AdaGraph: Unifying Predictive and Continuous Domain Adaptation through Graphs
The ability to categorize is a cornerstone of visual intelligence, and a key
functionality for artificial, autonomous visual machines. This problem will
never be solved without algorithms able to adapt and generalize across visual
domains. Within the context of domain adaptation and generalization, this paper
focuses on the predictive domain adaptation scenario, namely the case where no
target data are available and the system has to learn to generalize from
annotated source images plus unlabeled samples with associated metadata from
auxiliary domains. Our contributionis the first deep architecture that tackles
predictive domainadaptation, able to leverage over the information broughtby
the auxiliary domains through a graph. Moreover, we present a simple yet
effective strategy that allows us to take advantage of the incoming target data
at test time, in a continuous domain adaptation scenario. Experiments on three
benchmark databases support the value of our approach.Comment: CVPR 2019 (oral
Many Task Learning with Task Routing
Typical multi-task learning (MTL) methods rely on architectural adjustments
and a large trainable parameter set to jointly optimize over several tasks.
However, when the number of tasks increases so do the complexity of the
architectural adjustments and resource requirements. In this paper, we
introduce a method which applies a conditional feature-wise transformation over
the convolutional activations that enables a model to successfully perform a
large number of tasks. To distinguish from regular MTL, we introduce Many Task
Learning (MaTL) as a special case of MTL where more than 20 tasks are performed
by a single model. Our method dubbed Task Routing (TR) is encapsulated in a
layer we call the Task Routing Layer (TRL), which applied in an MaTL scenario
successfully fits hundreds of classification tasks in one model. We evaluate
our method on 5 datasets against strong baselines and state-of-the-art
approaches.Comment: 8 Pages, 5 Figures, 2 Table
- …