8,656 research outputs found
Open Set Domain Adaptation by Backpropagation
Numerous algorithms have been proposed for transferring knowledge from a
label-rich domain (source) to a label-scarce domain (target). Almost all of
them are proposed for a closed-set scenario, where the source and the target
domain completely share the class of their samples. We call the shared class
the \doublequote{known class.} However, in practice, when samples in target
domain are not labeled, we cannot know whether the domains share the class. A
target domain can contain samples of classes that are not shared by the source
domain. We call such classes the \doublequote{unknown class} and algorithms
that work well in the open set situation are very practical. However, most
existing distribution matching methods for domain adaptation do not work well
in this setting because unknown target samples should not be aligned with the
source.
In this paper, we propose a method for an open set domain adaptation scenario
which utilizes adversarial training. A classifier is trained to make a boundary
between the source and the target samples whereas a generator is trained to
make target samples far from the boundary. Thus, we assign two options to the
feature generator: aligning them with source known samples or rejecting them as
unknown target samples. This approach allows extracting features that separate
unknown target samples from known target samples. Our method was extensively
evaluated in domain adaptation setting and outperformed other methods with a
large margin in most settings.Comment: Accepted by ECCV201
Fast Context Adaptation via Meta-Learning
We propose CAVIA for meta-learning, a simple extension to MAML that is less
prone to meta-overfitting, easier to parallelise, and more interpretable. CAVIA
partitions the model parameters into two parts: context parameters that serve
as additional input to the model and are adapted on individual tasks, and
shared parameters that are meta-trained and shared across tasks. At test time,
only the context parameters are updated, leading to a low-dimensional task
representation. We show empirically that CAVIA outperforms MAML for regression,
classification, and reinforcement learning. Our experiments also highlight
weaknesses in current benchmarks, in that the amount of adaptation needed in
some cases is small.Comment: Published at the International Conference on Machine Learning (ICML)
201
Neuromorphic Hardware In The Loop: Training a Deep Spiking Network on the BrainScaleS Wafer-Scale System
Emulating spiking neural networks on analog neuromorphic hardware offers
several advantages over simulating them on conventional computers, particularly
in terms of speed and energy consumption. However, this usually comes at the
cost of reduced control over the dynamics of the emulated networks. In this
paper, we demonstrate how iterative training of a hardware-emulated network can
compensate for anomalies induced by the analog substrate. We first convert a
deep neural network trained in software to a spiking network on the BrainScaleS
wafer-scale neuromorphic system, thereby enabling an acceleration factor of 10
000 compared to the biological time domain. This mapping is followed by the
in-the-loop training, where in each training step, the network activity is
first recorded in hardware and then used to compute the parameter updates in
software via backpropagation. An essential finding is that the parameter
updates do not have to be precise, but only need to approximately follow the
correct gradient, which simplifies the computation of updates. Using this
approach, after only several tens of iterations, the spiking network shows an
accuracy close to the ideal software-emulated prototype. The presented
techniques show that deep spiking networks emulated on analog neuromorphic
devices can attain good computational performance despite the inherent
variations of the analog substrate.Comment: 8 pages, 10 figures, submitted to IJCNN 201
- …