534 research outputs found
Cross-stitch Networks for Multi-task Learning
Multi-task learning in Convolutional Networks has displayed remarkable
success in the field of recognition. This success can be largely attributed to
learning shared representations from multiple supervisory tasks. However,
existing multi-task approaches rely on enumerating multiple network
architectures specific to the tasks at hand, that do not generalize. In this
paper, we propose a principled approach to learn shared representations in
ConvNets using multi-task learning. Specifically, we propose a new sharing
unit: "cross-stitch" unit. These units combine the activations from multiple
networks and can be trained end-to-end. A network with cross-stitch units can
learn an optimal combination of shared and task-specific representations. Our
proposed method generalizes across multiple tasks and shows dramatically
improved performance over baseline methods for categories with few training
examples.Comment: To appear in CVPR 2016 (Spotlight
Beyond One-hot Encoding: lower dimensional target embedding
Target encoding plays a central role when learning Convolutional Neural
Networks. In this realm, One-hot encoding is the most prevalent strategy due to
its simplicity. However, this so widespread encoding schema assumes a flat
label space, thus ignoring rich relationships existing among labels that can be
exploited during training. In large-scale datasets, data does not span the full
label space, but instead lies in a low-dimensional output manifold. Following
this observation, we embed the targets into a low-dimensional space,
drastically improving convergence speed while preserving accuracy. Our
contribution is two fold: (i) We show that random projections of the label
space are a valid tool to find such lower dimensional embeddings, boosting
dramatically convergence rates at zero computational cost; and (ii) we propose
a normalized eigenrepresentation of the class manifold that encodes the targets
with minimal information loss, improving the accuracy of random projections
encoding while enjoying the same convergence rates. Experiments on CIFAR-100,
CUB200-2011, Imagenet, and MIT Places demonstrate that the proposed approach
drastically improves convergence speed while reaching very competitive accuracy
rates.Comment: Published at Image and Vision Computin
- …