208,230 research outputs found
EFANet: Exchangeable Feature Alignment Network for Arbitrary Style Transfer
Style transfer has been an important topic both in computer vision and
graphics. Since the seminal work of Gatys et al. first demonstrates the power
of stylization through optimization in the deep feature space, quite a few
approaches have achieved real-time arbitrary style transfer with
straightforward statistic matching techniques. In this work, our key
observation is that only considering features in the input style image for the
global deep feature statistic matching or local patch swap may not always
ensure a satisfactory style transfer; see e.g., Figure 1. Instead, we propose a
novel transfer framework, EFANet, that aims to jointly analyze and better align
exchangeable features extracted from content and style image pair. In this way,
the style features from the style image seek for the best compatibility with
the content information in the content image, leading to more structured
stylization results. In addition, a new whitening loss is developed for
purifying the computed content features and better fusion with styles in
feature space. Qualitative and quantitative experiments demonstrate the
advantages of our approach.Comment: Accepted by AAAI 202
Many Task Learning with Task Routing
Typical multi-task learning (MTL) methods rely on architectural adjustments
and a large trainable parameter set to jointly optimize over several tasks.
However, when the number of tasks increases so do the complexity of the
architectural adjustments and resource requirements. In this paper, we
introduce a method which applies a conditional feature-wise transformation over
the convolutional activations that enables a model to successfully perform a
large number of tasks. To distinguish from regular MTL, we introduce Many Task
Learning (MaTL) as a special case of MTL where more than 20 tasks are performed
by a single model. Our method dubbed Task Routing (TR) is encapsulated in a
layer we call the Task Routing Layer (TRL), which applied in an MaTL scenario
successfully fits hundreds of classification tasks in one model. We evaluate
our method on 5 datasets against strong baselines and state-of-the-art
approaches.Comment: 8 Pages, 5 Figures, 2 Table
Bandit Structured Prediction for Neural Sequence-to-Sequence Learning
Bandit structured prediction describes a stochastic optimization framework
where learning is performed from partial feedback. This feedback is received in
the form of a task loss evaluation to a predicted output structure, without
having access to gold standard structures. We advance this framework by lifting
linear bandit learning to neural sequence-to-sequence learning problems using
attention-based recurrent neural networks. Furthermore, we show how to
incorporate control variates into our learning algorithms for variance
reduction and improved generalization. We present an evaluation on a neural
machine translation task that shows improvements of up to 5.89 BLEU points for
domain adaptation from simulated bandit feedback.Comment: ACL 201
- …
