44 research outputs found
Contextual Augmentation: Data Augmentation by Words with Paradigmatic Relations
We propose a novel data augmentation for labeled sentences called contextual
augmentation. We assume an invariance that sentences are natural even if the
words in the sentences are replaced with other words with paradigmatic
relations. We stochastically replace words with other words that are predicted
by a bi-directional language model at the word positions. Words predicted
according to a context are numerous but appropriate for the augmentation of the
original words. Furthermore, we retrofit a language model with a
label-conditional architecture, which allows the model to augment sentences
without breaking the label-compatibility. Through the experiments for six
various different text classification tasks, we demonstrate that the proposed
method improves classifiers based on the convolutional or recurrent neural
networks.Comment: NAACL 201
Unsupervised Learning of Style-sensitive Word Vectors
This paper presents the first study aimed at capturing stylistic similarity
between words in an unsupervised manner. We propose extending the continuous
bag of words (CBOW) model (Mikolov et al., 2013) to learn style-sensitive word
vectors using a wider context window under the assumption that the style of all
the words in an utterance is consistent. In addition, we introduce a novel task
to predict lexical stylistic similarity and to create a benchmark dataset for
this task. Our experiment with this dataset supports our assumption and
demonstrates that the proposed extensions contribute to the acquisition of
style-sensitive word embeddings.Comment: 7 pages, Accepted at The 56th Annual Meeting of the Association for
Computational Linguistics (ACL 2018
ニューラルモデルの効率的訓練のためのサブネットワーク探索
要約のみTohoku University乾健太郎課