95 research outputs found
Semi-Supervised Learning for Neural Keyphrase Generation
We study the problem of generating keyphrases that summarize the key points
for a given document. While sequence-to-sequence (seq2seq) models have achieved
remarkable performance on this task (Meng et al., 2017), model training often
relies on large amounts of labeled data, which is only applicable to
resource-rich domains. In this paper, we propose semi-supervised keyphrase
generation methods by leveraging both labeled data and large-scale unlabeled
samples for learning. Two strategies are proposed. First, unlabeled documents
are first tagged with synthetic keyphrases obtained from unsupervised keyphrase
extraction methods or a selflearning algorithm, and then combined with labeled
samples for training. Furthermore, we investigate a multi-task learning
framework to jointly learn to generate keyphrases as well as the titles of the
articles. Experimental results show that our semi-supervised learning-based
methods outperform a state-of-the-art model trained with labeled data only.Comment: To appear in EMNLP 2018 (12 pages, 7 figures, 6 tables
WriterForcing: Generating more interesting story endings
We study the problem of generating interesting endings for stories. Neural
generative models have shown promising results for various text generation
problems. Sequence to Sequence (Seq2Seq) models are typically trained to
generate a single output sequence for a given input sequence. However, in the
context of a story, multiple endings are possible. Seq2Seq models tend to
ignore the context and generate generic and dull responses. Very few works have
studied generating diverse and interesting story endings for a given story
context. In this paper, we propose models which generate more diverse and
interesting outputs by 1) training models to focus attention on important
keyphrases of the story, and 2) promoting generation of non-generic words. We
show that the combination of the two leads to more diverse and interesting
endings.Comment: Accepted in ACL workshop on Storytelling 201
- …