2 research outputs found
Let Me Know What to Ask: Interrogative-Word-Aware Question Generation
Question Generation (QG) is a Natural Language Processing (NLP) task that
aids advances in Question Answering (QA) and conversational assistants.
Existing models focus on generating a question based on a text and possibly the
answer to the generated question. They need to determine the type of
interrogative word to be generated while having to pay attention to the grammar
and vocabulary of the question. In this work, we propose
Interrogative-Word-Aware Question Generation (IWAQG), a pipelined system
composed of two modules: an interrogative word classifier and a QG model. The
first module predicts the interrogative word that is provided to the second
module to create the question. Owing to an increased recall of deciding the
interrogative words to be used for the generated questions, the proposed model
achieves new state-of-the-art results on the task of QG in SQuAD, improving
from 46.58 to 47.69 in BLEU-1, 17.55 to 18.53 in BLEU-4, 21.24 to 22.33 in
METEOR, and from 44.53 to 46.94 in ROUGE-L.Comment: Accepted at 2nd Workshop on Machine Reading for Question Answering
(MRQA), EMNLP 201
Regularization of Distinct Strategies for Unsupervised Question Generation
Unsupervised question answering (UQA) has been proposed to avoid the high cost of creating high-quality datasets for QA. One approach to UQA is to train a QA model with questions generated automatically. However, the generated questions are either too similar to a word sequence in the context or too drifted from the semantics of the context, thereby making it difficult to train a robust QA model. We propose a novel regularization method based on teacher-student architecture to avoid bias toward a particular question generation strategy and modulate the process of generating individual words when a question is generated. Our experiments demonstrate that we have achieved the goal of generating higher-quality questions for UQA across diverse QA datasets and tasks. We also show that this method can be useful for creating a QA model with few-shot learning