32,749 research outputs found
Understanding language-elicited EEG data by predicting it from a fine-tuned language model
Electroencephalography (EEG) recordings of brain activity taken while
participants read or listen to language are widely used within the cognitive
neuroscience and psycholinguistics communities as a tool to study language
comprehension. Several time-locked stereotyped EEG responses to
word-presentations -- known collectively as event-related potentials (ERPs) --
are thought to be markers for semantic or syntactic processes that take place
during comprehension. However, the characterization of each individual ERP in
terms of what features of a stream of language trigger the response remains
controversial. Improving this characterization would make ERPs a more useful
tool for studying language comprehension. We take a step towards better
understanding the ERPs by fine-tuning a language model to predict them. This
new approach to analysis shows for the first time that all of the ERPs are
predictable from embeddings of a stream of language. Prior work has only found
two of the ERPs to be predictable. In addition to this analysis, we examine
which ERPs benefit from sharing parameters during joint training. We find that
two pairs of ERPs previously identified in the literature as being related to
each other benefit from joint training, while several other pairs of ERPs that
benefit from joint training are suggestive of potential relationships.
Extensions of this analysis that further examine what kinds of information in
the model embeddings relate to each ERP have the potential to elucidate the
processes involved in human language comprehension.Comment: To appear in Proceedings of the 2019 Conference of the North American
Chapter of the Association for Computational Linguistic
Incorporating Structured Commonsense Knowledge in Story Completion
The ability to select an appropriate story ending is the first step towards
perfect narrative comprehension. Story ending prediction requires not only the
explicit clues within the context, but also the implicit knowledge (such as
commonsense) to construct a reasonable and consistent story. However, most
previous approaches do not explicitly use background commonsense knowledge. We
present a neural story ending selection model that integrates three types of
information: narrative sequence, sentiment evolution and commonsense knowledge.
Experiments show that our model outperforms state-of-the-art approaches on a
public dataset, ROCStory Cloze Task , and the performance gain from adding the
additional commonsense knowledge is significant
Teaching Machines to Read and Comprehend
Teaching machines to read natural language documents remains an elusive
challenge. Machine reading systems can be tested on their ability to answer
questions posed on the contents of documents that they have seen, but until now
large scale training and test datasets have been missing for this type of
evaluation. In this work we define a new methodology that resolves this
bottleneck and provides large scale supervised reading comprehension data. This
allows us to develop a class of attention based deep neural networks that learn
to read real documents and answer complex questions with minimal prior
knowledge of language structure.Comment: Appears in: Advances in Neural Information Processing Systems 28
(NIPS 2015). 14 pages, 13 figure
- …