2 research outputs found
A Deep Reinforced Model for Abstractive Summarization
Attentional, RNN-based encoder-decoder models for abstractive summarization
have achieved good performance on short input and output sequences. For longer
documents and summaries however these models often include repetitive and
incoherent phrases. We introduce a neural network model with a novel
intra-attention that attends over the input and continuously generated output
separately, and a new training method that combines standard supervised word
prediction and reinforcement learning (RL). Models trained only with supervised
learning often exhibit "exposure bias" - they assume ground truth is provided
at each step during training. However, when standard word prediction is
combined with the global sequence prediction training of RL the resulting
summaries become more readable. We evaluate this model on the CNN/Daily Mail
and New York Times datasets. Our model obtains a 41.16 ROUGE-1 score on the
CNN/Daily Mail dataset, an improvement over previous state-of-the-art models.
Human evaluation also shows that our model produces higher quality summaries
Plain English Summarization of Contracts
Unilateral contracts, such as terms of service, play a substantial role in
modern digital life. However, few users read these documents before accepting
the terms within, as they are too long and the language too complicated. We
propose the task of summarizing such legal documents in plain English, which
would enable users to have a better understanding of the terms they are
accepting.
We propose an initial dataset of legal text snippets paired with summaries
written in plain English. We verify the quality of these summaries manually and
show that they involve heavy abstraction, compression, and simplification.
Initial experiments show that unsupervised extractive summarization methods do
not perform well on this task due to the level of abstraction and style
differences. We conclude with a call for resource and technique development for
simplification and style transfer for legal language.Comment: The first Workshop on Natural Legal Language Processing (NLLP) will
be co-located with NAACL 2019 in Minneapolis, Minnesota, US