2,965 research outputs found
Abstractive Text Summarization by Incorporating Reader Comments
In neural abstractive summarization field, conventional sequence-to-sequence
based models often suffer from summarizing the wrong aspect of the document
with respect to the main aspect. To tackle this problem, we propose the task of
reader-aware abstractive summary generation, which utilizes the reader comments
to help the model produce better summary about the main aspect. Unlike
traditional abstractive summarization task, reader-aware summarization
confronts two main challenges: (1) Comments are informal and noisy; (2) jointly
modeling the news document and the reader comments is challenging. To tackle
the above challenges, we design an adversarial learning model named
reader-aware summary generator (RASG), which consists of four components: (1) a
sequence-to-sequence based summary generator; (2) a reader attention module
capturing the reader focused aspects; (3) a supervisor modeling the semantic
gap between the generated summary and reader focused aspects; (4) a goal
tracker producing the goal for each generation step. The supervisor and the
goal tacker are used to guide the training of our framework in an adversarial
manner. Extensive experiments are conducted on our large-scale real-world text
summarization dataset, and the results show that RASG achieves the
state-of-the-art performance in terms of both automatic metrics and human
evaluations. The experimental results also demonstrate the effectiveness of
each module in our framework. We release our large-scale dataset for further
research.Comment: Accepted by AAAI 201
Deep Recurrent Generative Decoder for Abstractive Text Summarization
We propose a new framework for abstractive text summarization based on a
sequence-to-sequence oriented encoder-decoder model equipped with a deep
recurrent generative decoder (DRGN).
Latent structure information implied in the target summaries is learned based
on a recurrent latent random model for improving the summarization quality.
Neural variational inference is employed to address the intractable posterior
inference for the recurrent latent variables.
Abstractive summaries are generated based on both the generative latent
variables and the discriminative deterministic states.
Extensive experiments on some benchmark datasets in different languages show
that DRGN achieves improvements over the state-of-the-art methods.Comment: 10 pages, EMNLP 201
- …