1,116 research outputs found
HEGEL: Hypergraph Transformer for Long Document Summarization
Extractive summarization for long documents is challenging due to the
extended structured input context. The long-distance sentence dependency
hinders cross-sentence relations modeling, the critical step of extractive
summarization. This paper proposes HEGEL, a hypergraph neural network for long
document summarization by capturing high-order cross-sentence relations. HEGEL
updates and learns effective sentence representations with hypergraph
transformer layers and fuses different types of sentence dependencies,
including latent topics, keywords coreference, and section structure. We
validate HEGEL by conducting extensive experiments on two benchmark datasets,
and experimental results demonstrate the effectiveness and efficiency of HEGEL.Comment: EMNLP 202
Deep Recurrent Generative Decoder for Abstractive Text Summarization
We propose a new framework for abstractive text summarization based on a
sequence-to-sequence oriented encoder-decoder model equipped with a deep
recurrent generative decoder (DRGN).
Latent structure information implied in the target summaries is learned based
on a recurrent latent random model for improving the summarization quality.
Neural variational inference is employed to address the intractable posterior
inference for the recurrent latent variables.
Abstractive summaries are generated based on both the generative latent
variables and the discriminative deterministic states.
Extensive experiments on some benchmark datasets in different languages show
that DRGN achieves improvements over the state-of-the-art methods.Comment: 10 pages, EMNLP 201
- …