4,338 research outputs found
Query-Based Summarization using Rhetorical Structure Theory
Research on Question Answering is focused mainly on classifying the question type and finding
the answer. Presenting the answer in a way that suits the userâs needs has received little
attention. This paper shows how existing question answering systemsâwhich aim at finding
precise answers to questionsâcan be improved by exploiting summarization techniques to extract
more than just the answer from the document in which the answer resides. This is done
using a graph search algorithm which searches for relevant sentences in the discourse structure,
which is represented as a graph. The Rhetorical Structure Theory (RST) is used to create a
graph representation of a text document. The output is an extensive answer, which not only
answers the question, but also gives the user an opportunity to assess the accuracy of the answer
(is this what I am looking for?), and to find additional information that is related to the question,
and which may satisfy an information need. This has been implemented in a working multimodal
question answering system where it operates with two independently developed question
answering modules
Graph-based Neural Multi-Document Summarization
We propose a neural multi-document summarization (MDS) system that
incorporates sentence relation graphs. We employ a Graph Convolutional Network
(GCN) on the relation graphs, with sentence embeddings obtained from Recurrent
Neural Networks as input node features. Through multiple layer-wise
propagation, the GCN generates high-level hidden sentence features for salience
estimation. We then use a greedy heuristic to extract salient sentences while
avoiding redundancy. In our experiments on DUC 2004, we consider three types of
sentence relation graphs and demonstrate the advantage of combining sentence
relations in graphs with the representation power of deep neural networks. Our
model improves upon traditional graph-based extractive approaches and the
vanilla GRU sequence model with no graph, and it achieves competitive results
against other state-of-the-art multi-document summarization systems.Comment: In CoNLL 201
Query-based extracting: how to support the answer?
Human-made query-based summaries commonly contain information not explicitly asked for. They answer the user query, but also provide supporting information. In order to find this information in the source text, a graph is used to model the strength and type of relations between sentences of the query and document cluster, based on various features. The resulting extracts rank second in overall readability in the DUC 2006 evaluation. Employment of better question answering methods is the key to improve also content-based evaluation results
- âŚ