14 research outputs found
Entity Coherence for Descriptive Text Structuring
Institute for Communicating and Collaborative SystemsAlthough entity coherence, i.e. the coherence that arises from certain patterns of references to
entities, is of attested importance for characterising a descriptive text structure, whether and how current formal models of entity coherence such as Centering Theory can be used for the purposes of natural language generation remains unclear. This thesis investigates this issue and sets out to explore which of the many formulations of Centering best suits text structuring. In doing this, we assume text
structuring to be a search task where different orderings of propositions are evaluated according to scores assigned by a metric.
The main question behind this study is how to choose a metric of entity coherence among many
alternatives as the only guidance to the text structuring component of a system that produces descriptions of objects. Different ways of defining metrics of entity coherence using Centering’s notions are discussed and a general corpus-based methodology is introduced to identify which of these metrics constitute the most promising candidates for search-based text structuring before the actual generation
of the descriptive structure takes place.
The performance of a large set of metrics is estimated empirically in a series of computational
experiments using two kinds of data: (i) a reliably annotated corpus representing the genre of interest and (ii) data derived from an existing natural language generation system and ordered according to the instructions of a domain expert.
A final experiment supplements our main methodology by automatically evaluating the best scoring orderings of some of the best performing metrics in comparison to an upper bound defined by orderings produced by multiple experts on additional application-specific data and a lower bound defined by a random baseline.
The main findings are summarised as follows: In general, the simplest metric of entity coherence
constitutes a very robust baseline for both datasets. However, when the metrics are modified
according to an additional constraint on entity coherence, then the baseline is beaten in domain (ii).
The employed modification is supported by the subsidiary evaluation which renders all employed
metrics superior to the random baseline and helps identify the metric which overall constitutes the
most suitable candidate (among the ones investigated) for search-based descriptive text structuring in
domain (ii).
This thesis provides substantial insight into the role of entity coherence as a descriptive text structuring
constraint. Viewing Centering from an NLG perspective raises a series of interesting challenges
that the thesis identifies and attempts to investigate to a certain extent. The general evaluation methodology
and the results of the empirical studies are useful for any subsequent attempt to generate a descriptive
text structure in the context of an application that makes use of the notion of entity coherence
as modelled by Centering
Writing The Literature Review Section: Teaching Undergraduate Psychology Students Scientific Writing
Many undergraduate psychology students write the literature review section of a scientific paper as a list of summaries without direction or coherence. This paper proposes to teach students to write the literature review section as an argument instead of following the traditional hourglass metaphor approach
Evaluating Centering for Information Ordering Using Corpora
In this article we discuss several metrics of coherence defined using centering theory and investigate the usefulness of such metrics for information ordering in automatic text generation. We estimate empirically which is the most promising metric and how useful this metric is using a general methodology applied on several corpora. Our main result is that the simplest metric (which relies exclusively on NOCB transitions) sets a robust baseline that cannot be outperformed by other metrics which make use of additional centering-based features. This baseline can be used for the development of both text-to-text and concept-to-text generation systems. </jats:p
Literature Genre Effects on Memory and Influence
Superstructures are text structure relations commonly found in specific types of text such as narrative texts. Superstructures are important from a pedagogical standpoint because learners’ comprehension can be improved when they are taught about superstructures (Baumann & Bergeron, 1993; Calfee & Patrick, 1995; Dymock, 2005). The current study examined three types of texts with distinct superstructures—narrative, expository, and procedural. Undergraduate student participants (n=200) were randomly assigned to read a text that was written in the form of a narrative, expository, or procedural superstructure text. After reading, participants were asked to recall information from the text and rated their compliance level to the instructions provided in the text. Our results indicated a significant recall difference between narrative and expository superstructure texts. Future directions and implications are described in the discussion section
Anaphora resolution for Arabic machine translation :a case study of nafs
PhD ThesisIn the age of the internet, email, and social media there is an increasing need for processing online information, for example, to support education and business. This has led to the rapid development of natural language processing technologies such as computational linguistics, information retrieval, and data mining. As a branch of computational linguistics, anaphora resolution has attracted much interest. This is reflected in the large number of papers on the topic published in journals such as Computational Linguistics. Mitkov (2002) and Ji et al. (2005) have argued that the overall quality of anaphora resolution systems remains low, despite practical advances in the area, and that major challenges include dealing with real-world knowledge and accurate parsing.
This thesis investigates the following research question: can an algorithm be found for the resolution of the anaphor nafs in Arabic text which is accurate to at least 90%, scales linearly with text size, and requires a minimum of knowledge resources? A resolution algorithm intended to satisfy these criteria is proposed. Testing on a corpus of contemporary Arabic shows that it does indeed satisfy the criteria.Egyptian Government
The design and implementation of a system for the automatic generation of narrative debriefs for AUV Missions
Increased autonomy allows autonomous underwater vehicles to act without direct
support or supervision. This requires increased complexity, however, and a deficit
of trust may form between operators and these complex machines, though previous
research has shown this can be reduced through repeated experience with the system
in question. Regardless of whether a mission is performed with real vehicles or their
simulated counterparts, effective debrief represents the most efficient method for
performing an analysis of the mission.
A novel system is presented to maximise the effectiveness of a debrief by ordering
the mission events using a narrative structure, which has been shown to be the
quickest and most effective way of communicating information and building a situation
model inside a person’s mind. Mission logs are de-constructed and analysed,
then optimisation algorithms used to generate a coherent discourse based on the
events of the missions with any required exposition. This is then combined with
a timed mission playback and additional visual information to form an automated
mission debrief.
This approach was contrasted with two alternative techniques: a simpler chronological
ordering; and a facsimile of the current state of the art. Results show
that participant recall accuracy was higher and the need for redundant delivery of
information was lower when compared to either of the baselines. Also apparent is
a need for debriefs to be adapted to individual users and scenarios. Results are
discussed in full, along with suggestions for future avenues of research
Joint models for concept-to-text generation
Much of the data found on the world wide web is in numeric, tabular, or other nontextual
format (e.g., weather forecast tables, stock market charts, live sensor feeds), and
thus inaccessible to non-experts or laypersons. However, most conventional search engines
and natural language processing tools (e.g., summarisers) can only handle textual
input. As a result, data in non-textual form remains largely inaccessible. Concept-to-
text generation refers to the task of automatically producing textual output from
non-linguistic input, and holds promise for rendering non-linguistic data widely accessible.
Several successful generation systems have been produced in the past twenty
years. They mostly rely on human-crafted rules or expert-driven grammars, implement
a pipeline architecture, and usually operate in a single domain.
In this thesis, we present several novel statistical models that take as input a set
of database records and generate a description of them in natural language text. Our
unique idea is to combine the processes of structuring a document (document planning),
deciding what to say (content selection) and choosing the specific words and
syntactic constructs specifying how to say it (lexicalisation and surface realisation),
in a uniform joint manner. Rather than breaking up the generation process into a sequence
of local decisions, we define a probabilistic context-free grammar that globally
describes the inherent structure of the input (a corpus of database records and
text describing some of them). This joint representation allows individual processes
(i.e., document planning, content selection, and surface realisation) to communicate
and influence each other naturally.
We recast generation as the task of finding the best derivation tree for a set of input
database records and our grammar, and describe several algorithms for decoding in this
framework that allows to intersect the grammar with additional information capturing
fluency and syntactic well-formedness constraints. We implement our generators using
the hypergraph framework. Contrary to traditional systems, we learn all the necessary
document, structural and linguistic knowledge from unannotated data. Additionally,
we explore a discriminative reranking approach on the hypergraph representation of
our model, by including more refined content selection features. Central to our approach
is the idea of porting our models to various domains; we experimented on four
widely different domains, namely sportscasting, weather forecast generation, booking
flights, and troubleshooting guides. The performance of our systems is competitive
and often superior compared to state-of-the-art systems that use domain specific constraints,
explicit feature engineering or labelled data
Reordering metrics for statistical machine translation
Natural languages display a great variety of different word orders, and one of the
major challenges facing statistical machine translation is in modelling these differences.
This thesis is motivated by a survey of 110 different language pairs drawn
from the Europarl project, which shows that word order differences account for more
variation in translation performance than any other factor. This wide ranging analysis
provides compelling evidence for the importance of research into reordering.
There has already been a great deal of research into improving the quality of the
word order in machine translation output. However, there has been very little analysis
of how best to evaluate this research. Current machine translation metrics are largely
focused on evaluating the words used in translations, and their ability to measure the
quality of word order has not been demonstrated. In this thesis we introduce novel
metrics for quantitatively evaluating reordering.
Our approach isolates the word order in translations by using word alignments.
We reduce alignment information to permutations and apply standard distance metrics
to compare the word order in the reference to that of the translation. We show
that our metrics correlate more strongly with human judgements of word order quality
than current machine translation metrics. We also show that a combined lexical and
reordering metric, the LRscore, is useful for training translation model parameters.
Humans prefer the output of models trained using the LRscore as the objective function,
over those trained with the de facto standard translation metric, the BLEU score.
The LRscore thus provides researchers with a reliable metric for evaluating the impact
of their research on the quality of word order
Data-to-text generation with neural planning
In this thesis, we consider the task of data-to-text generation, which takes non-linguistic
structures as input and produces textual output. The inputs can take the form of
database tables, spreadsheets, charts, and so on. The main application of data-to-text
generation is to present information in a textual format which makes it accessible to
a layperson who may otherwise find it problematic to understand numerical figures.
The task can also automate routine document generation jobs, thus improving human
efficiency. We focus on generating long-form text, i.e., documents with multiple paragraphs. Recent approaches to data-to-text generation have adopted the very successful
encoder-decoder architecture or its variants. These models generate fluent (but often
imprecise) text and perform quite poorly at selecting appropriate content and ordering
it coherently. This thesis focuses on overcoming these issues by integrating content
planning with neural models. We hypothesize data-to-text generation will benefit from
explicit planning, which manifests itself in (a) micro planning, (b) latent entity planning, and (c) macro planning. Throughout this thesis, we assume the input to our
generator are tables (with records) in the sports domain. And the output are summaries
describing what happened in the game (e.g., who won/lost, ..., scored, etc.).
We first describe our work on integrating fine-grained or micro plans with data-to-text generation. As part of this, we generate a micro plan highlighting which records
should be mentioned and in which order, and then generate the document while taking
the micro plan into account.
We then show how data-to-text generation can benefit from higher level latent entity planning. Here, we make use of entity-specific representations which are dynam ically updated. The text is generated conditioned on entity representations and the
records corresponding to the entities by using hierarchical attention at each time step.
We then combine planning with the high level organization of entities, events, and
their interactions. Such coarse-grained macro plans are learnt from data and given
as input to the generator. Finally, we present work on making macro plans latent
while incrementally generating a document paragraph by paragraph. We infer latent
plans sequentially with a structured variational model while interleaving the steps of
planning and generation. Text is generated by conditioning on previous variational
decisions and previously generated text.
Overall our results show that planning makes data-to-text generation more interpretable, improves the factuality and coherence of the generated documents and re duces redundancy in the output document