254 research outputs found
Generating readable texts for readers with low basic skills
Most NLG systems generate texts for readers with good reading ability, but SkillSum adapts its output for readers with poor literacy. Evaluation with lowskilled readers confirms that SkillSum's knowledge-based microplanning choices enhance readability. We also discuss future readability improvements
Recommended from our members
Floating constraints in lexical choice
Lexical choice is a computationally complex task, requiring a generation system to consider a potentially large number of mappings between concepts and words. Constraints that aid in determining which word is best come from a wide variety of sources, including syntax, semantics, pragmatics, the lexicon, and the underlying domain. Furthermore, in some situations, different constraints come into play early on, while in others, they apply much later. This makes it difficult to determine a systematic ordering in which to apply constraints. In this paper, we present a general approach to lexical choice that can handle multiple, interacting constraints. We focus on the problem of floating constraints, semantic or pragmatic constraints that float, appearing at a variety of different syntactic ranks, often merged with other semantic constraints. This means that multiple content units can be realized by a single surface element, and conversely, that a single content unit can be realized by a variety of surface elements. Our approach uses the Functional Unification Formalism (FUF) to represent a generation lexicon, allowing for declarative and compositional representation of individual constraints
A Knowledge Multidimensional Representation Model for Automatic Text Analysis and Generation: Applications for Cultural Heritage
Knowledge is information that has been contextualized in a certain domain, where it can be used and applied. Natural Language provides a most direct way to transfer knowledge at different levels of conceptual density. The opportunity provided by the evolution of the technologies of Natural Language Processing is thus of making more fluid and universal the process of knowledge transfer. Indeed, unfolding domain knowledge is one way to bring to larger audiences contents that would be otherwise restricted to specialists. This has been done so far in a totally manual way through the skills of divulgators and popular science writers. Technology provides now a way to make this transfer both less expensive and more widespread. Extracting knowledge and then generating from it suitably communicable text in natural language are the two related subtasks that need be fulfilled in order to attain the general goal. To this aim, two fields from information technology have achieved the needed maturity and can therefore be effectively combined. In fact, on the one hand Information Extraction and Retrieval (IER) can extract knowledge from texts and map it into a neutral, abstract form, hence liberating it from the stylistic constraints into which it was originated. From there, Natural Language Generation can take charge, by regenerating automatically, or semi-automatically, the extracted knowledge into texts targeting new communities.
This doctoral thesis provides a contribution to making substantial this combination through the definition and implementation of a novel multidimensional model for the representation of conceptual knowledge and of a workflow that can produce strongly customized textual descriptions.
By exploiting techniques for the generation of paraphrases and by profiling target users, applications and domains, a target-driven approach is proposed to automatically generate multiple texts from the same information core. An extended case study is described to demonstrate the effectiveness of the proposed model and approach in the Cultural Heritage application domain, so as to compare and position this contribution within the current state of the art and to outline future directions
Recommended from our members
Planning multisentential English text using communicative acts
The goal of this research is to develop explanation presentation mechanisms for knowledge based
systems which enable them to define domain terminology and concepts, narrate events, elucidate plans,
processes, or propositions and argue to support a claim or advocate action. This requires the development
of devices which select, structure, order and then linguistically realize explanation content as coherent and
cohesive English text.
With the goal of identifying generic explanation presentation strategies, a wide range of naturally
occurring texts were analyzed with respect to their communicative sttucture, function, content and intended
effects on the reader. This motivated an integrated theory of communicative acts which characterizes text at
the level of rhetorical acts (e.g., describe, define, narrate), illocutionary acts (e.g., inform, request), and
locutionary acts (e.g., ask, command). Taken as a whole, the identified communicative acts characterize
the structure, content and intended effects of four types of text: description, narration, exposition,
argument. These text types have distinct effects such as getting the reader to know about entities, to know
about events, to understand plans, processes, or propositions, or to believe propositions or want to
perform actions. In addition to identifying the communicative function and effect of text at multiple levels
of abstraction, this dissertation details a tripartite theory of focus of attention (discourse focus, temporal
focus, and spatial focus) which constrains the planning and linguistic realization of text.
To test the integrated theory of communicative acts and tripartite theory of focus of attention, a text
generation system TEXPLAN (Textual EXplanation PLANner) was implemented that plans and
linguistically realizes multisentential and multiparagraph explanations from knowledge based systems. The
communicative acts identified during text analysis were formalized as over sixty compositional and (in
some cases) recursive plan operators in the library of a hierarchical planner. Discourse, temporal, and
spatial focus models were implemented to track and use attentional information to guide the organization
and realization of text. Because the plan operators distinguish between the communicative function (e.g.,
argue for a proposition) and the expected effect (e.g., the reader believes the proposition) of communicative
acts, the system is able to construct a discourse model of the structure and function of its textual responses
as well as a user model of the expected effects of its responses on the reader's knowledge, beliefs, and
desires. The system uses both the discourse model and user model to guide subsequent utterances. To test
its generality, the system was interfaced to a variety of domain applications including a neuropsychological
diagnosis system, a mission planning system, and a knowledge based mission simulator. The system
produces descriptions, narrations, expositions, and arguments from these applications, thus exhibiting a
broader range of rhetorical coverage than previous text generation systems
CHORUS Deliverable 2.1: State of the Art on Multimedia Search Engines
Based on the information provided by European projects and national initiatives related to multimedia search as well as domains experts that participated in the CHORUS Think-thanks and workshops, this document reports on the state of the art related to multimedia content search from, a technical, and socio-economic perspective.
The technical perspective includes an up to date view on content based indexing and retrieval technologies, multimedia search in the context of mobile devices and peer-to-peer networks, and an overview of current evaluation and benchmark inititiatives to measure the performance of multimedia search engines.
From a socio-economic perspective we inventorize the impact and legal consequences of these technical advances and point out future directions of research
Adaptive hypertext and hypermedia : workshop : proceedings, 3rd, Sonthofen, Germany, July 14, 2001 and Aarhus, Denmark, August 15, 2001
This paper presents two empirical usability studies based on techniques from Human-Computer Interaction (HeI) and software engineering, which were used to elicit requirements for the design of a hypertext generation system. Here we will discuss the findings of these studies, which were used to motivate the choice of adaptivity techniques. The results showed dependencies between different ways to adapt the explanation content and the document length and formatting. Therefore, the system's architecture had to be modified to cope with this requirement. In addition, the system had to be made adaptable, in addition to being adaptive, in order to satisfy the elicited users' preferences
Adaptive hypertext and hypermedia : workshop : proceedings, 3rd, Sonthofen, Germany, July 14, 2001 and Aarhus, Denmark, August 15, 2001
This paper presents two empirical usability studies based on techniques from Human-Computer Interaction (HeI) and software engineering, which were used to elicit requirements for the design of a hypertext generation system. Here we will discuss the findings of these studies, which were used to motivate the choice of adaptivity techniques. The results showed dependencies between different ways to adapt the explanation content and the document length and formatting. Therefore, the system's architecture had to be modified to cope with this requirement. In addition, the system had to be made adaptable, in addition to being adaptive, in order to satisfy the elicited users' preferences
- …