76,981 research outputs found

    Graphene: Semantically-Linked Propositions in Open Information Extraction

    Full text link
    We present an Open Information Extraction (IE) approach that uses a two-layered transformation stage consisting of a clausal disembedding layer and a phrasal disembedding layer, together with rhetorical relation identification. In that way, we convert sentences that present a complex linguistic structure into simplified, syntactically sound sentences, from which we can extract propositions that are represented in a two-layered hierarchy in the form of core relational tuples and accompanying contextual information which are semantically linked via rhetorical relations. In a comparative evaluation, we demonstrate that our reference implementation Graphene outperforms state-of-the-art Open IE systems in the construction of correct n-ary predicate-argument structures. Moreover, we show that existing Open IE approaches can benefit from the transformation process of our framework.Comment: 27th International Conference on Computational Linguistics (COLING 2018

    Domain transfer for deep natural language generation from abstract meaning representations

    Get PDF
    Stochastic natural language generation systems that are trained from labelled datasets are often domainspecific in their annotation and in their mapping from semantic input representations to lexical-syntactic outputs. As a result, learnt models fail to generalize across domains, heavily restricting their usability beyond single applications. In this article, we focus on the problem of domain adaptation for natural language generation. We show how linguistic knowledge from a source domain, for which labelled data is available, can be adapted to a target domain by reusing training data across domains. As a key to this, we propose to employ abstract meaning representations as a common semantic representation across domains. We model natural language generation as a long short-term memory recurrent neural network encoderdecoder, in which one recurrent neural network learns a latent representation of a semantic input, and a second recurrent neural network learns to decode it to a sequence of words. We show that the learnt representations can be transferred across domains and can be leveraged effectively to improve training on new unseen domains. Experiments in three different domains and with six datasets demonstrate that the lexical-syntactic constructions learnt in one domain can be transferred to new domains and achieve up to 75-100% of the performance of in-domain training. This is based on objective metrics such as BLEU and semantic error rate and a subjective human rating study. Training a policy from prior knowledge from a different domain is consistently better than pure in-domain training by up to 10%

    The Intelligent Web

    Get PDF
    Many people are working on the Semantic Web with the main objective being to enhance web searches. Our proposal is a new research strategy based on the existence of a discrete set of semantic relations for the creation and exploitation of semantic networks on the web. To do so, we have defined in a previous paper (Álamo, Martínez, Jaén) the Rhetoric-Semantic Relation (RSR) based on the results of the Rhetoric Structure Theory. We formulate a general set of RSR capable of building discourse and making it possible to express any concept, procedure or principle in terms of knowledge nodes and RSRs. These knowledge nodes can then be elaborated in the same way. This network structure in terms of RSR makes the objective of developing automatic answering systems possible as well as any other type of utilities oriented towards the exploitation of semantic structure, such as the automatic production of web pages or automatic e-learning generation

    Multiple Discourse Relations on the Sentential Level in Japanese

    Get PDF
    In the German government (BMBF) funded project Verbmobil, a semantic formalism Language for Underspecified Discourse Representation Structures (LUD) is used which describes several DRSs and allows for underspecification. Dealing with Japanese poses challenging problems. In this paper, a treatment of multiple discourse relation constructions on the sentential level is shown, which are common in Japanese but cause a problem for the formalism,. The problem is to distinguish discourse relations which take the widest scope compared with other scope-taking elements on the one hand and to have them underspecified among each other on the other hand. We also state a semantic constraint on the resolution of multiple discourse relations which seems to prevail over the syntactic c-command constraint.Comment: 6 pages, Postscrip

    The Narrator: NLG for digital storytelling

    Get PDF
    We present the Narrator, an NLG component used for the generation of narratives in a digital storytelling system. We describe how the Narrator works and show some examples of generated stories

    The automatic generation of narratives

    Get PDF
    We present the Narrator, a Natural Language Generation component used in a digital storytelling system. The system takes as input a formal representation of a story plot, in the form of a causal network relating the actions of the characters to their motives and their consequences. Based on this input, the Narrator generates a narrative in Dutch, by carrying out tasks such as constructing a Document Plan, performing aggregation and ellipsis and the generation of appropriate referring expressions. We describe how these tasks are performed and illustrate the process with examples, showing how this results in the generation of coherent and well-formed narrative texts
    • 

    corecore