18,463 research outputs found

    A Collaborative Approach to Describe the Domain Language through the Language Extended Lexicon

    Get PDF
    Software development is a succession of descriptions in different languages in which every description is based on a previous one. Since the requirements specification is one of the first descriptions, it is important to begin software development with requirements that are as correct and as complete as possible. Although some literature holds the belief that correctness and completeness are two attributes that requirements specifications must satisfy, we know that these attributes are very difficult to meet. However, we have to find ways to diminish the level of incompleteness and deal with the possible conflicts that do arise in the requirements context. Defining the domain language before specifying the requirements is a way of coping with these problems. Nowadays, software systems involve many stakeholders and it is hard to engage all of them to produce a domain language specification. We rely on collaboration to foster the involvement and cooperation of the stakeholders, thus they are able to explore the differences constructively and provide a common understanding of the domain language beyond their own limited views. In this paper, we propose a strategy to capture the domain language in a collaborative way using the Language Extended Lexicon and we show a validation of the proposed strategy.Laboratorio de Investigación y Formación en Informática Avanzad

    Natural language processing

    Get PDF
    Beginning with the basic issues of NLP, this chapter aims to chart the major research activities in this area since the last ARIST Chapter in 1996 (Haas, 1996), including: (i) natural language text processing systems - text summarization, information extraction, information retrieval, etc., including domain-specific applications; (ii) natural language interfaces; (iii) NLP in the context of www and digital libraries ; and (iv) evaluation of NLP systems

    Building a Generation Knowledge Source using Internet-Accessible Newswire

    Full text link
    In this paper, we describe a method for automatic creation of a knowledge source for text generation using information extraction over the Internet. We present a prototype system called PROFILE which uses a client-server architecture to extract noun-phrase descriptions of entities such as people, places, and organizations. The system serves two purposes: as an information extraction tool, it allows users to search for textual descriptions of entities; as a utility to generate functional descriptions (FD), it is used in a functional-unification based generation system. We present an evaluation of the approach and its applications to natural language generation and summarization.Comment: 8 pages, uses eps

    Web 2.0, language resources and standards to automatically build a multilingual named entity lexicon

    Get PDF
    This paper proposes to advance in the current state-of-the-art of automatic Language Resource (LR) building by taking into consideration three elements: (i) the knowledge available in existing LRs, (ii) the vast amount of information available from the collaborative paradigm that has emerged from the Web 2.0 and (iii) the use of standards to improve interoperability. We present a case study in which a set of LRs for different languages (WordNet for English and Spanish and Parole-Simple-Clips for Italian) are extended with Named Entities (NE) by exploiting Wikipedia and the aforementioned LRs. The practical result is a multilingual NE lexicon connected to these LRs and to two ontologies: SUMO and SIMPLE. Furthermore, the paper addresses an important problem which affects the Computational Linguistics area in the present, interoperability, by making use of the ISO LMF standard to encode this lexicon. The different steps of the procedure (mapping, disambiguation, extraction, NE identification and postprocessing) are comprehensively explained and evaluated. The resulting resource contains 974,567, 137,583 and 125,806 NEs for English, Spanish and Italian respectively. Finally, in order to check the usefulness of the constructed resource, we apply it into a state-of-the-art Question Answering system and evaluate its impact; the NE lexicon improves the system’s accuracy by 28.1%. Compared to previous approaches to build NE repositories, the current proposal represents a step forward in terms of automation, language independence, amount of NEs acquired and richness of the information represented

    Neurocognitive Informatics Manifesto.

    Get PDF
    Informatics studies all aspects of the structure of natural and artificial information systems. Theoretical and abstract approaches to information have made great advances, but human information processing is still unmatched in many areas, including information management, representation and understanding. Neurocognitive informatics is a new, emerging field that should help to improve the matching of artificial and natural systems, and inspire better computational algorithms to solve problems that are still beyond the reach of machines. In this position paper examples of neurocognitive inspirations and promising directions in this area are given

    Specification Cases: a lightweight approach based on Natural Language

    Get PDF
    Use Cases are one of the most used artifacts in software requirements specifications. Although there are a lot of templates that suggests how to describe Use Cases, as well as many quality inspection techniques, there are no many techniques to deal with the complexity and the effort to produce good quality Use Cases. On top of that, Use Cases are biased towards user interaction, leaving out important domain information. Thus, it is necessary to provide techniques to incrementally describe specifications that goes beyond user interaction but which can be derived from simpler artifacts in order to make the process organized and effective. This paper proposes an approach to begin with very simple sentences (kernel sentences), following with the LEL glossary and Scenarios, in order to describe Specification Cases. The approach relies on already defined kernel sentences and suggests three procedures to reorganize the knowledge captured in kernel sentences to describe the LEL, then Scenarios and finally Use Cases. This paper also reports a preliminary evaluation that supports the applicability and usability of the approach
    corecore