1,595 research outputs found

    The DeepThought Core Architecture Framework

    Get PDF
    The research performed in the DeepThought project aims at demonstrating the potential of deep linguistic processing if combined with shallow methods for robustness. Classical information retrieval is extended by high precision concept indexing and relation detection. On the basis of this approach, the feasibility of three ambitious applications will be demonstrated, namely: precise information extraction for business intelligence; email response management for customer relationship management; creativity support for document production and collective brainstorming. Common to these applications, and the basis for their development is the XML-based, RMRS-enabled core architecture framework that will be described in detail in this paper. The framework is not limited to the applications envisaged in the DeepThought project, but can also be employed e.g. to generate and make use of XML standoff annotation of documents and linguistic corpora, and in general for a wide range of NLP-based applications and research purposes

    Hybrid robust deep and shallow semantic processing for creativity support in document production

    Get PDF
    The research performed in the DeepThought project (http://www.project-deepthought.net) aims at demonstrating the potential of deep linguistic processing if added to existing shallow methods that ensure robustness. Classical information retrieval is extended by high precision concept indexing and relation detection. We use this approach to demonstrate the feasibility of three ambitious applications, one of which is a tool for creativity support in document production and collective brainstorming. This application is described in detail in this paper. Common to all three applications, and the basis for their development is a platform for integrated linguistic processing. This platform is based on a generic software architecture that combines multiple NLP components and on robust minimal recursive semantics (RMRS) as a uniform representation language

    Murine tumor models for the in vivo evaluation of natural compounds and their derivatives as new cancer therapeutics

    Get PDF

    An integrated architecture for shallow and deep processing

    Get PDF
    We present an architecture for the integration of shallow and deep NLP components which is aimed at flexible combination of different language technologies for a range of practical current and future applications. In particular, we describe the integration of a high-level HPSG parsing system with different high-performance shallow components, ranging from named entity recognition to chunk parsing and shallow clause recognition. The NLP components enrich a representation of natural language text with layers of new XML meta-information using a single shared data structure, called the text chart. We describe details of the integration methods, and show how information extraction and language checking applications for realworld German text benefit from a deep grammatical analysis

    Presentación del Plan de Digitalización para las bibliotecas del CSIC

    Get PDF
    Sección: La RedA finales del mes de abril ha visto la luz el Plan de Digitalización para la Red de Bibliotecas del CSIC. Este plan tiene como objetivo crear una política unitaria que guíe a las bibliotecas en la aplicación de criterios comunes y estándares de uso para el adecuado desarrollo de los proyectos de digitalización. Se busca con ello sentar las bases para la creación y desarrollo de una colección digital propia del CSIC de proyección internacional y con visos de perdurabilidad. Aunque el plan está concebido pensando principalmente en los fondos patrimoniales y libres de derechos de autor, las pautas técnicas y los criterios de selección ahí expuestos son una buena guía para el desarrollo de cualquier proyecto de digitalización.N

    Enzymes involved in DNA ligation and end-healing in the radioresistant bacterium Deinococcus radiodurans

    Get PDF
    BACKGROUND: Enzymes involved in DNA metabolic events of the highly radioresistant bacterium Deinococcus radiodurans are currently examined to understand the mechanisms that protect and repair the Deinococcus radiodurans genome after extremely high doses of gamma-irradiation. Although several Deinococcus radiodurans DNA repair enzymes have been characterised, no biochemical data is available for DNA ligation and DNA endhealing enzymes of Deinococcus radiodurans so far. DNA ligases are necessary to seal broken DNA backbones during replication, repair and recombination. In addition, ionizing radiation frequently leaves DNA strand-breaks that are not feasible for ligation and thus require end-healing by a 5'-polynucleotide kinase or a 3'-phosphatase. We expect that DNA ligases and end-processing enzymes play an important role in Deinococcus radiodurans DNA strand-break repair. RESULTS: In this report, we describe the cloning and expression of a Deinococcus radiodurans DNA ligase in Escherichia coli. This enzyme efficiently catalyses DNA ligation in the presence of Mn(II) and NAD+ as cofactors and lysine 128 was found to be essential for its activity. We have also analysed a predicted second DNA ligase from Deinococcus radiodurans that is part of a putative DNA repair operon and shows sequence similarity to known ATP-dependent DNA ligases. We show that this enzyme possesses an adenylyltransferase activity using ATP, but is not functional as a DNA ligase by itself. Furthermore, we identified a 5'-polynucleotide kinase similar to human polynucleotide kinase that probably prepares DNA termini for subsequent ligation. CONCLUSION: Deinococcus radiodurans contains a standard bacterial DNA ligase that uses NAD+ as a cofactor. Its enzymatic properties are similar to E. coli DNA ligase except for its preference for Mn(II) as a metal cofactor. The function of a putative second DNA ligase remains unclear, but its adenylyltransferase activity classifies it as a member of the nucleotidyltransferase family. Characterization of another protein from the same operon revealed a 5'-polynucleotide kinase with a possible role in DNA strand-break repair

    Corpora and evaluation tools for multilingual named entity grammar development

    Get PDF
    We present an effort for the development of multilingual named entity grammars in a unification-based finite-state formalism (SProUT). Following an extended version of the MUC7 standard, we have developed Named Entity Recognition grammars for German, Chinese, Japanese, French, Spanish, English, and Czech. The grammars recognize person names, organizations, geographical locations, currency, time and date expressions. Subgrammars and gazetteers are shared as much as possible for the grammars of the different languages. Multilingual corpora from the business domain are used for grammar development and evaluation. The annotation format (named entity and other linguistic information) is described. We present an evaluation tool which provides detailed statistics and diagnostics, allows for partial matching of annotations, and supports user-defined mappings between different annotation and grammar output formats

    Herausforderungen der Digitalisierung für die Zukunft der Arbeitswelt

    Full text link
    Der technologische Wandel erlaubt es zunehmend, Wertschöpfungsprozesse automatisiert und digital gestützt zu erbringen. Diese Digitalisierung und Automatisierung ist im Produktionsbereich beispielsweise gekennzeichnet durch sich zunehmend selbststeuernde Anlagen und Roboter bis hin zu vollautomatisierten und mit anderen vor- und nachgelagerten Geschäftsbereichen vernetzten "Smart Factories" (Industrie 4.0). Im Dienstleistungsbereich sorgen analog intelligente Software und Algorithmen mit Hilfe großer Datenmengen und Internetschnittstellen für die Digitalisierung und Automatisierung von Geschäftsprozessen. Zum Einsatz kommen etwa Analysetools mit Big Data, Cloud-Computing-Systeme oder Online-Plattformen. Vor dem Hintergrund dieser technologischen Entwicklungen – oftmals Technologien der 4. Industriellen Revolution genannt – werden in der öffentlichen Debatte vermehrt Befürchtungen laut, dass in Zukunft viele Arbeitsplätze überflüssig werden könnten. Dieses Szenario einer „technologischen Arbeitslosigkeit“ wird von Studien aus den USA gestützt, nach denen jeder zweite Arbeitsplatz gefährdet ist (Frey und Osborne, 2013). Für Politik und Gesellschaft drängen sich damit eine Reihe von Fragen auf: Führt die Automatisierung und Digitalisierung tatsächlich zum Abbau von Arbeitsplätzen? Welche Arbeitsplätze sind gefährdet? Wie verändern sich die Arbeitsprozesse und -inhalte im Zuge des Wandels? Wie verändern sich Qualifikations- und Kompetenzanforderungen? Besteht Anpassungsbedarf zur Sicherung der Beschäftigungsfähigkeit von Arbeitnehmern/-innen? Dieser Policy Brief fasst zentrale Erkenntnisse und Einschätzungen aus Sicht der Wissenschaft und den am ZEW entstandenen Arbeiten zusammen
    corecore