305 research outputs found

    From Text to Knowledge with Graphs: modelling, querying and exploiting textual content

    Full text link
    This paper highlights the challenges, current trends, and open issues related to the representation, querying and analytics of content extracted from texts. The internet contains vast text-based information on various subjects, including commercial documents, medical records, scientific experiments, engineering tests, and events that impact urban and natural environments. Extracting knowledge from this text involves understanding the nuances of natural language and accurately representing the content without losing information. This allows knowledge to be accessed, inferred, or discovered. To achieve this, combining results from various fields, such as linguistics, natural language processing, knowledge representation, data storage, querying, and analytics, is necessary. The vision in this paper is that graphs can be a well-suited text content representation once annotated and the right querying and analytics techniques are applied. This paper discusses this hypothesis from the perspective of linguistics, natural language processing, graph models and databases and artificial intelligence provided by the panellists of the DOING session in the MADICS Symposium 2022

    Deployment and Operation of Complex Software in Heterogeneous Execution Environments

    Get PDF
    This open access book provides an overview of the work developed within the SODALITE project, which aims at facilitating the deployment and operation of distributed software on top of heterogeneous infrastructures, including cloud, HPC and edge resources. The experts participating in the project describe how SODALITE works and how it can be exploited by end users. While multiple languages and tools are available in the literature to support DevOps teams in the automation of deployment and operation steps, still these activities require specific know-how and skills that cannot be found in average teams. The SODALITE framework tackles this problem by offering modelling and smart editing features to allow those we call Application Ops Experts to work without knowing low level details about the adopted, potentially heterogeneous, infrastructures. The framework offers also mechanisms to verify the quality of the defined models, generate the corresponding executable infrastructural code, automatically wrap application components within proper execution containers, orchestrate all activities concerned with deployment and operation of all system components, and support on-the-fly self-adaptation and refactoring

    A survey on 3D CAD model quality assurance and testing

    Get PDF
    [EN] A new taxonomy of issues related to CAD model quality is presented, which distinguishes between explicit and procedural models. For each type of model, morphologic, syntactic, and semantic errors are characterized. The taxonomy was validated successfully when used to classify quality testing tools, which are aimed at detecting and repairing data errors that may affect the simplification, interoperability, and reusability of CAD models. The study shows that low semantic level errors that hamper simplification are reasonably covered in explicit representations, although many CAD quality testers are still unaffordable for Small and Medium Enterprises, both in terms of cost and training time. Interoperability has been reasonably solved by standards like STEP AP 203 and AP214, but model reusability is not feasible in explicit representations. Procedural representations are promising, as interactive modeling editors automatically prevent most morphologic errors derived from unsuitable modeling strategies. Interoperability problems between procedural representations are expected to decrease dramatically with STEP AP242. Higher semantic aspects of quality such as assurance of design intent, however, are hardly supported by current CAD quality testers. (C) 2016 Elsevier Ltd. All rights reserved.This work was supported by the Spanish Ministry of Economy and Competitiveness and the European Regional Development Fund, through the ANNOTA project (Ref. TIN2013-46036-C3-1-R).GonzĂĄlez-Lluch, C.; Company, P.; Contero, M.; Camba, J.; Plumed, R. (2017). A survey on 3D CAD model quality assurance and testing. Computer-Aided Design. 83:64-79. https://doi.org/10.1016/j.cad.2016.10.003S64798

    The Scalable Brain Atlas: instant web-based access to public brain atlases and related content

    Get PDF
    The Scalable Brain Atlas (SBA) is a collection of web services that provide unified access to a large collection of brain atlas templates for different species. Its main component is an atlas viewer that displays brain atlas data as a stack of slices in which stereotaxic coordinates and brain regions can be selected. These are subsequently used to launch web queries to resources that require coordinates or region names as input. It supports plugins which run inside the viewer and respond when a new slice, coordinate or region is selected. It contains 20 atlas templates in six species, and plugins to compute coordinate transformations, display anatomical connectivity and fiducial points, and retrieve properties, descriptions, definitions and 3d reconstructions of brain regions. The ambition of SBA is to provide a unified representation of all publicly available brain atlases directly in the web browser, while remaining a responsive and light weight resource that specializes in atlas comparisons, searches, coordinate transformations and interactive displays.Comment: Rolf K\"otter sadly passed away on June 9th, 2010. He co-initiated this project and played a crucial role in the design and quality assurance of the Scalable Brain Atla

    Improving the Quality and Utility of Electronic Health Record Data through Ontologies

    Get PDF
    The translational research community, in general, and the Clinical and Translational Science Awards (CTSA) community, in particular, share the vision of repurposing EHRs for research that will improve the quality of clinical practice. Many members of these communities are also aware that electronic health records (EHRs) suffer limitations of data becoming poorly structured, biased, and unusable out of original context. This creates obstacles to the continuity of care, utility, quality improvement, and translational research. Analogous limitations to sharing objective data in other areas of the natural sciences have been successfully overcome by developing and using common ontologies. This White Paper presents the authors’ rationale for the use of ontologies with computable semantics for the improvement of clinical data quality and EHR usability formulated for researchers with a stake in clinical and translational science and who are advocates for the use of information technology in medicine but at the same time are concerned by current major shortfalls. This White Paper outlines pitfalls, opportunities, and solutions and recommends increased investment in research and development of ontologies with computable semantics for a new generation of EHRs

    Machine Learning Methods for Product Quality Monitoring in Electric Resistance Welding

    Get PDF
    Elektrisches Widerstandsschweißen (Englisch: Electric Resistance Welding, ERW) ist eine Gruppe von vollautomatisierten Fertigungsprozessen, bei denen metallische Werkstoffe durch WĂ€rme verbunden werden, die von elektrischem Strom und Widerstand erzeugt wird. Eine genaue QualitĂ€tsüberwachung von ERW kann oft nur teilweise mit destruktiven Methoden durchgeführt werden. Es besteht ein großes industrielles und wirtschaftliches Potenzial, datengetriebene AnsĂ€tze für die QualitĂ€tsüberwachung in ERW zu entwickeln, um die Wartungskosten zu senken und die QualitĂ€tskontrolle zu verbessern. Datengetriebene AnsĂ€tze wie maschinelles Lernen (ML) haben aufgrund der enormen Menge verfügbarer Daten, die von Technologien der Industrie 4.0 bereitgestellt werden, viel Aufmerksamkeit auf sich gezogen. Datengetriebene AnsĂ€tze ermöglichen eine zerstörungsfreie, umfassende und prĂ€zise QualitĂ€tsüberwachung, wenn eine bestimmte Menge prĂ€ziser Daten verfügbar ist. Dies kann eine umfassende Online-QualitĂ€tsüberwachung ermöglichen, die ansonsten mit herkömmlichen empirischen Methoden Ă€ußerst schwierig ist. Es gibt jedoch noch viele Herausforderungen bei der Adoption solcher AnsĂ€tze in der Fertigungsindustrie. Zu diesen Herausforderungen gehören: effiziente Datensammlung, die dasWissen von erforderlichen Datenmengen und relevanten Sensoren für erfolgreiches maschinelles Lernen verlangt; das anspruchsvolle Verstehen von komplexen Prozessen und facettenreichen Daten; eine geschickte Selektion geeigneter ML-Methoden und die Integration von DomĂ€nenwissen für die prĂ€diktive QualitĂ€tsüberwachung mit inhomogenen Datenstrukturen, usw. Bestehende ML-Lösungen für ERW liefern keine systematische Vorgehensweise für die Methodenauswahl. Jeder Prozess der ML-Entwicklung erfordert ein umfassendes Prozess- und DatenverstĂ€ndnis und ist auf ein bestimmtes Szenario zugeschnitten, das schwer zu verallgemeinern ist. Es existieren semantische Lösungen für das Prozess- und DatenverstĂ€ndnis und Datenmanagement. Diese betrachten die Datenanalyse als eine isolierte Phase. Sie liefern keine Systemlösungen für das Prozess- und DatenverstĂ€ndnis, die Datenaufbereitung und die ML-Verbesserung, die konfigurierbare und verallgemeinerbare Lösungen für maschinelles Lernen ermöglichen. Diese Arbeit versucht, die obengenannten Herausforderungen zu adressieren, indem ein Framework fĂŒr maschinelles Lernen für ERW vorgeschlagen wird, und demonstriert fünf industrielle AnwendungsfĂ€lle, die das Framework anwenden und validieren. Das Framework ĂŒberprĂŒft die Fragen und DatenspezifitĂ€ten, schlĂ€gt eine simulationsunterstützte Datenerfassung vor und erörtert Methoden des maschinellen Lernens, die in zwei Gruppen unterteilt sind: Feature Engineering und Feature Learning. Das Framework basiert auf semantischen Technologien, die eine standardisierte Prozess- und Datenbeschreibung, eine Ontologie-bewusste Datenaufbereitung sowie halbautomatisierte und Nutzer-konfigurierbare ML-Lösungen ermöglichen. Diese Arbeit demonstriert außerdem die Übertragbarkeit des Frameworks auf einen hochprĂ€zisen Laserprozess. Diese Arbeit ist ein Beginn des Wegs zur intelligenten Fertigung von ERW, der mit dem Trend der vierten industriellen Revolution korrespondiert

    Ontology-Based Open-Corpus Personalization for E-Learning

    Get PDF
    Conventional closed-corpus adaptive information systems control limited sets of documents in predefined domains and cannot provide access to the external content. Such restrictions contradict the requirements of today, when most of the information systems are implemented in the open document space of the World Wide Web and are expected to operate on the open-corpus content. In order to provide personalized access to open-corpus documents, an adaptive system should be able to maintain modeling of new documents in terms of domain knowledge automatically and dynamically. This dissertation explores the problem of open-corpus personalization and semantic modeling of open-corpus content in the context of e-Learning. Information on the World Wide Web is not without structure. Many collections of online instructional material (tutorials, electronic books, digital libraries, etc.) have been provided with implicit knowledge models encoded in form of tables of content, indexes, headers of chapters, links between pages, and different styles of text fragments. The main dissertation approach tries to leverage this layer of hidden semantics by extracting and representing it as coarse-grained models of content collections. A central domain ontology is used to maintain overlay modeling of students’ knowledge and serves as a reference point for multiple collections of external instructional material. In order to establish the link between the ontology and the open-corpus content models a special ontology mapping algorithm has been developed. The proposed approach has been applied in the Ontology-based Open-corpus Personalization Service that recommends and adaptively annotates online reading material. The domain of Java programming has been chosen for the proof-of-concept implementation. A controlled experiment has been organized to evaluate the developed adaptive system and the proposed approach overall. The results of the evaluation have demonstrated several significant learning effects of the implemented open-corpus personalization. The analysis of log-based data has also shown that the open-corpus version of the system is capable of providing personalization of similar quality to the close-corpus one. Such results indicate that the proposed approach successfully supports open-corpus personalization for e-Learning. Further research is required to verify if the approach remains effective in other subject domains and with other types of instructional content
    • 

    corecore