305 research outputs found
From Text to Knowledge with Graphs: modelling, querying and exploiting textual content
This paper highlights the challenges, current trends, and open issues related
to the representation, querying and analytics of content extracted from texts.
The internet contains vast text-based information on various subjects,
including commercial documents, medical records, scientific experiments,
engineering tests, and events that impact urban and natural environments.
Extracting knowledge from this text involves understanding the nuances of
natural language and accurately representing the content without losing
information. This allows knowledge to be accessed, inferred, or discovered. To
achieve this, combining results from various fields, such as linguistics,
natural language processing, knowledge representation, data storage, querying,
and analytics, is necessary. The vision in this paper is that graphs can be a
well-suited text content representation once annotated and the right querying
and analytics techniques are applied. This paper discusses this hypothesis from
the perspective of linguistics, natural language processing, graph models and
databases and artificial intelligence provided by the panellists of the DOING
session in the MADICS Symposium 2022
Deployment and Operation of Complex Software in Heterogeneous Execution Environments
This open access book provides an overview of the work developed within the SODALITE project, which aims at facilitating the deployment and operation of distributed software on top of heterogeneous infrastructures, including cloud, HPC and edge resources. The experts participating in the project describe how SODALITE works and how it can be exploited by end users. While multiple languages and tools are available in the literature to support DevOps teams in the automation of deployment and operation steps, still these activities require specific know-how and skills that cannot be found in average teams. The SODALITE framework tackles this problem by offering modelling and smart editing features to allow those we call Application Ops Experts to work without knowing low level details about the adopted, potentially heterogeneous, infrastructures. The framework offers also mechanisms to verify the quality of the defined models, generate the corresponding executable infrastructural code, automatically wrap application components within proper execution containers, orchestrate all activities concerned with deployment and operation of all system components, and support on-the-fly self-adaptation and refactoring
A survey on 3D CAD model quality assurance and testing
[EN] A new taxonomy of issues related to CAD model quality is presented, which distinguishes between explicit and procedural models. For each type of model, morphologic, syntactic, and semantic errors are characterized. The taxonomy was validated successfully when used to classify quality testing tools, which are aimed at detecting and repairing data errors that may affect the simplification, interoperability, and reusability of CAD models.
The study shows that low semantic level errors that hamper simplification are reasonably covered in explicit representations, although many CAD quality testers are still unaffordable for Small and Medium Enterprises, both in terms of cost and training time. Interoperability has been reasonably solved by standards like STEP AP 203 and AP214, but model reusability is not feasible in explicit representations.
Procedural representations are promising, as interactive modeling editors automatically prevent most morphologic errors derived from unsuitable modeling strategies. Interoperability problems between procedural representations are expected to decrease dramatically with STEP AP242. Higher semantic aspects of quality such as assurance of design intent, however, are hardly supported by current CAD quality testers. (C) 2016 Elsevier Ltd. All rights reserved.This work was supported by the Spanish Ministry of Economy and Competitiveness and the European Regional Development Fund, through the ANNOTA project (Ref. TIN2013-46036-C3-1-R).GonzĂĄlez-Lluch, C.; Company, P.; Contero, M.; Camba, J.; Plumed, R. (2017). A survey on 3D CAD model quality assurance and testing. Computer-Aided Design. 83:64-79. https://doi.org/10.1016/j.cad.2016.10.003S64798
The Scalable Brain Atlas: instant web-based access to public brain atlases and related content
The Scalable Brain Atlas (SBA) is a collection of web services that provide
unified access to a large collection of brain atlas templates for different
species. Its main component is an atlas viewer that displays brain atlas data
as a stack of slices in which stereotaxic coordinates and brain regions can be
selected. These are subsequently used to launch web queries to resources that
require coordinates or region names as input. It supports plugins which run
inside the viewer and respond when a new slice, coordinate or region is
selected. It contains 20 atlas templates in six species, and plugins to compute
coordinate transformations, display anatomical connectivity and fiducial
points, and retrieve properties, descriptions, definitions and 3d
reconstructions of brain regions. The ambition of SBA is to provide a unified
representation of all publicly available brain atlases directly in the web
browser, while remaining a responsive and light weight resource that
specializes in atlas comparisons, searches, coordinate transformations and
interactive displays.Comment: Rolf K\"otter sadly passed away on June 9th, 2010. He co-initiated
this project and played a crucial role in the design and quality assurance of
the Scalable Brain Atla
Improving the Quality and Utility of Electronic Health Record Data through Ontologies
The translational research community, in general, and the Clinical and Translational Science Awards (CTSA) community, in particular, share the vision of repurposing EHRs for research that will improve the quality of clinical practice. Many members of these communities are also aware that electronic health records (EHRs) suffer limitations of data becoming poorly structured, biased, and unusable out of original context. This creates obstacles to the continuity of care, utility, quality improvement, and translational research. Analogous limitations to sharing objective data in other areas of the natural sciences have been successfully overcome by developing and using common ontologies. This White Paper presents the authorsâ rationale for the use of ontologies with computable semantics for the improvement of clinical data quality and EHR usability formulated for researchers with a stake in clinical and translational science and who are advocates for the use of information technology in medicine but at the same time are concerned by current major shortfalls. This White Paper outlines pitfalls, opportunities, and solutions and recommends increased investment in research and development of ontologies with computable semantics for a new generation of EHRs
Machine Learning Methods for Product Quality Monitoring in Electric Resistance Welding
Elektrisches WiderstandsschweiĂen (Englisch: Electric Resistance Welding, ERW) ist eine Gruppe von vollautomatisierten Fertigungsprozessen, bei denen metallische Werkstoffe durch WĂ€rme verbunden werden, die von elektrischem Strom und Widerstand erzeugt wird. Eine genaue QualitĂ€tsuÌberwachung von ERW kann oft nur teilweise mit destruktiven Methoden durchgefuÌhrt werden. Es besteht ein groĂes industrielles und wirtschaftliches Potenzial, datengetriebene AnsĂ€tze fuÌr die QualitĂ€tsuÌberwachung in ERW zu entwickeln, um die Wartungskosten zu senken und die QualitĂ€tskontrolle zu verbessern. Datengetriebene AnsĂ€tze wie maschinelles Lernen (ML) haben aufgrund der enormen Menge verfuÌgbarer Daten, die von Technologien der Industrie 4.0 bereitgestellt werden, viel Aufmerksamkeit auf sich gezogen. Datengetriebene AnsĂ€tze ermöglichen eine zerstörungsfreie, umfassende und prĂ€zise QualitĂ€tsuÌberwachung, wenn eine bestimmte Menge prĂ€ziser Daten verfuÌgbar ist. Dies kann eine umfassende Online-QualitĂ€tsuÌberwachung ermöglichen, die ansonsten mit herkömmlichen empirischen Methoden Ă€uĂerst schwierig ist.
Es gibt jedoch noch viele Herausforderungen bei der Adoption solcher AnsĂ€tze in der Fertigungsindustrie. Zu diesen Herausforderungen gehören: effiziente Datensammlung, die dasWissen von erforderlichen Datenmengen und relevanten Sensoren fuÌr erfolgreiches maschinelles Lernen verlangt; das anspruchsvolle Verstehen von komplexen Prozessen und facettenreichen Daten; eine geschickte Selektion geeigneter ML-Methoden und die Integration von DomĂ€nenwissen fuÌr die prĂ€diktive QualitĂ€tsuÌberwachung mit inhomogenen Datenstrukturen, usw.
Bestehende ML-Lösungen fuÌr ERW liefern keine systematische Vorgehensweise fuÌr die Methodenauswahl. Jeder Prozess der ML-Entwicklung erfordert ein umfassendes Prozess- und DatenverstĂ€ndnis und ist auf ein bestimmtes Szenario zugeschnitten, das schwer zu verallgemeinern ist. Es existieren semantische Lösungen fuÌr das Prozess- und DatenverstĂ€ndnis und Datenmanagement. Diese betrachten die Datenanalyse als eine isolierte Phase. Sie liefern keine Systemlösungen fuÌr das Prozess- und DatenverstĂ€ndnis, die Datenaufbereitung und die ML-Verbesserung, die konfigurierbare und verallgemeinerbare Lösungen fuÌr maschinelles Lernen ermöglichen.
Diese Arbeit versucht, die obengenannten Herausforderungen zu adressieren, indem ein Framework fĂŒr maschinelles Lernen fuÌr ERW vorgeschlagen wird, und demonstriert fuÌnf industrielle AnwendungsfĂ€lle, die das Framework anwenden und validieren. Das Framework ĂŒberprĂŒft die Fragen und DatenspezifitĂ€ten, schlĂ€gt eine simulationsunterstuÌtzte Datenerfassung vor und erörtert Methoden des maschinellen Lernens, die in zwei Gruppen unterteilt sind: Feature Engineering und Feature Learning. Das Framework basiert auf semantischen Technologien, die eine standardisierte Prozess- und Datenbeschreibung, eine Ontologie-bewusste Datenaufbereitung sowie halbautomatisierte und Nutzer-konfigurierbare ML-Lösungen ermöglichen. Diese Arbeit demonstriert auĂerdem die Ăbertragbarkeit des Frameworks auf einen hochprĂ€zisen Laserprozess.
Diese Arbeit ist ein Beginn des Wegs zur intelligenten Fertigung von ERW, der mit dem Trend der vierten industriellen Revolution korrespondiert
Ontology-Based Open-Corpus Personalization for E-Learning
Conventional closed-corpus adaptive information systems control limited sets of documents in predefined domains and cannot provide access to the external content. Such restrictions contradict the requirements of today, when most of the information systems are implemented in the open document space of the World Wide Web and are expected to operate on the open-corpus content. In order to provide personalized access to open-corpus documents, an adaptive system should be able to maintain modeling of new documents in terms of domain knowledge automatically and dynamically. This dissertation explores the problem of open-corpus personalization and semantic modeling of open-corpus content in the context of e-Learning.
Information on the World Wide Web is not without structure. Many collections of online instructional material (tutorials, electronic books, digital libraries, etc.) have been provided with implicit knowledge models encoded in form of tables of content, indexes, headers of chapters, links between pages, and different styles of text fragments. The main dissertation approach tries to leverage this layer of hidden semantics by extracting and representing it as coarse-grained models of content collections. A central domain ontology is used to maintain overlay modeling of studentsâ knowledge and serves as a reference point for multiple collections of external instructional material. In order to establish the link between the ontology and the open-corpus content models a special ontology mapping algorithm has been developed.
The proposed approach has been applied in the Ontology-based Open-corpus Personalization Service that recommends and adaptively annotates online reading material. The domain of Java programming has been chosen for the proof-of-concept implementation. A controlled experiment has been organized to evaluate the developed adaptive system and the proposed approach overall. The results of the evaluation have demonstrated several significant learning effects of the implemented open-corpus personalization. The analysis of log-based data has also shown that the open-corpus version of the system is capable of providing personalization of similar quality to the close-corpus one. Such results indicate that the proposed approach successfully supports open-corpus personalization for e-Learning. Further research is required to verify if the approach remains effective in other subject domains and with other types of instructional content
- âŠ