171 research outputs found

    Conflict resolution in clinical treatments

    Get PDF
    Dissertação de mestrado integrado em Engenharia InformáticaCurrently, in the health area, there is a need for systems that provide support for the decision of health professionals through specific recommendations for each patient based on Clinical Practice Guidelines (CPGs) for automatic interpretation. CPGs are documents that have enormous importance in the daily life of health professionals, playing a key role in reducing variations in medical practice, improving the quality of health care, and reducing health care costs. These documents reflect knowledge about how best to diagnose and treat diseases in the form of a list of clinical recommendations. However, there may be conflicts and interactions in the application of these clinical recommendations, that which in their maximum exponent may impair the patient’s clinical condition. These conflicts are transported to decision support systems, creating the need to develop computational methods to solve these same conflicts. In the case of multimorbid patients, this resolution of conflicts can be very problematic because these patients suffer from several pathologies at the same time, and that the use of a drug for one particular pathology may have a detrimental effect on the application of another drug in another pathology. Therefore, the objective of this dissertation topic is the determination of conflicts and interactions between drugs and the determination of these same alternatives.Atualmente na área da saúde, existe uma necessidade de existirem sistemas que forneçam apoio à decisão dos profissionais de saúde através de recomendações específicas para cada paciente com base em protocolos clínicos para interpretação automática. Os protocolos clínicos são documentos que têm enorme importância no dia-a-dia dos profissionais de saúde, desempenhando um papel fundamental na redução das variações na prática médica, na melhoria da qualidade dos cuidados de saúde e na redução dos custos de saúde. Estes documentos reflectem o conhecimento sobre a melhor forma de diagnosticar e tratar doenças na forma de uma lista de recomendações clínicas. Contudo, podem existir conflitos e interações na aplicação destas recomendações clínicas, que no seu expoente máximo poderão levar a um agravamento do estado clínico do paciente, nomeadamente no caso da aplicação de diferentes fármacos. Estes conflitos são transportados para os sistemas de apoio à decisão, criando a necessidade de desenvolver métodos computacionais de resolução destes mesmos conflitos. No caso dos pacientes multimórbidos esta resolução de conflitos pode ser bastante problemática devido ao facto destes pacientes sofrerem de várias patologias ao mesmo tempo, e que a utilização de um fármaco para uma determinada patologia possa vir a ter um efeito nocivo na aplicação de outro fármaco noutra patologia. Sendo assim, o objetivo deste tema de dissertação é a determinação dos conflitos e interações entre fármacos e a determinação dessas mesmas alternativas

    A abordagem POESIA para a integração de dados e serviços na Web semantica

    Get PDF
    Orientador: Claudia Bauzer MedeirosTese (doutorado) - Universidade Estadual de Campinas, Instituto de ComputaçãoResumo: POESIA (Processes for Open-Ended Systems for lnformation Analysis), a abordagem proposta neste trabalho, visa a construção de processos complexos envolvendo integração e análise de dados de diversas fontes, particularmente em aplicações científicas. A abordagem é centrada em dois tipos de mecanismos da Web semântica: workflows científicos, para especificar e compor serviços Web; e ontologias de domínio, para viabilizar a interoperabilidade e o gerenciamento semânticos dos dados e processos. As principais contribuições desta tese são: (i) um arcabouço teórico para a descrição, localização e composição de dados e serviços na Web, com regras para verificar a consistência semântica de composições desses recursos; (ii) métodos baseados em ontologias de domínio para auxiliar a integração de dados e estimar a proveniência de dados em processos cooperativos na Web; (iii) implementação e validação parcial das propostas, em urna aplicação real no domínio de planejamento agrícola, analisando os benefícios e as limitações de eficiência e escalabilidade da tecnologia atual da Web semântica, face a grandes volumes de dadosAbstract: POESIA (Processes for Open-Ended Systems for Information Analysis), the approach proposed in this work, supports the construction of complex processes that involve the integration and analysis of data from several sources, particularly in scientific applications. This approach is centered in two types of semantic Web mechanisms: scientific workflows, to specify and compose Web services; and domain ontologies, to enable semantic interoperability and management of data and processes. The main contributions of this thesis are: (i) a theoretical framework to describe, discover and compose data and services on the Web, inc1uding mIes to check the semantic consistency of resource compositions; (ii) ontology-based methods to help data integration and estimate data provenance in cooperative processes on the Web; (iii) partial implementation and validation of the proposal, in a real application for the domain of agricultural planning, analyzing the benefits and scalability problems of the current semantic Web technology, when faced with large volumes of dataDoutoradoCiência da ComputaçãoDoutor em Ciência da Computaçã

    HybridMDSD: Multi-Domain Engineering with Model-Driven Software Development using Ontological Foundations

    Get PDF
    Software development is a complex task. Executable applications comprise a mutlitude of diverse components that are developed with various frameworks, libraries, or communication platforms. The technical complexity in development retains resources, hampers efficient problem solving, and thus increases the overall cost of software production. Another significant challenge in market-driven software engineering is the variety of customer needs. It necessitates a maximum of flexibility in software implementations to facilitate the deployment of different products that are based on one single core. To reduce technical complexity, the paradigm of Model-Driven Software Development (MDSD) facilitates the abstract specification of software based on modeling languages. Corresponding models are used to generate actual programming code without the need for creating manually written, error-prone assets. Modeling languages that are tailored towards a particular domain are called domain-specific languages (DSLs). Domain-specific modeling (DSM) approximates technical solutions with intentional problems and fosters the unfolding of specialized expertise. To cope with feature diversity in applications, the Software Product Line Engineering (SPLE) community provides means for the management of variability in software products, such as feature models and appropriate tools for mapping features to implementation assets. Model-driven development, domain-specific modeling, and the dedicated management of variability in SPLE are vital for the success of software enterprises. Yet, these paradigms exist in isolation and need to be integrated in order to exhaust the advantages of every single approach. In this thesis, we propose a way to do so. We introduce the paradigm of Multi-Domain Engineering (MDE) which means model-driven development with multiple domain-specific languages in variability-intensive scenarios. MDE strongly emphasize the advantages of MDSD with multiple DSLs as a neccessity for efficiency in software development and treats the paradigm of SPLE as indispensable means to achieve a maximum degree of reuse and flexibility. We present HybridMDSD as our solution approach to implement the MDE paradigm. The core idea of HybidMDSD is to capture the semantics of particular DSLs based on properly defined semantics for software models contained in a central upper ontology. Then, the resulting semantic foundation can be used to establish references between arbitrary domain-specific models (DSMs) and sophisticated instance level reasoning ensures integrity and allows to handle partiucular change adaptation scenarios. Moreover, we present an approach to automatically generate composition code that integrates generated assets from separate DSLs. All necessary development tasks are arranged in a comprehensive development process. Finally, we validate the introduced approach with a profound prototypical implementation and an industrial-scale case study.Softwareentwicklung ist komplex: ausführbare Anwendungen beinhalten und vereinen eine Vielzahl an Komponenten, die mit unterschiedlichen Frameworks, Bibliotheken oder Kommunikationsplattformen entwickelt werden. Die technische Komplexität in der Entwicklung bindet Ressourcen, verhindert effiziente Problemlösung und führt zu insgesamt hohen Kosten bei der Produktion von Software. Zusätzliche Herausforderungen entstehen durch die Vielfalt und Unterschiedlichkeit an Kundenwünschen, die der Entwicklung ein hohes Maß an Flexibilität in Software-Implementierungen abverlangen und die Auslieferung verschiedener Produkte auf Grundlage einer Basis-Implementierung nötig machen. Zur Reduktion der technischen Komplexität bietet sich das Paradigma der modellgetriebenen Softwareentwicklung (MDSD) an. Software-Spezifikationen in Form abstrakter Modelle werden hier verwendet um Programmcode zu generieren, was die fehleranfällige, manuelle Programmierung ähnlicher Komponenten überflüssig macht. Modellierungssprachen, die auf eine bestimmte Problemdomäne zugeschnitten sind, nennt man domänenspezifische Sprachen (DSLs). Domänenspezifische Modellierung (DSM) vereint technische Lösungen mit intentionalen Problemen und ermöglicht die Entfaltung spezialisierter Expertise. Um der Funktionsvielfalt in Software Herr zu werden, bietet der Forschungszweig der Softwareproduktlinienentwicklung (SPLE) verschiedene Mittel zur Verwaltung von Variabilität in Software-Produkten an. Hierzu zählen Feature-Modelle sowie passende Werkzeuge, um Features auf Implementierungsbestandteile abzubilden. Modellgetriebene Entwicklung, domänenspezifische Modellierung und eine spezielle Handhabung von Variabilität in Softwareproduktlinien sind von entscheidender Bedeutung für den Erfolg von Softwarefirmen. Zur Zeit bestehen diese Paradigmen losgelöst voneinander und müssen integriert werden, damit die Vorteile jedes einzelnen für die Gesamtheit der Softwareentwicklung entfaltet werden können. In dieser Arbeit wird ein Ansatz vorgestellt, der dies ermöglicht. Es wird das Multi-Domain Engineering Paradigma (MDE) eingeführt, welches die modellgetriebene Softwareentwicklung mit mehreren domänenspezifischen Sprachen in variabilitätszentrierten Szenarien beschreibt. MDE stellt die Vorteile modellgetriebener Entwicklung mit mehreren DSLs als eine Notwendigkeit für Effizienz in der Entwicklung heraus und betrachtet das SPLE-Paradigma als unabdingbares Mittel um ein Maximum an Wiederverwendbarkeit und Flexibilität zu erzielen. In der Arbeit wird ein Ansatz zur Implementierung des MDE-Paradigmas, mit dem Namen HybridMDSD, vorgestellt

    Verification of knowledge shared across design and manufacture using a foundation ontology

    Get PDF
    Seamless computer-based knowledge sharing between departments of a manufacturing enterprise is useful in preventing unnecessary design revisions. A lack of interoperability between independently developed knowledge bases, however, is a major impediment in the development of a seamless knowledge sharing system. Interoperability, being an ability to overcome semantic and syntactic differences during computer-based knowledge sharing can be enhanced through the use of ontologies. Ontologies in computer science terms are hierarchical structures of knowledge stored in a computer-based knowledge base. Ontologies have been accepted by all as an interoperable medium to provide a non-subjective way of storing and sharing knowledge across diverse domains. Some semantic and syntactic differences, however, still crop up when these ontological knowledge bases are developed independently. A case study in an aerospace components manufacturing company suggests that shape features of a component are perceived differently by the designing and manufacturing departments. These differences cause further misunderstanding and misinterpretation when computer-based knowledge sharing systems are used across the two domains. Foundation or core ontologies can be used to overcome these differences and to ensure a seamless sharing of knowledge. This is because these ontologies provide a common grounding for domain ontologies to be used by individual domains or department. This common grounding can be used by the mediation and knowledge verification systems to authenticate the meaning of knowledge understood across different domains. For this reason, this research proposes a knowledge verification framework for developing a system capable of verifying knowledge between those domain ontologies which are developed out of a common core or foundation ontology. This framework makes use of ontology logic to standardize the way concepts from a foundation and core-concepts ontology are used in domain ontologies and then by using the same principles the knowledge being shared is verified. The Knowledge Frame Language which is based on Common Logic is used for formalizing example ontologies. The ontology editor used for browsing and querying ontologies is the Integrated Ontology Development Environment (IODE) by Highfleet Inc. An ontological product modelling technique is also developed in this research, to test the proposed framework in the scenario of manufacturability analysis. The proposed framework is then validated through a Java API specially developed for this purpose. Real industrial examples experienced during the case study are used for validation

    Knowledge-based product support systems

    Get PDF
    This research helps bridge the gap between conventional product support, where the support system is considered as a stand-alone application, and the new paradigm of responsive one, where the support system frequently communicates with its environment and reacts to stimuli. This new paradigm would enable product support knowledge to be captured, stored, processed, and updated automatically, being delivered to the users when, where and in the form they need it. The research reported in this thesis first defines Product Support Systems (PRSSs) as electronic means that provide accurate and up-to-date information to the user in a coherent and personalised manner. Product support knowledge is then identified as the integration of product, task, user, and support documentation knowledge. Next, the thesis focuses on an ontology-based model of the structure, relations, and attributes of product support knowledge. In that model product support virtual documentation (PSVD) is presented as an aggregation of Information Objects (IOs) and Information Object Clusters (IOCs). The description of PSVD is followed by an analysis of the relation between IOs, IOCs, and domain knowledge. Then, the thesis builds on the ontology-based representation of product support knowledge and explores the synergy between product support, problem solving, and knowledge engineering. As a result, a structured problem solving approach is introduced that combines case-based adaptation and model-based generation techniques. Based on that approach a knowledge engineering framework for product support systems is developed. A conceptual model of context-aware product support systems that extends the framework is then introduced. The conceptual model includes an ontology-based representation of knowledge related to the users, their activities, the support environment, and the device being used. An approach to semi-automatically integrating design and documentation data is also proposed as part of context-aware product support systems development process.EThOS - Electronic Theses Online ServiceGBUnited Kingdo

    Knowledge extraction from unstructured data and classification through distributed ontologies

    Get PDF
    The World Wide Web has changed the way humans use and share any kind of information. The Web removed several access barriers to the information published and has became an enormous space where users can easily navigate through heterogeneous resources (such as linked documents) and can easily edit, modify, or produce them. Documents implicitly enclose information and relationships among them which become only accessible to human beings. Indeed, the Web of documents evolved towards a space of data silos, linked each other only through untyped references (such as hypertext references) where only humans were able to understand. A growing desire to programmatically access to pieces of data implicitly enclosed in documents has characterized the last efforts of the Web research community. Direct access means structured data, thus enabling computing machinery to easily exploit the linking of different data sources. It has became crucial for the Web community to provide a technology stack for easing data integration at large scale, first structuring the data using standard ontologies and afterwards linking them to external data. Ontologies became the best practices to define axioms and relationships among classes and the Resource Description Framework (RDF) became the basic data model chosen to represent the ontology instances (i.e. an instance is a value of an axiom, class or attribute). Data becomes the new oil, in particular, extracting information from semi-structured textual documents on the Web is key to realize the Linked Data vision. In the literature these problems have been addressed with several proposals and standards, that mainly focus on technologies to access the data and on formats to represent the semantics of the data and their relationships. With the increasing of the volume of interconnected and serialized RDF data, RDF repositories may suffer from data overloading and may become a single point of failure for the overall Linked Data vision. One of the goals of this dissertation is to propose a thorough approach to manage the large scale RDF repositories, and to distribute them in a redundant and reliable peer-to-peer RDF architecture. The architecture consists of a logic to distribute and mine the knowledge and of a set of physical peer nodes organized in a ring topology based on a Distributed Hash Table (DHT). Each node shares the same logic and provides an entry point that enables clients to query the knowledge base using atomic, disjunctive and conjunctive SPARQL queries. The consistency of the results is increased using data redundancy algorithm that replicates each RDF triple in multiple nodes so that, in the case of peer failure, other peers can retrieve the data needed to resolve the queries. Additionally, a distributed load balancing algorithm is used to maintain a uniform distribution of the data among the participating peers by dynamically changing the key space assigned to each node in the DHT. Recently, the process of data structuring has gained more and more attention when applied to the large volume of text information spread on the Web, such as legacy data, news papers, scientific papers or (micro-)blog posts. This process mainly consists in three steps: \emph{i)} the extraction from the text of atomic pieces of information, called named entities; \emph{ii)} the classification of these pieces of information through ontologies; \emph{iii)} the disambigation of them through Uniform Resource Identifiers (URIs) identifying real world objects. As a step towards interconnecting the web to real world objects via named entities, different techniques have been proposed. The second objective of this work is to propose a comparison of these approaches in order to highlight strengths and weaknesses in different scenarios such as scientific and news papers, or user generated contents. We created the Named Entity Recognition and Disambiguation (NERD) web framework, publicly accessible on the Web (through REST API and web User Interface), which unifies several named entity extraction technologies. Moreover, we proposed the NERD ontology, a reference ontology for comparing the results of these technologies. Recently, the NERD ontology has been included in the NIF (Natural language processing Interchange Format) specification, part of the Creating Knowledge out of Interlinked Data (LOD2) project. Summarizing, this dissertation defines a framework for the extraction of knowledge from unstructured data and its classification via distributed ontologies. A detailed study of the Semantic Web and knowledge extraction fields is proposed to define the issues taken under investigation in this work. Then, it proposes an architecture to tackle the single point of failure issue introduced by the RDF repositories spread within the Web. Although the use of ontologies enables a Web where data is structured and comprehensible by computing machinery, human users may take advantage of it especially for the annotation task. Hence, this work describes an annotation tool for web editing, audio and video annotation in a web front end User Interface powered on the top of a distributed ontology. Furthermore, this dissertation details a thorough comparison of the state of the art of named entity technologies. The NERD framework is presented as technology to encompass existing solutions in the named entity extraction field and the NERD ontology is presented as reference ontology in the field. Finally, this work highlights three use cases with the purpose to reduce the amount of data silos spread within the Web: a Linked Data approach to augment the automatic classification task in a Systematic Literature Review, an application to lift educational data stored in Sharable Content Object Reference Model (SCORM) data silos to the Web of data and a scientific conference venue enhancer plug on the top of several data live collectors. Significant research efforts have been devoted to combine the efficiency of a reliable data structure and the importance of data extraction techniques. This dissertation opens different research doors which mainly join two different research communities: the Semantic Web and the Natural Language Processing community. The Web provides a considerable amount of data where NLP techniques may shed the light within it. The use of the URI as a unique identifier may provide one milestone for the materialization of entities lifted from a raw text to real world object

    Enforcing Customization in e-Learning Systems: an ontology and product line-based approach

    Full text link
    In the era of e-Learning, educational materials are considered a crucial point for all the stakeholders. On the one hand, instructors aim at creating learning materials that meet the needs and expectations of learners easily and effec-tively; On the other hand, learners want to acquire knowledge in a way that suits their characteristics and preferences. Consequently, the provision and customization of educational materials to meet the needs of learners is a constant challenge and is currently synonymous with technological devel-opment. Promoting the personalization of learning materials, especially dur-ing their development, will help to produce customized learning materials for specific learners' needs. The main objective of this thesis is to reinforce and strengthen Reuse, Cus-tomization and Ease of Production issues in e-Learning materials during the development process. The thesis deals with the design of a framework based on ontologies and product lines to develop customized Learning Objects (LOs). With this framework, the development of learning materials has the following advantages: (i) large-scale production, (ii) faster development time, (iii) greater (re) use of resources. The proposed framework is the main contribution of this thesis, and is char-acterized by the combination of three models: the Content Model, which addresses important points related to the structure of learning materials, their granularity and levels of aggregation; the Customization Model, which con-siders specific learner characteristics and preferences to customize the learn-ing materials; and the LO Product Line (LOPL) model, which handles the subject of variability and creates matter-them in an easy and flexible way. With these models, instructors can not only develop learning materials, but also reuse and customize them during development. An additional contribution is the Customization Model, which is based on the Learning Style Model (LSM) concept. Based on the study of seven of them, a Global Learning Style Model Ontology (GLSMO) has been con-structed to help instructors with information on the apprentice's characteris-tics and to recommend appropriate LOs for customization. The results of our work have been reflected in the design of an authoring tool for learning materials called LOAT. They have described their require-ments, the elements of their architecture, and some details of their user inter-face. As an example of its use, it includes a case study that shows how its use in the development of some learning components.En la era del e¿Learning, los materiales educativos se consideran un punto crucial para todos los participantes. Por un lado, los instructores tienen como objetivo crear materiales de aprendizaje que satisfagan las necesidades y ex-pectativas de los alumnos de manera fácil y efectiva; por otro lado, los alumnos quieren adquirir conocimientos de una manera que se adapte a sus características y preferencias. En consecuencia, la provisión y personaliza-ción de materiales educativos para satisfacer las necesidades de los estudian-tes es un desafío constante y es actualmente sinónimo de desarrollo tecnoló-gico. El fomento de la personalización de los materiales de aprendizaje, es-pecialmente durante su desarrollo, ayudará a producir materiales de aprendi-zaje específicos para las necesidades específicas de los alumnos. El objetivo fundamental de esta tesis es reforzar y fortalecer los temas de Reutilización, Personalización y Facilidad de Producción en materiales de e-Learning durante el proceso de desarrollo. La tesis se ocupa del diseño de un marco basado en ontologías y líneas de productos para desarrollar objetos de aprendizaje personalizados. Con este marco, el desarrollo de materiales de aprendizaje tiene las siguientes ventajas: (i) producción a gran escala, (ii) tiempo de desarrollo más rápido, (iii) mayor (re)uso de recursos. El marco propuesto es la principal aportación de esta tesis, y se caracteriza por la combinación de tres modelos: el Modelo de Contenido, que aborda puntos importantes relacionados con la estructura de los materiales de aprendizaje, su granularidad y niveles de agregación, el Modelo de Persona-lización, que considera las características y preferencias específicas del alumno para personalizar los materiales de aprendizaje, y el modelo de Línea de productos LO (LOPL), que maneja el tema de la variabilidad y crea ma-teriales de manera fácil y flexible. Con estos modelos, los instructores no sólo pueden desarrollar materiales de aprendizaje, sino también reutilizarlos y personalizarlos durante el desarrollo. Una contribución adicional es el modelo de personalización, que se basa en el concepto de modelo de estilo de aprendizaje. A partir del estudio de siete de ellos, se ha construido una Ontología de Modelo de Estilo de Aprendiza-je Global para ayudar a los instructores con información sobre las caracterís-ticas del aprendiz y recomendarlos apropiados para personalización. Los resultados de nuestro trabajo se han plasmado en el diseño de una he-rramienta de autor de materiales de aprendizaje llamada LOAT. Se han des-crito sus requisitos, los elementos de su arquitectura, y algunos detalles de su interfaz de usuario. Como ejemplo de su uso, se incluye un caso de estudio que muestra cómo su empleo en el desarrollo de algunos componentes de aprendizaje.En l'era de l'e¿Learning, els materials educatius es consideren un punt crucial per a tots els participants. D'una banda, els instructors tenen com a objectiu crear materials d'aprenentatge que satisfacen les necessitats i expectatives dels alumnes de manera fàcil i efectiva; d'altra banda, els alumnes volen ad-quirir coneixements d'una manera que s'adapte a les seues característiques i preferències. En conseqüència, la provisio' i personalitzacio' de materials edu-catius per a satisfer les necessitats dels estudiants és un desafiament constant i és actualment sinònim de desenvolupament tecnològic. El foment de la personalitzacio' dels materials d'aprenentatge, especialment durant el seu desenvolupament, ajudarà a produir materials d'aprenentatge específics per a les necessitats concretes dels alumnes. L'objectiu fonamental d'aquesta tesi és reforçar i enfortir els temes de Reutilització, Personalització i Facilitat de Producció en materials d'e-Learning durant el procés de desenvolupament. La tesi s'ocupa del disseny d'un marc basat en ontologies i línia de productes per a desenvolupar objec-tes d'aprenentatge personalitzats. Amb aquest marc, el desenvolupament de materials d'aprenentatge té els següents avantatges: (i) produccio' a gran esca-la, (ii) temps de desenvolupament mes ràpid, (iii) major (re)ús de recursos. El marc proposat és la principal aportacio' d'aquesta tesi, i es caracteritza per la combinacio' de tres models: el Model de Contingut, que aborda punts im-portants relacionats amb l'estructura dels materials d'aprenentatge, la se-ua granularitat i nivells d'agregació, el Model de Línia de Producte, que ges-tiona el tema de la variabilitat i crea materials d'aprenentatge de manera fàcil i flexible. Amb aquests models, els instructors no solament poden desenvolu-par materials d'aprenentatge, sinó que també poden reutilitzar-los i personalit-zar-los durant el desenvolupament. Una contribucio' addicional és el Model de Personalitzacio', que es basa en el concepte de model d'estil d'aprenentatge. A partir de l'estudi de set d'ells, s'ha construït una Ontologia de Model d'Estil d'Aprenentatge Global per a ajudar als instructors amb informacio' sobre les característiques de l'aprenent i recomanar els apropiats per a personalitzacio'. Els resultats del nostre treball s'han plasmat en el disseny d'una eina d'autor de materials d'aprenentatge anomenada LOAT. S'han descrit els seus requi-sits, els elements de la seua arquitectura, i alguns detalls de la seua interfície d'usuari. Com a exemple del seu ús, s'inclou un cas d'estudi que mostra com és el desenvolupament d'alguns components d'aprenentatge.Ezzat Labib Awad, A. (2017). Enforcing Customization in e-Learning Systems: an ontology and product line-based approach [Tesis doctoral no publicada]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/90515TESI

    AH 2004 : 3rd international conference on adaptive hypermedia and adaptive web-based systems : workshop proceedings part 2

    Get PDF

    Knowledge-based product support systems

    Get PDF
    This research helps bridge the gap between conventional product support, where the support system is considered as a stand-alone application, and the new paradigm of responsive one, where the support system frequently communicates with its environment and reacts to stimuli. This new paradigm would enable product support knowledge to be captured, stored, processed, and updated automatically, being delivered to the users when, where and in the form they need it. The research reported in this thesis first defines Product Support Systems (PRSSs) as electronic means that provide accurate and up-to-date information to the user in a coherent and personalised manner. Product support knowledge is then identified as the integration of product, task, user, and support documentation knowledge. Next, the thesis focuses on an ontology-based model of the structure, relations, and attributes of product support knowledge. In that model product support virtual documentation (PSVD) is presented as an aggregation of Information Objects (IOs) and Information Object Clusters (IOCs). The description of PSVD is followed by an analysis of the relation between IOs, IOCs, and domain knowledge. Then, the thesis builds on the ontology-based representation of product support knowledge and explores the synergy between product support, problem solving, and knowledge engineering. As a result, a structured problem solving approach is introduced that combines case-based adaptation and model-based generation techniques. Based on that approach a knowledge engineering framework for product support systems is developed. A conceptual model of context-aware product support systems that extends the framework is then introduced. The conceptual model includes an ontology-based representation of knowledge related to the users, their activities, the support environment, and the device being used. An approach to semi-automatically integrating design and documentation data is also proposed as part of context-aware product support systems development process
    corecore