1,389 research outputs found

    Tools for enterprises collaboration in virtual enterprises

    Get PDF
    Virtual Enterprise (VE) is an organizational collaboration concept which provides a competitive edge in the globalized business environment. The life cycle of a VE consists of four stages i.e. opportunity identification (Pre-Creation), partner selection (Creation), operation and dissolution. The success of VEs depends upon the efficient execution of their VE-lifecycles along with knowledge enhancement for the partner enterprises to facilitate the future formation of efficient VEs. This research aims to study the different issues which occur in the VE lifecycle and provides a platform for the formation of high performance enterprises and VEs. In the pre-creation stage, enterprises look for suitable partners to create their VE and to exploit a market opportunity. This phase requires explicit and implicit information extraction from enterprise data bases (ECOS-ontology) for the identification of suitable partners. A description logic (DL) based query system is developed to extract explicit and implicit information and to identify potential partners for the creation of the VE. In the creation phase, the identified partners are analysed using different risks paradigms and a cooperative game theoretic approach is used to develop a revenue sharing mechanism based on enterprises inputs and risk minimization for optimal partner selection. In the operation phases, interoperability remains a key issue for seamless transfer of knowledge information and data. DL-based ontology mapping is applied in this research to provide interoperability in the VE between enterprises with different domains of expertise. In the dissolution stage, knowledge acquired in the VE lifecycle needs to be disseminated among the enterprises to enhance their competitiveness. A DL-based ontology merging approach is provided to accommodate new knowledge with existing data bases with logical consistency. Finally, the proposed methodologies are validated using the case study. The results obtained in the case study illustrate the applicability and effectiveness of proposed methodologies in each stage of the VE life cycle

    Talking about Forests: an Example of Sharing Information Expressed with Vague Terms

    Get PDF
    Most natural language terms do not have precise universally agreed definitions that fix their meanings. Even when conversation participants share the same vocabulary and agree on taxonomic relationships (such as subsumption and mutual exclusivity, which might be encoded in an ontology), they may differ greatly in the specific semantics they give to the terms. We illustrate this with the example of `forest', for which the problematic arising of the assignation of different meanings is repeatedly reported in the literature. This is especially the case in the context of an unprecedented scale of publicly available geographic data, where information and databases, even when tagged to ontologies, may present a substantial semantic variation, which challenges interoperability and knowledge exchange. Our research addresses the issue of conceptual vagueness in ontology by providing a framework based on supervaluation semantics that explicitly represents the semantic variability of a concept as a set of admissible precise interpretations. Moreover, we describe the tools that support the conceptual negotiation between an agent and the system, and the specification and reasoning within standpoints

    Compliance flow: an intelligent workflow management system to support engineering processes

    Get PDF
    This work is about extending the scope of current workflow management systems to support engineering processes. On the one hand engineering processes are relatively dynamic, and on the other their specification and performance are constrained by industry standards and guidelines for the sake of product acceptability, such as IEC 61508 for safety and ISO 9001 for quality. A number of technologies have been proposed to increase the adaptability of current workflow systems to deal with dynamic situations. A primary concern is how to support open-ended processes that cannot be completely specified in detail prior to their execution. A survey of adaptive workflow systems is given and the enabling technologies are discussed. Engineering processes are studied and their characteristics are identified and discussed. Current workflow systems have been successfully used in managing "administrative" processes for some time, but they lack the flexibility to support dynamic, unpredictable, collaborative, and highly interdependent engineering processes. [Continues.

    Semantic heterogeneity: comparing new semantic web approaches with those of digital libraries

    Full text link
    To demonstrate that newer developments in the semantic web community, particularly those based on ontologies (simple knowledge organization system and others) mitigate common arguments from the digital library (DL) community against participation in the Semantic web. The approach is a semantic web discussion focusing on the weak structure of the Web and the lack of consideration given to the semantic content during indexing. The points criticised by the semantic web and ontology approaches are the same as those of the DL ‘‘Shell model approach’’ from the mid-1990s, with emphasis on the centrality of its heterogeneity components (used, for example, in vascoda). The Shell model argument began with the ‘‘invisible web’’, necessitating the restructuring of DL approaches. The conclusion is that both approaches fit well together and that the Shell model, with its semantic heterogeneity components, can be reformulated on the semantic web basis. A reinterpretation of the DL approaches of semantic heterogeneity and adapting to standards and tools supported by the W3C should be the best solution. It is therefore recommended that – although most of the semantic web standards are not technologically refined for commercial applications at present – all individual DL developments should be checked for their adaptability to the W3C standards of the semantic web. A unique conceptual analysis of the parallel developments emanating from the digital library and semantic web communities. (author's abstract

    Modeling and improving Spatial Data Infrastructure (SDI)

    Get PDF
    Spatial Data Infrastructure (SDI) development is widely known to be a challenging process owing to its complex and dynamic nature. Although great effort has been made to conceptually explain the complexity and dynamics of SDIs, few studies thus far have actually modeled these complexities. In fact, better modeling of SDI complexities will lead to more reliable plans for its development. A state-of-the-art simulation model of SDI development, hereafter referred to as SMSDI, was created by using the system dynamics (SD) technique. The SMSDI enables policy-makers to test various investment scenarios in different aspects of SDI and helps them to determine the optimum policy for further development of an SDI. This thesis begins with adaption of the SMSDI to a new case study in Tanzania by using the community of participant concept, and further development of the model is performed by using fuzzy logic. It is argued that the techniques and models proposed in this part of the study enable SDI planning to be conducted in a more reliable manner, which facilitates receiving the support of stakeholders for the development of SDI.Developing a collaborative platform such as SDI would highlight the differences among stakeholders including the heterogeneous data they produce and share. This makes the reuse of spatial data difficult mainly because the shared data need to be integrated with other datasets and used in applications that differ from those originally produced for. The integration of authoritative data and Volunteered Geographic Information (VGI), which has a lower level structure and production standards, is a new, challenging area. The second part of this study focuses on proposing techniques to improve the matching and integration of spatial datasets. It is shown that the proposed solutions, which are based on pattern recognition and ontology, can considerably improve the integration of spatial data in SDIs and enable the reuse or multipurpose usage of available data resources

    An Ontology-Based Interpretable Fuzzy Decision Support System for Diabetes Diagnosis

    Get PDF
    Diabetes is a serious chronic disease. The importance of clinical decision support systems (CDSSs) to diagnose diabetes has led to extensive research efforts to improve the accuracy, applicability, interpretability, and interoperability of these systems. However, this problem continues to require optimization. Fuzzy rule-based systems are suitable for the medical domain, where interpretability is a main concern. The medical domain is data-intensive, and using electronic health record data to build the FRBS knowledge base and fuzzy sets is critical. Multiple variables are frequently required to determine a correct and personalized diagnosis, which usually makes it difficult to arrive at accurate and timely decisions. In this paper, we propose and implement a new semantically interpretable FRBS framework for diabetes diagnosis. The framework uses multiple aspects of knowledge-fuzzy inference, ontology reasoning, and a fuzzy analytical hierarchy process (FAHP) to provide a more intuitive and accurate design. First, we build a two-layered hierarchical and interpretable FRBS; then, we improve this by integrating an ontology reasoning process based on SNOMED CT standard ontology. We incorporate FAHP to determine the relative medical importance of each sub-FRBS. The proposed system offers numerous unique and critical improvements regarding the implementation of an accurate, dynamic, semantically intelligent, and interpretable CDSS. The designed system considers the ontology semantic similarity of diabetes complications and symptoms concepts in the fuzzy rules' evaluation process. The framework was tested using a real data set, and the results indicate how the proposed system helps physicians and patients to accurately diagnose diabetes mellitusThis work was supported by National Research Foundation of Korea-Grant funded by the Korean Government (Ministry of Science, ICT and Future Planning)-NRF-2017R1A2B2012337)S

    Towards a core ontology for information integration

    Get PDF
    In this paper, we argue that a core ontology is one of the key building blocks necessary to enable the scalable assimilation of information from diverse sources. A complete and extensible ontology that expresses the basic concepts that are common across a variety of domains and can provide the basis for specialization into domain-specific concepts and vocabularies, is essential for well-defined mappings between domain-specific knowledge representations (i.e., metadata vocabularies) and the subsequent building of a variety of services such as cross-domain searching, browsing, data mining and knowledge extraction. This paper describes the results of a series of three workshops held in 2001 and 2002 which brought together representatives from the cultural heritage and digital library communities with the goal of harmonizing their knowledge perspectives and producing a core ontology. The knowledge perspectives of these two communities were represented by the CIDOC/CRM [31], an ontology for information exchange in the cultural heritage and museum community, and the ABC ontology [33], a model for the exchange and integration of digital library information. This paper describes the mediation process between these two different knowledge biases and the results of this mediation - the harmonization of the ABC and CIDOC/CRM ontologies, which we believe may provide a useful basis for information integration in the wider scope of the involved communities
    • …
    corecore