54,404 research outputs found

    Translating semantic web service based business process models

    Get PDF
    We describe a model-driven translation approach between Semantic Web Service based business process models in the context of the SUPER project. In SUPER we provide a set of business process ontologies for enabling access to the business process space inside the organisation at the semantic level. One major task in this context is to handle the translations between the provided ontologies in order to navigate from different views at the business level to the IT view at the execution level. In this paper we present the results of our translation approach, which transforms instances of BPMO to instances of sBPEL

    Cloud service localisation

    Get PDF
    The essence of cloud computing is the provision of software and hardware services to a range of users in dierent locations. The aim of cloud service localisation is to facilitate the internationalisation and localisation of cloud services by allowing their adaption to dierent locales. We address the lingual localisation by providing service-level language translation techniques to adopt services to dierent languages and regulatory localisation by providing standards-based mappings to achieve regulatory compliance with regionally varying laws, standards and regulations. The aim is to support and enforce the explicit modelling of aspects particularly relevant to localisation and runtime support consisting of tools and middleware services to automating the deployment based on models of locales, driven by the two localisation dimensions. We focus here on an ontology-based conceptual information model that integrates locale specication in a coherent way

    The business process modelling ontology

    Get PDF
    In this paper we describe the Business Process Modelling Ontology (BPMO), which is part of an approach to modelling business processes at the semantic level, integrating knowledge about the organisational context, workflow activities and Semantic Web Services. We harness knowledge representation and reasoning techniques so that business process workflows can: be exposed and shared through semantic descriptions; refer to semantically annotated data and services; incorporate heterogeneous data though semantic mappings; and be queried using a reasoner or inference engine. In this paper we describe our approach and evaluate BPMO through a use case

    Transitioning Applications to Semantic Web Services: An Automated Formal Approach

    No full text
    Semantic Web Services have been recognized as a promising technology that exhibits huge commercial potential, and attract significant attention from both industry and the research community. Despite expectations being high, the industrial take-up of Semantic Web Service technologies has been slower than expected. One of the main reasons is that many systems have been developed without considering the potential of the web in integrating services and sharing resources. Without a systematic methodology and proper tool support, the migration from legacy systems to Semantic Web Service-based systems can be a very tedious and expensive process, which carries a definite risk of failure. There is an urgent need to provide strategies which allow the migration of legacy systems to Semantic Web Services platforms, and also tools to support such a strategy. In this paper we propose a methodology for transitioning these applications to Semantic Web Services by taking the advantage of rigorous mathematical methods. Our methodology allows users to migrate their applications to Semantic Web Services platform automatically or semi-automatically

    Ontology-based patterns for the integration of business processes and enterprise application architectures

    Get PDF
    Increasingly, enterprises are using Service-Oriented Architecture (SOA) as an approach to Enterprise Application Integration (EAI). SOA has the potential to bridge the gap between business and technology and to improve the reuse of existing applications and the interoperability with new ones. In addition to service architecture descriptions, architecture abstractions like patterns and styles capture design knowledge and allow the reuse of successfully applied designs, thus improving the quality of software. Knowledge gained from integration projects can be captured to build a repository of semantically enriched, experience-based solutions. Business patterns identify the interaction and structure between users, business processes, and data. Specific integration and composition patterns at a more technical level address enterprise application integration and capture reliable architecture solutions. We use an ontology-based approach to capture architecture and process patterns. Ontology techniques for pattern definition, extension and composition are developed and their applicability in business process-driven application integration is demonstrated

    TimeSeq: sequences with normalized time information for Seq2Seq translation task

    Get PDF
    During last years, researchers have focused on developing Deep Learning models to translate natural language questions to SQL structured queries, what is known as the "text-to-sql" translation task. By 2020, DL models have reached outstanding results in WikiSQL and Spider challenges. However, there are still some challenges they have not addressed. Time normalization and small size of training samples stand out among these challenges. Nowadays, state-of-the-art models for text-to-sql task are not able to deal with common questions like, "what were the average sales for last three Christmas?" or "how much did I sell in every weekend of the last quarter of previous year?". On the other hand, models are trained with relatively small number of samples compared to the number of parameters in the deep learning models and collecting more questions and annotate the dataset is not easy. For these reasons, we propose TimeSeq, a new standard for annotating temporal information as normalized token sequences (time sequences), which can take advantage of DL models that perform seq-to-seq translation for automatically normalizing relative time information contained in common questions about business databases. Currently this standard applies to the context of questions made about the content of a transactional database; but could grow to other domains. We demonstrated with our experiments that deep learning models can learn to summarize the temporal information contained in a natural language question into a time sequence. We developed a process for data augmentation to increase the examples of pairs of questions and time sequences, based in the substitution of paraphrases corresponding to 3 ontologies we integrated: (1) Ontology of Paraphrases for Temporal Expressions, (2) Ontology of Paraphrases for General Business Database Querying, and (3) Ontology of Paraphrases for Specific Industry Domain, and an algorithm to perform the data augmentation, by creating new pairs of questions and time sequences without affecting the fidelity of the temporal information, as result of controlled combinations of paraphrases substituted at original pairs and at the new pairs generated. Our experiments also demonstrate the usefulness of paraphrases substitution for providing data to train models in the translation from questions to time sequences, that can generalize part of the learning to unseen datasets (0.65 accuracy in validation and test sets vs. 0.48 accuracy when training without data augmentation). However, training variability is not enough to represent the whole variability of unseen data, i.e., the models generalize what they can with the variability present in the data they were trained. Although our data augmentation method based in paraphrases substitution still needs to improve, it provides a relevant contribution for developing new data augmentation methods that enable the use of DL models in cases where the number of examples is too few for training

    A Process Modelling Framework Based on Point Interval Temporal Logic with an Application to Modelling Patient Flows

    Get PDF
    This thesis considers an application of a temporal theory to describe and model the patient journey in the hospital accident and emergency (A&E) department. The aim is to introduce a generic but dynamic method applied to any setting, including healthcare. Constructing a consistent process model can be instrumental in streamlining healthcare issues. Current process modelling techniques used in healthcare such as flowcharts, unified modelling language activity diagram (UML AD), and business process modelling notation (BPMN) are intuitive and imprecise. They cannot fully capture the complexities of the types of activities and the full extent of temporal constraints to an extent where one could reason about the flows. Formal approaches such as Petri have also been reviewed to investigate their applicability to the healthcare domain to model processes. Additionally, to schedule patient flows, current modelling standards do not offer any formal mechanism, so healthcare relies on critical path method (CPM) and program evaluation review technique (PERT), that also have limitations, i.e. finish-start barrier. It is imperative to specify the temporal constraints between the start and/or end of a process, e.g., the beginning of a process A precedes the start (or end) of a process B. However, these approaches failed to provide us with a mechanism for handling these temporal situations. If provided, a formal representation can assist in effective knowledge representation and quality enhancement concerning a process. Also, it would help in uncovering complexities of a system and assist in modelling it in a consistent way which is not possible with the existing modelling techniques. The above issues are addressed in this thesis by proposing a framework that would provide a knowledge base to model patient flows for accurate representation based on point interval temporal logic (PITL) that treats point and interval as primitives. These objects would constitute the knowledge base for the formal description of a system. With the aid of the inference mechanism of the temporal theory presented here, exhaustive temporal constraints derived from the proposed axiomatic system’ components serves as a knowledge base. The proposed methodological framework would adopt a model-theoretic approach in which a theory is developed and considered as a model while the corresponding instance is considered as its application. Using this approach would assist in identifying core components of the system and their precise operation representing a real-life domain deemed suitable to the process modelling issues specified in this thesis. Thus, I have evaluated the modelling standards for their most-used terminologies and constructs to identify their key components. It will also assist in the generalisation of the critical terms (of process modelling standards) based on their ontology. A set of generalised terms proposed would serve as an enumeration of the theory and subsume the core modelling elements of the process modelling standards. The catalogue presents a knowledge base for the business and healthcare domains, and its components are formally defined (semantics). Furthermore, a resolution theorem-proof is used to show the structural features of the theory (model) to establish it is sound and complete. After establishing that the theory is sound and complete, the next step is to provide the instantiation of the theory. This is achieved by mapping the core components of the theory to their corresponding instances. Additionally, a formal graphical tool termed as point graph (PG) is used to visualise the cases of the proposed axiomatic system. PG facilitates in modelling, and scheduling patient flows and enables analysing existing models for possible inaccuracies and inconsistencies supported by a reasoning mechanism based on PITL. Following that, a transformation is developed to map the core modelling components of the standards into the extended PG (PG*) based on the semantics presented by the axiomatic system. A real-life case (from the King’s College hospital accident and emergency (A&E) department’s trauma patient pathway) is considered to validate the framework. It is divided into three patient flows to depict the journey of a patient with significant trauma, arriving at A&E, undergoing a procedure and subsequently discharged. Their staff relied upon the UML-AD and BPMN to model the patient flows. An evaluation of their representation is presented to show the shortfalls of the modelling standards to model patient flows. The last step is to model these patient flows using the developed approach, which is supported by enhanced reasoning and scheduling
    corecore