304 research outputs found

    A Survey on Service Composition Middleware in Pervasive Environments

    Get PDF
    The development of pervasive computing has put the light on a challenging problem: how to dynamically compose services in heterogeneous and highly changing environments? We propose a survey that defines the service composition as a sequence of four steps: the translation, the generation, the evaluation, and finally the execution. With this powerful and simple model we describe the major service composition middleware. Then, a classification of these service composition middleware according to pervasive requirements - interoperability, discoverability, adaptability, context awareness, QoS management, security, spontaneous management, and autonomous management - is given. The classification highlights what has been done and what remains to do to develop the service composition in pervasive environments

    Selection of Composable Web Services Driven by User Requirements

    Get PDF
    International audienceBuilding a composite application based on Web services has become a real challenge regarding the large and diverse service space nowadays. Especially when considering the various functional and non-functional capabilities that Web services may afford and users may require. In this paper, we propose an approach for facilitating Web service selection according to user requirements. These requirements specify the needed functionality and expected QoS, as well as the composability between each pair of services. The originality of our approach is embodied in the use of Relational Concept Analysis (RCA), an extension of Formal Concept Analysis (FCA). Using RCA, we classify services by their calculated QoS levels and composability modes. We use a real case study of 901 services to show how to accomplish an efficient selection of services satisfying a specified set of functional and non-functional requirements

    Applying the levels of conceptual interoperability model in support of integratability, interoperability, and composability for system-of-systems engineering

    Get PDF
    The Levels of Conceptual Interoperability Model (LCIM) was developed to cope with the different layers of interoperation of modeling & simulation applications. It introduced technical, syntactic, semantic, pragmatic, dynamic, and conceptual layers of interoperation and showed how they are related to the ideas of integratability, interoperability, and composability. The model was successfully applied in various domains of systems, cybernetics, and informatics

    Applying the Levels of Conceptual Interoperability Model in Support of Integratability, Interoperability, and Composability for System-of-Systems Engineering

    Get PDF
    The Levels of Conceptual Interoperability Model (LCIM) was developed to cope with the different layers of interoperation of modeling & simulation applications. It introduced technical, syntactic, semantic, pragmatic, dynamic, and conceptual layers of interoperation and showed how they are related to the ideas of integratability, interoperability, and composability. The model was successfully applied in various domains of systems, cybernetics, and informatics

    Un environnement de spécification et de découverte pour la réutilisation des composants logiciels dans le développement des logiciels distribués

    Get PDF
    Notre travail vise Ă  Ă©laborer une solution efficace pour la dĂ©couverte et la rĂ©utilisation des composants logiciels dans les environnements de dĂ©veloppement existants et couramment utilisĂ©s. Nous proposons une ontologie pour dĂ©crire et dĂ©couvrir des composants logiciels Ă©lĂ©mentaires. La description couvre Ă  la fois les propriĂ©tĂ©s fonctionnelles et les propriĂ©tĂ©s non fonctionnelles des composants logiciels exprimĂ©es comme des paramĂštres de QoS. Notre processus de recherche est basĂ© sur la fonction qui calcule la distance sĂ©mantique entre la signature d'un composant et la signature d'une requĂȘte donnĂ©e, rĂ©alisant ainsi une comparaison judicieuse. Nous employons Ă©galement la notion de " subsumption " pour comparer l'entrĂ©e-sortie de la requĂȘte et des composants. AprĂšs sĂ©lection des composants adĂ©quats, les propriĂ©tĂ©s non fonctionnelles sont employĂ©es comme un facteur distinctif pour raffiner le rĂ©sultat de publication des composants rĂ©sultats. Nous proposons une approche de dĂ©couverte des composants composite si aucun composant Ă©lĂ©mentaire n'est trouvĂ©, cette approche basĂ©e sur l'ontologie commune. Pour intĂ©grer le composant rĂ©sultat dans le projet en cours de dĂ©veloppement, nous avons dĂ©veloppĂ© l'ontologie d'intĂ©gration et les deux services " input/output convertor " et " output Matching ".Our work aims to develop an effective solution for the discovery and the reuse of software components in existing and commonly used development environments. We propose an ontology for describing and discovering atomic software components. The description covers both the functional and non functional properties which are expressed as QoS parameters. Our search process is based on the function that calculates the semantic distance between the component interface signature and the signature of a given query, thus achieving an appropriate comparison. We also use the notion of "subsumption" to compare the input/output of the query and the components input/output. After selecting the appropriate components, the non-functional properties are used to refine the search result. We propose an approach for discovering composite components if any atomic component is found, this approach based on the shared ontology. To integrate the component results in the project under development, we developed the ontology integration and two services " input/output convertor " and " output Matching "

    Simulation Software as a Service and Service-Oriented Simulation Experiment

    Get PDF
    Simulation software is being increasingly used in various domains for system analysis and/or behavior prediction. Traditionally, researchers and field experts need to have access to the computers that host the simulation software to do simulation experiments. With recent advances in cloud computing and Software as a Service (SaaS), a new paradigm is emerging where simulation software is used as services that are composed with others and dynamically influence each other for service-oriented simulation experiment on the Internet. The new service-oriented paradigm brings new research challenges in composing multiple simulation services in a meaningful and correct way for simulation experiments. To systematically support simulation software as a service (SimSaaS) and service-oriented simulation experiment, we propose a layered framework that includes five layers: an infrastructure layer, a simulation execution engine layer, a simulation service layer, a simulation experiment layer and finally a graphical user interface layer. Within this layered framework, we provide a specification for both simulation experiment and the involved individual simulation services. Such a formal specification is useful in order to support systematic compositions of simulation services as well as automatic deployment of composed services for carrying out simulation experiments. Built on this specification, we identify the issue of mismatch of time granularity and event granularity in composing simulation services at the pragmatic level, and develop four types of granularity handling agents to be associated with the couplings between services. The ultimate goal is to achieve standard and automated approaches for simulation service composition in the emerging service-oriented computing environment. Finally, to achieve more efficient service-oriented simulation, we develop a profile-based partitioning method that exploits a system’s dynamic behavior and uses it as a profile to guide the spatial partitioning for more efficient parallel simulation. We develop the work in this dissertation within the application context of wildfire spread simulation, and demonstrate the effectiveness of our work based on this application

    Tradespace and Affordability – Phase 1

    Get PDF
    One of the key elements of the SERC’s research strategy is transforming the practice of systems engineering – “SE Transformation.” The Grand Challenge goal for SE Transformation is to transform the DoD community’s current systems engineering and management methods, processes, and tools (MPTs) and practices away from sequential, single stovepipe system, hardware-first, outside-in, document-driven, point-solution, acquisition-oriented approaches; and toward concurrent, portfolio and enterprise-oriented, hardware-software-human engineered, balanced outside-in and inside-out, model-driven, set-based, full life cycle approaches.This material is based upon work supported, in whole or in part, by the U.S. Department of Defense through the Office of the Assistant Secretary of Defense for Research and Engineering (ASD(R&E)) under Contract H98230-08- D-0171 (Task Order 0031, RT 046).This material is based upon work supported, in whole or in part, by the U.S. Department of Defense through the Office of the Assistant Secretary of Defense for Research and Engineering (ASD(R&E)) under Contract H98230-08- D-0171 (Task Order 0031, RT 046)

    Orchestration de web services fiables

    Get PDF
    L’Informatique OrientĂ© Services reprĂ©sente un paradigme pour construire des applications distribuĂ©es sur Internet. L’Architecture OrientĂ©e Services(SOA) est un style architectural qui permet le dĂ©veloppement de ces applications Ă  base de services. Au cours de la derniĂšre dĂ©cennie, l’orchestration des services Web est devenue un domaine trĂšs actif dans la recherche scientifique et acadĂ©mique. Bien que de nombreux dĂ©fis liĂ©s Ă  l’orchestration aient Ă©tĂ© abordĂ©s, la fiabilitĂ© de l’orchestration et de sa vĂ©rification restent encore un sujet ouvert, prĂ©requis et important de fait que ces orchestrations affectent aujourd’hui plusieurs activitĂ©s quotidiennes. Cette thĂšse focalise sur le sujet d’orchestration des Services Web Fiables. En particulier, elle contribue avec un ensemble d’approches, de techniques et d’outils pour amĂ©liorer la sĂ©lection et l’orchestration des services web fiables. PremiĂšrement, elle affine les phases du cycle de vie d’orchestration de services web afin d’assurer une vĂ©rification continuĂ©e de fiabilitĂ© lors des phases de conception et d’exĂ©cution. En outre, elle propose une architecture conceptuelle basĂ©e sur un registre de service amĂ©liorĂ©, pour la mise en Ɠuvre d’orchestrations fiables. DeuxiĂšmement, elle prĂ©sente une approche de mesure de similaritĂ© entre les services web. L’approche repose sur la comparaison des interfaces WSDL de services. L’approche sert Ă  identifier les relations de similaritĂ©, de substituabilitĂ© et de composabilitĂ© entre services. L’outil WSSIM a Ă©tĂ© dĂ©veloppĂ© pour mettre en Ɠuvre l’approche proposĂ©e. Pour validation, l’outil a Ă©tĂ© expĂ©rimentĂ© avec un ensemble important de services web rĂ©els. TroisiĂšmement, la thĂšse contribue avec une approche pour l’identification des substituts de services simples et complexes. L’approche utilise les techniques de mesure de similaritĂ©, la classification de service avec FCA et l’analyse de fiabilitĂ© pour identifier et sĂ©lectionner les meilleures substitutes. Un ensemble d’algorithmes aient Ă©tĂ© proposĂ©s pour dĂ©crire le processus d’identification. QuatriĂšmement, pour examiner la rĂ©putation des services comme un autre critĂšre de fiabilitĂ©, la thĂšse introduit un Framework et un modĂšle mathĂ©matique pour la gestion de rĂ©putation de service We

    Digital Twin in the IoT context: a survey on technical features, scenarios and architectural models

    Get PDF
    Digital Twin is an emerging concept that is gaining attention in various industries. It refers to the ability to clone a physical object into a software counterpart. The softwarized object, termed logical object, reflects all the important properties and characteristics of the original object within a specific application context. To fully determine the expected properties of the Digital Twin, this paper surveys the state of the art starting from the original definition within the manufacturing industry. It takes into account related proposals emerging in other fields, namely, Augmented and Virtual Reality (e.g., avatars), Multi-agent systems, and virtualization. This survey thereby allows for the identification of an extensive set of Digital Twin features that point to the “softwarization” of physical objects. To properly consolidate a shared Digital Twin definition, a set of foundational properties is identified and proposed as a common ground outlining the essential characteristics (must-haves) of a Digital Twin. Once the Digital Twin definition has been consolidated, its technical and business value is discussed in terms of applicability and opportunities. Four application scenarios illustrate how the Digital Twin concept can be used and how some industries are applying it. The scenarios also lead to a generic DT architectural Model. This analysis is then complemented by the identification of software architecture models and guidelines in order to present a general functional framework for the Digital Twin. The paper, eventually, analyses a set of possible evolution paths for the Digital Twin considering its possible usage as a major enabler for the softwarization process
    • 

    corecore