128 research outputs found

    Combinatorial Auction-based Mechanisms for Composite Web Service Selection

    Get PDF
    Composite service selection presents the opportunity for the rapid development of complex applications using existing web services. It refers to the problem of selecting a set of web services from a large pool of available candidates to logically compose them to achieve value-added composite services. The aim of service selection is to choose the best set of services based on the functional and non-functional (quality related) requirements of a composite service requester. The current service selection approaches mostly assume that web services are offered as single independent entities; there is no possibility for bundling. Moreover, the current research has mainly focused on solving the problem for a single composite service. There is a limited research to date on how the presence of multiple requests for composite services affects the performance of service selection approaches. Addressing these two aspects can significantly enhance the application of composite service selection approaches in the real-world. We develop new approaches for the composite web service selection problem by addressing both the bundling and multiple requests issues. In particular, we propose two mechanisms based on combinatorial auction models, where the provisioning of multiple services are auctioned simultaneously and service providers can bid to offer combinations of web services. We mapped these mechanisms to Integer Linear Programing models and conducted extensive simulations to evaluate them. The results of our experimentation show that bundling can lead to cost reductions compared to when services are offered independently. Moreover, the simultaneous consideration of a set of requests enhances the success rate of the mechanism in allocating services to requests. By considering all composite service requests at the same time, the mechanism achieves more homogenous prices which can be a determining factor for the service requester in choosing the best composite service selection mechanism to deploy

    Automatic quality of service adaptation for composite web services

    Get PDF
    Quality of Services (QoS) management has become an important issue for Web services. Indeed, QoS is becoming a crucial and a distinguishing criterion among functionally equivalent Web services. QoS Management consists of two complementary tasks: monitoring and adaptation. Both are very challenging because of the unpredictable and dynamic nature of Web service composition. We are motivated to solve the QoS problem by taking advantage of some characteristics of composite Web services, such as their similarity to traditional workflows. In this thesis, we propose a broker based architecture that enables dynamic QoS monitoring and adaptation for composite Web services. Our approach consists of dynamically changing the execution paths of composed Web services by instrumenting the BPEL process. A new construct flexPath is introduced for supporting alternate execution paths definition in BPEL. We developed a BPEL compiler allowing automatic instrumentation for BPEL definition files. The BPEL process is deployed using the instrumented definition files in order to interact with the QoS broker during execution. The QoS broker is a key component in our architecture and is responsible of monitoring the QoS and managing the adaptation. We propose a broker that enables runtime monitoring of QoS, prediction of potential QoS violation, and the selection of the best execution path of the process in order to improve QoS when needed. We developed a prototype to evaluate our proposed architecture. A case study is also presented through an example BPEL process and a number of partner Web services. The performance of the QoS adaptation has been analyzed and the results showed that the QoS of the BPEL process has been considerably adapted and improved comparing to the original one. In addition, we analyzed the major factors that affect the performance of our prototype tool

    A self-learning framework for validation of runtime adaptation in service-oriented systems

    Get PDF
    Ensuring that service-oriented systems can adapt quickly and effectively to changes in service quality, business needs and their runtime environment is an increasingly important research problem. However, while considerable research has focused on developing runtime adaptation frameworks for service-oriented systems, there has been little work on assessing how effective the adaptations are. Effective adaptation ensures the system remains relevant in a changing environment. One way to address the problem is through validation. Validation allows us to assess how well a recommended adaptation addresses the concerns for which the system is reconfigured and provides us with insights into the nature of problems for which different adaptations are suited. However, the dynamic nature of runtime adaptation and the changeable contexts in which service-oriented systems operate make it difficult to specify appropriate validation mechanisms in advance. This thesis describes a novel consumer-centred approach that uses machine learning to continuously validate and refine runtime adaptation in service-oriented systems, through model-based clustering and deep learning. To evaluate the efficacy of the approach a medium sized health care case study was devised and implemented. The results obtained show that self-validation significantly improves the dynamic adaptation process by autonomously addressing changing user requirements at runtime. Further work in this area can improve the framework by integrating other learning algorithms as well as testing the framework on a larger case study

    Security policy architecture for web services environment

    Get PDF
    An enhanced observer is model that observes behaviour of a service and then automatically reports any changes in the state of the service to evaluator model. The e-observer observes the state of a service to determine whether it conforms to and obeys its intended behaviour or policy rules. E-observer techniques address most problems, govern and provide a proven solution that is re-usable in a similar context. This leads to an organisation and formalisation policy which is the engine of the e-observer model. Policies are used to refer to specific security rules for particular systems. They are derived from the goals of management that describe the desired behaviour of distributed heterogeneous systems and networks. These policies should be defended by security which has become a coherent and crucial issue. Security aims to protect these policies whenever possible. It is the first line of protection for resources or assets against events such as loss of availability, unauthorised access or modification of data. The techniques devised to protect information from intruders are general purpose in nature and, therefore, cannot directly enforce security that has no universal definition, the high degree of assurance of security properties of systems used in security-critical areas, such as business, education and financial, is usually achieved by verification. In addition, security policies express the protection requirements of a system in a precise and unambiguous form. They describe the requirements and mechanisms for securing the resources and assets between the sharing parties of a business transaction. However, Service-Oriented Computing (SOC) is a new paradigm of computing that considers "services" as fundamental elements for developing applications/solutions. SOC has many advantages that support IT to improve and increase its capabilities. SOC allows flexibility to be integrated into application development. This allows services to be provided in a highly distributed manner by Web services. Many organisations and enterprises have undertaken developments using SOC. Web services (WSs) are examples of SOC. WSs have become more powerful and sophisticated in recent years and are being used successfully for inter-operable solutions across various networks. The main benefit of web services is that they use machine-to-machine interaction. This leads initially to explore the "Quality" aspect of the services. Quality of Service (QoS) describes many techniques that prioritise one type of traffic or programme that operates across a network connection. Hence, QoS has rules to determine which requests have priority and uses these rules in order to specify their priority to real-time communications. In addition, these rules can be sophisticated and expressed as policies that constrain the behaviour of these services. The rules (policies) should be addressed and enforced by the security mechanism. Moreover, in SOC and in particular web services, services are black boxes where behaviour may be completely determined by its interaction with other services under confederation system. Therefore, we propose the design and implementation of the “behaviour of services,” which is constrained by QoS policies. We formulate and implement novel techniques for web service policy-based QoS, which leads to the development of a framework for observing services. These services interact with each other by verifying them in a formal and systematic manner. This framework can be used to specify security policies in a succinct and unambiguous manner; thus, we developed a set of rules that can be applied inductively to verify the set of traces generated by the specification of our model’s policy. These rules could be also used for verifying the functionality of the system. In order to demonstrate the protection features of information system that is able to specify and concisely describe a set of traces generated, we subsequently consider the design and management of Ponder policy language to express QoS and its associated based on criteria, such as, security. An algorithm was composed for analysing the observations that are constrained by policies, and then a prototype system for demonstrating the observation architecture within the education sector. Finally, an enforcement system was used to successfully deploy the prototype’s infrastructure over Web services in order to define an optimisation model that would capture efficiency requirements. Therefore, our assumption is, tracing and observing the communication between services and then takes the decision based on their behaviour and history. Hence, the big issue here is how do we ensure that some given security requirements are satisfied and enforced? The scenario here is under confederation system and based on the following: System’s components are Web-services. These components are black boxes and designed/built by various vendors. Topology is highly changeable. Consequently, the main issues are: • The proposal, design and development of a prototype of observation system that manages security policy and its associated aspects by evaluating the outcome results via the evaluator model. • Taming the design complexity of the observation system by leaving considerable degrees of freedom for their structure and behaviour and by bestowing upon them certain characteristics, and to learn and adapt with respect to dynamically changing environments.Saudi Arabian Cultural Burea

    Service Quality and Profit Control in Utility Computing Service Life Cycles

    Get PDF
    Utility Computing is one of the most discussed business models in the context of Cloud Computing. Service providers are more and more pushed into the role of utilities by their customer's expectations. Subsequently, the demand for predictable service availability and pay-per-use pricing models increases. Furthermore, for providers, a new opportunity to optimise resource usage offers arises, resulting from new virtualisation techniques. In this context, the control of service quality and profit depends on a deep understanding of the representation of the relationship between business and technique. This research analyses the relationship between the business model of Utility Computing and Service-oriented Computing architectures hosted in Cloud environments. The relations are clarified in detail for the entire service life cycle and throughout all architectural layers. Based on the elaborated relations, an approach to a delivery framework is evolved, in order to enable the optimisation of the relation attributes, while the service implementation passes through business planning, development, and operations. Related work from academic literature does not cover the collected requirements on service offers in this context. This finding is revealed by a critical review of approaches in the fields of Cloud Computing, Grid Computing, and Application Clusters. The related work is analysed regarding appropriate provision architectures and quality assurance approaches. The main concepts of the delivery framework are evaluated based on a simulation model. To demonstrate the ability of the framework to model complex pay-per-use service cascades in Cloud environments, several experiments have been conducted. First outcomes proof that the contributions of this research undoubtedly enable the optimisation of service quality and profit in Cloud-based Service-oriented Computing architectures

    Automated Improvement of Software Architecture Models for Performance and Other Quality Attributes

    Get PDF
    • …
    corecore