6,505 research outputs found
Cloud service localisation
The essence of cloud computing is the provision of software
and hardware services to a range of users in dierent locations. The aim of cloud service localisation is to facilitate the internationalisation and localisation of cloud services by allowing their adaption to dierent locales.
We address the lingual localisation by providing service-level language translation techniques to adopt services to dierent languages and regulatory localisation by providing standards-based mappings to achieve regulatory compliance with regionally varying laws, standards and regulations. The aim is to support and enforce the explicit modelling of
aspects particularly relevant to localisation and runtime support consisting of tools and middleware services to automating the deployment based on models of locales, driven by the two localisation dimensions.
We focus here on an ontology-based conceptual information model that integrates locale specication in a coherent way
Cloud service discovery and analysis: a unified framework
Over the past few years, cloud computing has been more and more attractive as a new
computing paradigm due to high flexibility for provisioning on-demand computing
resources that are used as services through the Internet. The issues around cloud service
discovery have considered by many researchers in the recent years. However,
in cloud computing, with the highly dynamic, distributed, the lack of standardized
description languages, diverse services offered at different levels and non-transparent
nature of cloud services, this research area has gained a significant attention. Robust
cloud service discovery approaches will assist the promotion and growth of cloud
service customers and providers, but will also provide a meaningful contribution to
the acceptance and development of cloud computing. In this dissertation, we have
proposed an automated cloud service discovery approach of cloud services. We have
also conducted extensive experiments to validate our proposed approach. The results
demonstrate the applicability of our approach and its capability of effectively identifying
and categorizing cloud services on the Internet. Firstly, we develop a novel
approach to build cloud service ontology. Cloud service ontology initially is built
based on the National Institute of Standards and Technology (NIST) cloud computing
standard. Then, we add new concepts to ontology by automatically analyzing real
cloud services based on cloud service ontology Algorithm. We also propose cloud
service categorization that use Term Frequency to weigh cloud service ontology concepts
and calculate cosine similarity to measure the similarity between cloud services.
The cloud service categorization algorithm is able to categorize cloud services to clusters for effective categorization of cloud services. In addition, we use Machine
Learning techniques to identify cloud service in real environment. Our cloud service
identifier is built by utilizing cloud service features extracted from the real cloud service
providers. We determine several features such as similarity function, semantic
ontology, cloud service description and cloud services components, to be used effectively
in identifying cloud service on the Web. Also, we build a unified model to
expose the cloud service’s features to a cloud service search user to ease the process of
searching and comparison among a large amount of cloud services by building cloud
service’s profile. Furthermore, we particularly develop a cloud service discovery Engine
that has capability to crawl the Web automatically and collect cloud services.
The collected datasets include meta-data of nearly 7,500 real-world cloud services
providers and nearly 15,000 services (2.45GB). The experimental results show that
our approach i) is able to effectively build automatic cloud service ontology, ii) is
robust in identifying cloud service in real environment and iii) is more scalable in
providing more details about cloud services.Thesis (Ph.D.) -- University of Adelaide, School of Computer Science, 201
A study of existing Ontologies in the IoT-domain
Several domains have adopted the increasing use of IoT-based devices to
collect sensor data for generating abstractions and perceptions of the real
world. This sensor data is multi-modal and heterogeneous in nature. This
heterogeneity induces interoperability issues while developing cross-domain
applications, thereby restricting the possibility of reusing sensor data to
develop new applications. As a solution to this, semantic approaches have been
proposed in the literature to tackle problems related to interoperability of
sensor data. Several ontologies have been proposed to handle different aspects
of IoT-based sensor data collection, ranging from discovering the IoT sensors
for data collection to applying reasoning on the collected sensor data for
drawing inferences. In this paper, we survey these existing semantic ontologies
to provide an overview of the recent developments in this field. We highlight
the fundamental ontological concepts (e.g., sensor-capabilities and
context-awareness) required for an IoT-based application, and survey the
existing ontologies which include these concepts. Based on our study, we also
identify the shortcomings of currently available ontologies, which serves as a
stepping stone to state the need for a common unified ontology for the IoT
domain.Comment: Submitted to Elsevier JWS SI on Web semantics for the Internet/Web of
Thing
Smart Ontology Framework for Multi-Tenant Cloud Architecture
The exponential growth of data complexity in an era marked by the rapid expansion of the computer environment has led to an increase in the demand for scalable and effective systems. The crucial stage of data management, which acts as a vital conduit for accelerating the processing of enormous amounts of data, is at the centre of this paradigm. Scientific workflows must be coordinated in order to orchestrate the management of large datasets within this complex ecosystem. These workflows differ from generic workflows in that they involve a complex interplay of scheduling, algorithms, data flow, processes, operational protocols, and a focused attention on data-intensive systems. Software as a Service's (SaaS) distinctive feature of multi-tenancy is inextricably related to the growth of the industry. In this complex fabric, the investigation of scientific processes reveals a mutually beneficial relationship with the multi-tenant cloud orchestration environment, revealing a realm that goes beyond simple control and data propagation. It opens a fresh path for system development and makes service delivery's previously hidden facets visible. This study pioneers an exploration into a thorough framework for scientific operations within the context of multi-tenant cloud orchestration. Semantics-based workflows, which leverage semantics to help users manage the complexities of data orchestration, form the basis of this paradigm. In addition, policy-based processes provide another level of intricacy, giving users a flexible way to manoeuvre the complex environment of multi-tenancy, orchestration, and service identification. The study focuses on the fundamentals of orchestrating scientific workflows in a multi-tenant cloud environment, where the creative, scalable, and effective composition results from the harmonious integration of data and semantics under the guidance of rules
- …