157 research outputs found

    Orchestration of heterogeneous middleware services and its application to a comand and control platform

    Get PDF
    MSC Dissertation in Computer EngineeringDistributed objects was, until recently, the leading technology in the design and implementation of component-based architectures, such as the ones based on services, better known as Service-Oriented Architectures (SOA). Although established in the market for more than a decade, and therefore mature, these technologies have failed to overcome the porting of the SOA concept to the Web. Web services are a recent technology that has been growing in the last few years. Their acceptance has increased over enterprises and organizations as they seem to overcome the Web and interoperability related problems of the Distributed Objects technology. Web services provide interoperability between systems and that is undoubtedly a strength of this technology since this is a crucial aspect of nowadays business. Moreover, the widespread of services led to the recent introduction of the service composition concept, that although being a technology independent concept,is closely related to Web services and there is no tool support for other technologies. Nonetheless, distributed objects still play an important role in the development of distributed systems, namely due to performance issues that are important when it comes to the internals of a platform. However, the use of service composition in these distributed object-based platforms requires the exposure of their composing services as Web services. The main objective of this masters thesis is improve the state-of-the-art in the support for the composition of services originating from distributed objects-based platforms. Bearing in mind that these kind of platforms are composed by several services, the idea is to present a platform as a set of Web services in order to be able to orchestrate them

    Supporting service discovery, querying and interaction in ubiquitous computing environments.

    Get PDF
    In this paper, we contend that ubiquitous computing environments will be highly heterogeneous, service rich domains. Moreover, future applications will consequently be required to interact with multiple, specialised service location and interaction protocols simultaneously. We argue that existing service discovery techniques do not provide sufficient support to address the challenges of building applications targeted to these emerging environments. This paper makes a number of contributions. Firstly, using a set of short ubiquitous computing scenarios we identify several key limitations of existing service discovery approaches that reduce their ability to support ubiquitous computing applications. Secondly, we present a detailed analysis of requirements for providing effective support in this domain. Thirdly, we provide the design of a simple extensible meta-service discovery architecture that uses database techniques to unify service discovery protocols and addresses several of our key requirements. Lastly, we examine the lessons learnt through the development of a prototype implementation of our architecture

    Open Home Networks: the TEAHA Approach

    Get PDF
    The current trend for home appliances is networking. Although more and more of these appliances are networked, there is not a standard way of interaction, which restrains the development of services for in-home networks. The lack of standardisation is partly due to a legacy of business interests; white goods, audio video equipment, security, and personal digital appliances all have a different background and have different business models. Rather than profound standardisation we propose secure seamless interworking of technologies, applications, and business interests. In this paper we present an architecture which is embedded in legacy technology. Our approach combines known design patterns, augments existing technology, and facilitates so-called business clusters. Further, we discuss a prototype implementation that integrates as an example OSGI, ZIGBEE, and UPNP technology with CECED (white goods) business interests. The work reported in this paper has been executed in an international industrial project: TEAHA

    An empirical, in-depth investigation into service creation in H.323 Version 4 Networks

    Get PDF
    Over the past few years there has been an increasing tendency to carry voice on IP networks as opposed to the PSTN and other switched circuit networks. Initially this trend was favoured due to reduced costs but occurred at the expense of sacrificing the quality of the voice communications. Switched circuit networks have therefore remained the preferred carrier-grade voice communication network, but this is again changing. The advancement in improved quality of service (QoS) of real-time traffic on the IP network is a contributing factor to the anticipated future of the IP network supplying carrier-grade voice communications. Another contributing factor is the possibility of creating a new range of innovative, state-of-the-art telephony and communications services that acquire leverage through the intelligence and flexibility of the IP network. The latter has yet to be fully explored. Various protocols exist that facilitate the transport of voice and other media on IP networks. The most well known and widely supported of these is H.323. This work presents and discusses H.323 version 4 service creation. The work also categorises the various H.323 services and presents the mechanisms provided by H.323 version 4 that have facilitated the development of the three services I have developed, EmailReader, Telgo323 and CANS

    Assessing and Improving Interoperability of Distributed Systems

    Get PDF
    Interoperabilität von verteilten Systemen ist eine Grundlage für die Entwicklung von neuen und innovativen Geschäftslösungen. Sie erlaubt es existierende Dienste, die auf verschiedenen Systemen angeboten werden, so miteinander zu verknüpfen, dass neue oder erweiterte Dienste zur Verfügung gestellt werden können. Außerdem kann durch diese Integration die Zuverlässigkeit von Diensten erhöht werden. Das Erreichen und Bewerten von Interoperabilität stellt jedoch eine finanzielle und zeitliche Herausforderung dar. Zur Sicherstellung und Bewertung von Interoperabilität werden systematische Methoden benötigt. Um systematisch Interoperabilität von Systemen erreichen und bewerten zu können, wurde im Rahmen der vorliegenden Arbeit ein Prozess zur Verbesserung und Beurteilung von Interoperabilität (IAI) entwickelt. Der IAI-Prozess beinhaltet drei Phasen und kann die Interoperabilität von verteilten, homogenen und auch heterogenen Systemen bewerten und verbessern. Die Bewertung erfolgt dabei durch Interoperabilitätstests, die manuell oder automatisiert ausgeführt werden können. Für die Automatisierung von Interoperabilitätstests wird eine neue Methodik vorgestellt, die einen Entwicklungsprozess für automatisierte Interoperabilitätstestsysteme beinhaltet. Die vorgestellte Methodik erleichtert die formale und systematische Bewertung der Interoperabilität von verteilten Systemen. Im Vergleich zur manuellen Prüfung von Interoperabilität gewährleistet die hier vorgestellte Methodik eine höhere Testabdeckung, eine konsistente Testdurchführung und wiederholbare Interoperabilitätstests. Die praktische Anwendbarkeit des IAI-Prozesses und der Methodik für automatisierte Interoperabilitätstests wird durch drei Fallstudien belegt. In der ersten Fallstudie werden Prozess und Methodik für Internet Protocol Multimedia Subsystem (IMS) Netzwerke instanziiert. Die Interoperabilität von IMS-Netzwerken wurde bisher nur manuell getestet. In der zweiten und dritten Fallstudie wird der IAI-Prozess zur Beurteilung und Verbesserung der Interoperabilität von Grid- und Cloud-Systemen angewendet. Die Bewertung und Verbesserung dieser Interoperabilität ist eine Herausforderung, da Grid- und Cloud-Systeme im Gegensatz zu IMS-Netzwerken heterogen sind. Im Rahmen der Fallstudien werden Möglichkeiten für Integrations- und Interoperabilitätslösungen von Grid- und Infrastructure as a Service (IaaS) Cloud-Systemen sowie von Grid- und Platform as a Service (PaaS) Cloud-Systemen aufgezeigt. Die vorgestellten Lösungen sind in der Literatur bisher nicht dokumentiert worden. Sie ermöglichen die komplementäre Nutzung von Grid- und Cloud-Systemen, eine vereinfachte Migration von Grid-Anwendungen in ein Cloud-System sowie eine effiziente Ressourcennutzung. Die Interoperabilitätslösungen werden mit Hilfe des IAI-Prozesses bewertet. Die Durchführung der Tests für Grid-IaaS-Cloud-Systeme erfolgte manuell. Die Interoperabilität von Grid-PaaS-Cloud-Systemen wird mit Hilfe der Methodik für automatisierte Interoperabilitätstests bewertet. Interoperabilitätstests und deren Beurteilung wurden bisher in der Grid- und Cloud-Community nicht diskutiert, obwohl sie eine Basis für die Entwicklung von standardisierten Schnittstellen zum Erreichen von Interoperabilität zwischen Grid- und Cloud-Systemen bieten.Achieving interoperability of distributed systems offers means for the development of new and innovative business solutions. Interoperability allows the combination of existing services provided on different systems, into new or extended services. Such an integration can also increase the reliability of the provided service. However, achieving and assessing interoperability is a technical challenge that requires high effort regarding time and costs. The reasons are manifold and include differing implementations of standards as well as the provision of proprietary interfaces. The implementations need to be engineered to be interoperable. Techniques that assess and improve interoperability systematically are required. For the assurance of reliable interoperation between systems, interoperability needs to be assessed and improved in a systematic manner. To this aim, we present the Interoperability Assessment and Improvement (IAI) process, which describes in three phases how interoperability of distributed homogeneous and heterogeneous systems can be improved and assessed systematically. The interoperability assessment is achieved by means of interoperability testing, which is typically performed manually. For the automation of interoperability test execution, we present a new methodology including a generic development process for a complete and automated interoperability test system. This methodology provides means for a formalized and systematic assessment of systems' interoperability in an automated manner. Compared to manual interoperability testing, the application of our methodology has the following benefits: wider test coverage, consistent test execution, and test repeatability. We evaluate the IAI process and the methodology for automated interoperability testing in three case studies. Within the first case study, we instantiate the IAI process and the methodology for Internet Protocol Multimedia Subsystem (IMS) networks, which were previously assessed for interoperability only in a manual manner. Within the second and third case study, we apply the IAI process to assess and improve the interoperability of grid and cloud computing systems. Their interoperability assessment and improvement is challenging, since cloud and grid systems are, in contrast to IMS networks, heterogeneous. We develop integration and interoperability solutions for grids and Infrastructure as a Service (IaaS) clouds as well as for grids and Platform as a Service (PaaS) clouds. These solutions are unique and foster complementary usage of grids and clouds, simplified migration of grid applications into the cloud, as well as efficient resource utilization. In addition, we assess the interoperability of the grid-cloud interoperability solutions. While the tests for grid-IaaS clouds are performed manually, we applied our methodology for automated interoperability testing for the assessment of interoperability to grid-PaaS cloud interoperability successfully. These interoperability assessments are unique in the grid-cloud community and provide a basis for the development of standardized interfaces improving the interoperability between grids and clouds

    Dynamisch erweiterbares Trading mit heterogenen Softwarekomponenten

    Get PDF
    Universal Component Trading (UComT) ist ein neuer Ansatz zum Trading mit (objektorientierten) Softwarekomponenten. Component Trading steht für die Anwendung des Trading-Ansatzes auf objektorientierte Komponentenmodelle (z.B. Java Beans/EJB, CORBA Components, COM, CLI/.NET). Der Begriff Universal in dem Namen bedeutet, dass die Trading-Architektur in der Lage ist, mit beliebigen Komponentenmodellen sowohl als Dienstanbieter (Exporter) als auch als Dienstnutzer (Importer) zusammenzuarbeiten. Ausgehend von einem Metamodell, das die Gemeinsamkeiten der objektorientierten Softwarekomponentenmodelle abbildet, wird mit der Simple XML-based Component Description Language (SXCDL) eine modellübergreifende Strukturbeschreibungssprache in XML-Form entwickelt, in die beliebige Dienstbeschreibungssprachen dynamisch zur Laufzeit integriert werden können. Eine einzelne Komponente kann zahlreiche Dienstbeschreibungen besitzen, die gleiche oder unterschiedliche Sprachen verwenden und von gleichen oder unterschiedlichen Autoren stammen. Der Trader im UComT-Modell bietet dem Importer ferner eine vielfältige Unterstützung für die Nutzung der Komponenten. Zur Validierung der Forschungsergebnisse wird im Rahmen der Arbeit ein Prototyp für den Trader-Server (in C#) und zwei Prototypen für den Trader-Client (in C# und in Java) entwickelt, die die zentralen Konzepte des UComT-Modells realisiere

    Network and service monitoring in heterogeneous home networks

    Get PDF
    Home networks are becoming dynamic and technologically heterogeneous. They consist of an increasing number of devices which offer several functionalities and can be used for many different services. In the home, these devices are interconnected using a mixture of networking technologies (for example, Ethernet, Wifi, coaxial cable, or power-line). However, interconnecting these devices is often not easy. The increasing heterogeneity has led to significant device- and service-management complexity. In addition, home networks provide a critical "last meters" access to the public telecom and Internet infrastructure and have a dramatic impact on to the end-to-end reliability and performance of services from these networks. This challenges service providers not only to maintain a satisfactory quality of service level in such heterogeneous home networks, but also to remotely monitor and troubleshoot them. The present thesis work contributes research and several solutions in the field of network and service monitoring in home networks, mainly in three areas: (1) providing automatic device- and service-discovery and configuration, (2) remote management, and (3) providing quality of service (QoS). With regard to the first area, current service discovery technology is designed to relieve the increasing human role in network and service administration. However, the relevant Service Discovery Protocols (SDPs) are lacking crucial features namely: (1) they are not platform- and network-independent, and (2) they do not provide sufficient mechanisms for (device) resource reservation. Consequently, devices implementing different SDPs cannot communicate with each other and share their functionalities and resources in a managed way, especially when they use different network technologies. As a solution to the first problem, we propose a new proxy server architecture that enables IP-based devices and services to be discovered on non-IP based network and vice versa. We implemented the proxy architecture using UPnP respectively Bluetooth SDP as IP- and non-IP-based SDPs. The proxy allows Bluetooth devices and UPnP control points to discover, access, and utilize services located on the other network. Validation experiments with the proxy prototype showed that seamless inter-working can be achieved keeping all proxy functionalities on a single device, thus not requiring modification of currently existing UPnP and Bluetooth end devices. Although the proxy itself taxes the end-to-end performance of the service, it is shown to be still acceptable for an end user. For mitigating resource conflicts in SDPs, we propose a generic resource reservation scheme with properties derived from common SDP operation. Performance studies with a prototype showed that this reservation scheme significantly improves the scalability and sustainability of service access in SDPs, at a minor computational cost. With regard to the second area, it is known that the end-to-end quality of Internet services depends crucially on the performance of the home network. Consequently, service providers require the ability to monitor and configure devices in the home network, behind the home gateway (HG). However, they can only put limited requirements to these off-the-shelf devices, as the consumer electronics market is largely outside their span of control. Therefore they have to make intelligent use of the given device control and management protocols. In this work, we propose an architecture for remote discovery and management of devices in a highly heterogeneous home network. A proof-of-concept is developed for the remote management of UPnP devices in the home with a TR-069/UPnP proxy on the HG. Although this architecture is protocol specific, it can be easily adapted to other web-services based protocols. Service providers are also asking for diagnostic tools with which they can remotely troubleshoot the home networks. One of these tools should be able to gather information about the topology of the home network. Although topology discovery protocols already exist, nothing is known yet about their performance. In this work we propose a set of key performance indicators for home network topology discovery architectures, and how they should be measured. We applied them to the Link-Layer Topology Discovery (LLTD) protocol and the Link-Layer Discovery Protocol (LLDP). Our performance measurement results show that these protocols do not fulfill all the requirements as formulated by the service providers. With regard to the third area, current QoS solutions are mostly based on traffic classification. Because they need to be supported by all devices in the network, they are relatively expensive for home networks. Furthermore, they are not interoperable between different networking technologies. Alternative QoS provision techniques have been proposed in the literature. These techniques require end-user services to pragmatically adapt their properties to the actual condition of the network. For this, the condition of the home network in terms of its available bandwidth, delay, jitter, etc., needs to be known in real time. Appropriate tools for determining the available home network resources do not yet exist. In this work we propose a new method to probe the path capacity and available bandwidth between a server and a client in a home network. The main features of this method are: (a) it does not require adaptation of existing end devices, (b) it does not require pre-knowledge of the link-layer network topology, and (c) it is accurate enough to make reliable QoS predictions for the most relevant home applications. To use these predictions for effective service- or content-adaptation or admission control, one should also know how the state of the home network is expected to change immediately after the current state has been probed. However, not much is known about the stochastic properties of traffic in home networks. Based on a relatively small set of traffic observations in several home networks in the Netherlands, we were able to build a preliminary model for home network traffic dynamics

    Investigating call control using MGCP in conjuction with SIP and H.323

    Get PDF
    Telephony used to mean using a telephone to call another telephone on the Public Switched Telephone Network (PSTN), and data networks were used purely to allow computers to communicate. However, with the advent of the Internet, telephony services have been extended to run on data networks. Telephone calls within the IP network are known as Voice over IP. These calls are carried by a number of protocols, with the most popular ones currently being Session Initiation Protocol (SIP) and H.323. Calls can be made from the IP network to the PSTN and vice versa through the use of a gateway. The gateway translates the packets from the IP network to circuits on the PSTN and vice versa to facilitate calls between the two networks. Gateways have evolved and are now split into two entities using the master/slave architecture. The master is an intelligent Media Gateway Controller (MGC) that handles the call control and signalling. The slave is a "dumb" Media Gateway (MG) that handles the translation of the media. The current gateway control protocols in use are Megaco/H.248, MGCP and Skinny. These protocols have proved themselves on the edge of the network. Furthermore, since they communicate with the call signalling VoIP protocols as well as the PSTN, they have to be the lingua franca between the two networks. Within the VoIP network, the numbers of call signalling protocols make it difficult to communicate with each other and to create services. This research investigates the use of Gateway Control Protocols as the lowest common denominator between the call signalling protocols SIP and H.323. More specifically, it uses MGCP to investigate service creation. It also considers the use of MGCP as a protocol translator between SIP and H.323. A service was created using MGCP to allow H.323 endpoints to send Short Message Service (SMS) messages. This service was then extended with minimal effort to SIP endpoints. This service investigated MGCP’s ability to handle call control from the H.323 and SIP endpoints. An MGC was then successfully used to perform as a protocol translator between SIP and H.323
    • …
    corecore