76 research outputs found

    Quality of Context Enhancements and Cost Effective Radio over Fiber Network Planning

    Get PDF

    A Control-Theoretic Methodology for Adaptive Structured Parallel Computations

    Get PDF
    Adaptivity for distributed parallel applications is an essential feature whose impor- tance has been assessed in many research fields (e.g. scientific computations, large- scale real-time simulation systems and emergency management applications). Especially for high-performance computing, this feature is of special interest in order to properly and promptly respond to time-varying QoS requirements, to react to uncontrollable environ- mental effects influencing the underlying execution platform and to efficiently deal with highly irregular parallel problems. In this scenario the Structured Parallel Programming paradigm is a cornerstone for expressing adaptive parallel programs: the high-degree of composability of parallelization schemes, their QoS predictability formally expressed by performance models, are basic tools in order to introduce dynamic reconfiguration processes of adaptive applications. These reconfigurations are not only limited to imple- mentation aspects (e.g. parallelism degree modifications), but also parallel versions with different structures can be expressed for the same computation, featuring different levels of performance, memory utilization, energy consumption, and exploitation of the memory hierarchies. Over the last decade several programming models and research frameworks have been developed aimed at the definition of tools and strategies for expressing adaptive parallel applications. Notwithstanding this notable research effort, properties like the optimal- ity of the application execution and the stability of control decisions are not sufficiently studied in the existing work. For this reason this thesis exploits a pioneer research in the context of providing formal theoretical tools founded on Control Theory and Game Theory techniques. Based on these approaches, we introduce a formal model for control- ling distributed parallel applications represented by computational graphs of structured parallelism schemes (also called skeleton-based parallelism). Starting out from the performance predictability of structured parallelism schemes, in this thesis we provide a formalization of the concept of adaptive parallel module per- forming structured parallel computations. The module behavior is described in terms of a Hybrid System abstraction and reconfigurations are driven by a Predictive Control ap- proach. Experimental results show the effectiveness of this work, in terms of execution cost reduction as well as the stability degree of a system reconfiguration: i.e. how long a reconfiguration choice is useful for targeting the required QoS levels. This thesis also faces with the issue of controlling large-scale distributed applications composed of several interacting adaptive components. After a panoramic view of the existing control-theoretic approaches (e.g. based on decentralized, distributed or hierar- chical structures of controllers), we introduce a methodology for the distributed predictive control. For controlling computational graphs, the overall control problem consists in a set of coupled control sub-problems for each application module. The decomposition is- sue has a twofold nature: first of all we need to model the coupling relationships between control sub-problems, furthermore we need to introduce proper notions of negotiation and convergence in the control decisions collectively taken by the parallel modules of the application graph. This thesis provides a formalization through basic concepts of Non-cooperative Games and Cooperative Optimization. In the notable context of the dis- tributed control of performance and resource utilization, we exploit a formal description of the control problem providing results for equilibrium point existence and the compari- son of the control optimality with different adaptation strategies and interaction protocols. Discussions and a first validation of the proposed techniques are exploited through exper- iments performed in a simulation environment

    Enhancing infotainment applications quality of service in vehicular ad hoc networks

    Full text link
    Les réseaux ad hoc de véhicules accueillent une multitude d’applications intéressantes. Parmi celles-ci, les applications d’info-divertissement visent à améliorer l’expérience des passagers. Ces applications ont des exigences rigides en termes de délai de livraison et de débit. De nombreuses approches ont été proposées pour assurer la qualité du service des dites applications. Elles sont réparties en deux couches : réseau et contrôle d’accès. Toutefois, ces méthodes présentent plusieurs lacunes. Cette thèse a trois volets. Le premier aborde la question du routage dans le milieu urbain. A cet égard, un nouveau protocole, appelé SCRP, a été proposé. Il exploite l’information sur la circulation des véhicules en temps réel pour créer des épines dorsales sur les routes et les connecter aux intersections à l’aide des nœuds de pont. Ces derniers collectent des informations concernant la connectivité et le délai, utilisées pour choisir les chemins de routage ayant un délai de bout-en-bout faible. Le deuxième s’attaque au problème d’affectation des canaux de services afin d’augmenter le débit. A cet effet, un nouveau mécanisme, appelé ASSCH, a été conçu. ASSCH collecte des informations sur les canaux en temps réel et les donne à un modèle stochastique afin de prédire leurs états dans l’avenir. Les canaux les moins encombrés sont sélectionnés pour être utilisés. Le dernier volet vise à proposer un modèle analytique pour examiner la performance du mécanisme EDCA de la norme IEEE 802.11p. Ce modèle tient en compte plusieurs facteurs, dont l’opportunité de transmission, non exploitée dans IEEE 802.11p.The fact that vehicular ad hoc network accommodates two types of communications, Vehicle-to-Vehicle and Vehicle-to-Infrastructure, has opened the door for a plethora of interesting applications to thrive. Some of these applications, known as infotainment applications, focus on enhancing the passengers' experience. They have rigid requirements in terms of delivery delay and throughput. Numerous approaches have been proposed, at medium access control and routing layers, to enhance the quality of service of such applications. However, existing schemes have several shortcomings. Subsequently, the design of new and efficient approaches is vital for the proper functioning of infotainment applications. This work proposes three schemes. The first is a novel routing protocol, labeled SCRP. It leverages real-time vehicular traffic information to create backbones over road segments and connect them at intersections using bridge nodes. These nodes are responsible for collecting connectivity and delay information, which are used to select routing paths with low end-to-end delay. The second is an altruistic service channel selection scheme, labeled ASSCH. It first collects real-time service channels information and feeds it to a stochastic model that predicts the state of these channels in the near future. The least congested channels are then selected to be used. The third is an analytical model for the performance of the IEEE 802.11p Enhanced Distributed Channel Access mechanism that considers various factors, including the transmission opportunity (TXOP), unexploited by IEEE 802.11p

    Orchestration de web services fiables

    Get PDF
    L’Informatique Orienté Services représente un paradigme pour construire des applications distribuées sur Internet. L’Architecture Orientée Services(SOA) est un style architectural qui permet le développement de ces applications à base de services. Au cours de la dernière décennie, l’orchestration des services Web est devenue un domaine très actif dans la recherche scientifique et académique. Bien que de nombreux défis liés à l’orchestration aient été abordés, la fiabilité de l’orchestration et de sa vérification restent encore un sujet ouvert, prérequis et important de fait que ces orchestrations affectent aujourd’hui plusieurs activités quotidiennes. Cette thèse focalise sur le sujet d’orchestration des Services Web Fiables. En particulier, elle contribue avec un ensemble d’approches, de techniques et d’outils pour améliorer la sélection et l’orchestration des services web fiables. Premièrement, elle affine les phases du cycle de vie d’orchestration de services web afin d’assurer une vérification continuée de fiabilité lors des phases de conception et d’exécution. En outre, elle propose une architecture conceptuelle basée sur un registre de service amélioré, pour la mise en œuvre d’orchestrations fiables. Deuxièmement, elle présente une approche de mesure de similarité entre les services web. L’approche repose sur la comparaison des interfaces WSDL de services. L’approche sert à identifier les relations de similarité, de substituabilité et de composabilité entre services. L’outil WSSIM a été développé pour mettre en œuvre l’approche proposée. Pour validation, l’outil a été expérimenté avec un ensemble important de services web réels. Troisièmement, la thèse contribue avec une approche pour l’identification des substituts de services simples et complexes. L’approche utilise les techniques de mesure de similarité, la classification de service avec FCA et l’analyse de fiabilité pour identifier et sélectionner les meilleures substitutes. Un ensemble d’algorithmes aient été proposés pour décrire le processus d’identification. Quatrièmement, pour examiner la réputation des services comme un autre critère de fiabilité, la thèse introduit un Framework et un modèle mathématique pour la gestion de réputation de service We

    Search improvement within the geospatial web in the context of spatial data infrastructures

    Get PDF
    El trabajo desarrollado en esta tesis doctoral demuestra que es posible mejorar la búsqueda en el contexto de las Infraestructuras de Datos Espaciales mediante la aplicación de técnicas y buenas prácticas de otras comunidades científicas, especialmente de las comunidades de la Web y de la Web Semántica (por ejemplo, Linked Data). El uso de las descripciones semánticas y las aproximaciones basadas en el contenido publicado por la comunidad geoespacial pueden ayudar en la búsqueda de información sobre los fenómenos geográficos, y en la búsqueda de recursos geoespaciales en general. El trabajo comienza con un análisis de una aproximación para mejorar la búsqueda de las entidades geoespaciales desde la perspectiva de geocodificación tradicional. La arquitectura de geocodificación compuesta propuesta en este trabajo asegura una mejora de los resultados de geocodificación gracias a la utilización de diferentes proveedores de información geográfica. En este enfoque, el uso de patrones estructurales de diseño y ontologías en esta aproximación permite una arquitectura avanzada en términos de extensibilidad, flexibilidad y adaptabilidad. Además, una arquitectura basada en la selección de servicio de geocodificación permite el desarrollo de una metodología de la georreferenciación de diversos tipos de información geográfica (por ejemplo, direcciones o puntos de interés). A continuación, se presentan dos aplicaciones representativas que requieren una caracterización semántica adicional de los recursos geoespaciales. El enfoque propuesto en este trabajo utiliza contenidos basados en heurísticas para el muestreo de un conjunto de recursos geopesaciales. La primera parte se dedica a la idea de la abstracción de un fenómeno geográfico de su definición espacial. La investigación muestra que las buenas prácticas de la Web Semántica se puede reutilizar en el ámbito de una Infraestructura de Datos Espaciales para describir los servicios geoespaciales estandarizados por Open Geospatial Consortium por medio de geoidentificadores (es decir, por medio de las entidades de una ontología geográfica). La segunda parte de este capítulo desglosa la aquitectura y componentes de un servicio de geoprocesamiento para la identificación automática de ortoimágenes ofrecidas a través de un servicio estándar de publicación de mapas (es decir, los servicios que siguen la especificación OGC Web Map Service). Como resultado de este trabajo se ha propuesto un método para la identificación de los mapas ofrecidos por un Web Map Service que son ortoimágenes. A continuación, el trabajo se dedica al análisis de cuestiones relacionadas con la creación de los metadatos de recursos de la Web en el contexto del dominio geográfico. Este trabajo propone una arquitectura para la generación automática de conocimiento geográfico de los recursos Web. Ha sido necesario desarrollar un método para la estimación de la cobertura geográfica de las páginas Web. Las heurísticas propuestas están basadas en el contenido publicado por os proveedores de información geográfica. El prototipo desarrollado es capaz de generar metadatos. El modelo generado contiene el conjunto mínimo recomendado de elementos requeridos por un catálogo que sigue especificación OGC Catalogue Service for the Web, el estandar recomendado por deiferentes Infraestructuras de Datos Espaciales (por ejemplo, the Infrastructure for Spatial Information in the European Community (INSPIRE)). Además, este estudio determina algunas características de la Web Geoespacial actual. En primer lugar, ofrece algunas características del mercado de los proveedores de los recursos Web de la información geográfica. Este estudio revela algunas prácticas de la comunidad geoespacial en la producción de metadatos de las páginas Web, en particular, la falta de metadatos geográficos. Todo lo anterior es la base del estudio de la cuestión del apoyo a los usuarios no expertos en la búsqueda de recursos de la Web Geoespacial. El motor de búsqueda dedicado a la Web Geoespacial propuesto en este trabajo es capaz de usar como base un motor de búsqueda existente. Por otro lado, da soporte a la búsqueda exploratoria de los recursos geoespaciales descubiertos en la Web. El experimento sobre la precisión y la recuperación ha demostrado que el prototipo desarrollado en este trabajo es al menos tan bueno como el motor de búsqueda remoto. Un estudio dedicado a la utilidad del sistema indica que incluso los no expertos pueden realizar una tarea de búsqueda con resultados satisfactorios

    Localization of nodes in wired and wireless networks

    Get PDF
    This thesis focuses on the implementation of algorithms for localization of nodes in wired and wireless networks. The thesis is organized into two papers. The first paper presents the localization algorithms based on time of arrival (TOA) and time difference of arrival (TDOA) techniques for computer networks such as the Internet by using round-trip-time (RTT) measurements obtained from known positions of the gateway nodes. The RTT values provide an approximate measure of distance between the gateway nodes and an unknown node. The least squares technique is then used to obtain an estimated position of the unknown node. The second paper presents localization of an unknown node during route setup messages in wireless ad hoc and sensor networks using a new routing protocol. A proactive multi-interface multichannel routing (MMCR) protocol, recently developed at Missouri S&T, was implemented on the Missouri S&T motes. This protocol calculates link costs based on a composite metric defined using the available end-to-end delay, energy utilization, and bandwidth, and it chooses the path that minimizes the link cost factor to effectively route the information to the required destination. Experimental results indicate enhanced performance in terms of quality of service, and implementation of this protocol requires no modification to the current IEEE 802.11 MAC protocol. Received signal strength indicator (RSSI) values are recorded from the relay nodes (gateway nodes) to the unknown node during route setup messages. The location of the unknown node is estimated using these values with some a priori profiling and the known positions of the relay nodes as inputs to the least squares technique --Abstract, page iv

    Ontology-based Consistent Specification and Scalable Execution of Sensor Data Acquisition Plans in Cross-Domain loT Platforms

    Get PDF
    Nowadays there is an increased number of vertical Internet of Things (IoT) applications that have been developed within IoT Platforms that often do not interact with each other because of the adoption of different standards and formats. Several efforts are devoted to the construction of software infrastructures that facilitate the interoperability among heterogeneous cross-domain IoT platforms for the realization of horizontal applications. Even if their realization poses different challenges across all layers of the network stack, in this thesis we focus on the interoperability issues that arise at the data management layer. Starting from a flexible multi-granular Spatio-Temporal-Thematic data model according to which events generated by different kinds of sensors can be represented, we propose a Semantic Virtualization approach according to which the sensors belonging to different IoT platforms and the schema of the produced event streams are described in a Domain Ontology, obtained through the extension of the well-known ontologies (SSN and IoT-Lite ontologies) to the needs of a specific domain. Then, these sensors can be exploited for the creation of Data Acquisition Plans (DAPs) by means of which the streams of events can be filtered, merged, and aggregated in a meaningful way. Notions of soundness and consistency are introduced to bind the output streams of the services contained in the DAP with the Domain Ontology for providing a semantic description of its final output. The facilities of the \streamLoader prototype are finally presented for supporting the domain experts in the Semantic Virtualization of the sensors and for the construction of meaningful DAPs. Different graphical facilities have been developed for supporting domain experts in the development of complex DAPs. The system provides also facilities for their syntax-based translations in the Apache Spark Streaming language and execution in real time in a distributed cluster of machines

    Cloud Services Brokerage for Mobile Ubiquitous Computing

    Get PDF
    Recently, companies are adopting Mobile Cloud Computing (MCC) to efficiently deliver enterprise services to users (or consumers) on their personalized devices. MCC is the facilitation of mobile devices (e.g., smartphones, tablets, notebooks, and smart watches) to access virtualized services such as software applications, servers, storage, and network services over the Internet. With the advancement and diversity of the mobile landscape, there has been a growing trend in consumer attitude where a single user owns multiple mobile devices. This paradigm of supporting a single user or consumer to access multiple services from n-devices is referred to as the Ubiquitous Cloud Computing (UCC) or the Personal Cloud Computing. In the UCC era, consumers expect to have application and data consistency across their multiple devices and in real time. However, this expectation can be hindered by the intermittent loss of connectivity in wireless networks, user mobility, and peak load demands. Hence, this dissertation presents an architectural framework called, Cloud Services Brokerage for Mobile Ubiquitous Cloud Computing (CSB-UCC), which ensures soft real-time and reliable services consumption on multiple devices of users. The CSB-UCC acts as an application middleware broker that connects the n-devices of users to the multi-cloud services. The designed system determines the multi-cloud services based on the user's subscriptions and the n-devices are determined through device registration on the broker. The preliminary evaluations of the designed system shows that the following are achieved: 1) high scalability through the adoption of a distributed architecture of the brokerage service, 2) providing soft real-time application synchronization for consistent user experience through an enhanced mobile-to-cloud proximity-based access technique, 3) reliable error recovery from system failure through transactional services re-assignment to active nodes, and 4) transparent audit trail through access-level and context-centric provenance

    Intelligent Circuits and Systems

    Get PDF
    ICICS-2020 is the third conference initiated by the School of Electronics and Electrical Engineering at Lovely Professional University that explored recent innovations of researchers working for the development of smart and green technologies in the fields of Energy, Electronics, Communications, Computers, and Control. ICICS provides innovators to identify new opportunities for the social and economic benefits of society.  This conference bridges the gap between academics and R&D institutions, social visionaries, and experts from all strata of society to present their ongoing research activities and foster research relations between them. It provides opportunities for the exchange of new ideas, applications, and experiences in the field of smart technologies and finding global partners for future collaboration. The ICICS-2020 was conducted in two broad categories, Intelligent Circuits & Intelligent Systems and Emerging Technologies in Electrical Engineering
    • …
    corecore