1,457 research outputs found

    Exploring heterogeneity of unreliable machines for p2p backup

    Full text link
    P2P architecture is a viable option for enterprise backup. In contrast to dedicated backup servers, nowadays a standard solution, making backups directly on organization's workstations should be cheaper (as existing hardware is used), more efficient (as there is no single bottleneck server) and more reliable (as the machines are geographically dispersed). We present the architecture of a p2p backup system that uses pairwise replication contracts between a data owner and a replicator. In contrast to standard p2p storage systems using directly a DHT, the contracts allow our system to optimize replicas' placement depending on a specific optimization strategy, and so to take advantage of the heterogeneity of the machines and the network. Such optimization is particularly appealing in the context of backup: replicas can be geographically dispersed, the load sent over the network can be minimized, or the optimization goal can be to minimize the backup/restore time. However, managing the contracts, keeping them consistent and adjusting them in response to dynamically changing environment is challenging. We built a scientific prototype and ran the experiments on 150 workstations in the university's computer laboratories and, separately, on 50 PlanetLab nodes. We found out that the main factor affecting the quality of the system is the availability of the machines. Yet, our main conclusion is that it is possible to build an efficient and reliable backup system on highly unreliable machines (our computers had just 13% average availability)

    Active architecture for pervasive contextual services

    Get PDF
    International Workshop on Middleware for Pervasive and Ad-hoc Computing MPAC 2003), ACM/IFIP/USENIX International Middleware Conference (Middleware 2003), Rio de Janeiro, Brazil This work was supported by the FP5 Gloss project IST2000-26070, with partners at Trinity College Dublin and Université Joseph Fourier, and by EPSRC grants GR/M78403/GR/M76225, Supporting Internet Computation in Arbitrary Geographical Locations, and GR/R45154, Bulk Storage of XML Documents.Pervasive services may be defined as services that are available "to any client (anytime, anywhere)". Here we focus on the software and network infrastructure required to support pervasive contextual services operating over a wide area. One of the key requirements is a matching service capable of as-similating and filtering information from various sources and determining matches relevant to those services. We consider some of the challenges in engineering a globally distributed matching service that is scalable, manageable, and able to evolve incrementally as usage patterns, data formats, services, network topologies and deployment technologies change. We outline an approach based on the use of a peer-to-peer architecture to distribute user events and data, and to support the deployment and evolution of the infrastructure itself.Peer reviewe

    Active architecture for pervasive contextual services

    Get PDF
    Pervasive services may be defined as services that are available to any client (anytime, anywhere). Here we focus on the software and network infrastructure required to support pervasive contextual services operating over a wide area. One of the key requirements is a matching service capable of assimilating and filtering information from various sources and determining matches relevant to those services. We consider some of the challenges in engineering a globally distributed matching service that is scalable, manageable, and able to evolve incrementally as usage patterns, data formats, services, network topologies and deployment technologies change. We outline an approach based on the use of a peer-to-peer architecture to distribute user events and data, and to support the deployment and evolution of the infrastructure itself

    Reliable distributed data stream management in mobile environments

    Get PDF
    The proliferation of sensor technology, especially in the context of embedded systems, has brought forward novel types of applications that make use of streams of continuously generated sensor data. Many applications like telemonitoring in healthcare or roadside traffic monitoring and control particularly require data stream management (DSM) to be provided in a distributed, yet reliable way. This is even more important when DSM applications are deployed in a failure-prone distributed setting including resource-limited mobile devices, for instance in applications which aim at remotely monitoring mobile patients. In this paper, we introduce a model for distributed and reliable DSM. The contribution of this paper is threefold. First, in analogy to the SQL isolation levels, we define levels of reliability and describe necessary consistency constraints for distributed DSM that specify the tolerated loss, delay, or re-ordering of data stream elements, respectively. Second, we use this model to design and analyze an algorithm for reliable distributed DSM, namely efficient coordinated operator checkpointing (ECOC). We show that ECOC provides lossless and delay-limited reliable data stream management and thus can be used in critical application domains such as healthcare, where the loss of data stream elements can not be tolerated. Third, we present detailed performance evaluations of the ECOC algorithm running on mobile, resource-limited devices. In particular, we can show that ECOC provides a high level of reliability while, at the same time, featuring good performance characteristics with moderate resource consumption

    Unattended network operations technology assessment study. Technical support for defining advanced satellite systems concepts

    Get PDF
    The results are summarized of an unattended network operations technology assessment study for the Space Exploration Initiative (SEI). The scope of the work included: (1) identified possible enhancements due to the proposed Mars communications network; (2) identified network operations on Mars; (3) performed a technology assessment of possible supporting technologies based on current and future approaches to network operations; and (4) developed a plan for the testing and development of these technologies. The most important results obtained are as follows: (1) addition of a third Mars Relay Satellite (MRS) and MRS cross link capabilities will enhance the network's fault tolerance capabilities through improved connectivity; (2) network functions can be divided into the six basic ISO network functional groups; (3) distributed artificial intelligence technologies will augment more traditional network management technologies to form the technological infrastructure of a virtually unattended network; and (4) a great effort is required to bring the current network technology levels for manned space communications up to the level needed for an automated fault tolerance Mars communications network

    Self-management of machine-to-machine communications: a multi-models approach

    Get PDF
    International audienceMachine-to-Machine (M2M) paradigm apply to systems composed by numerous devices sharing information and making cooperative decisions with little or no human intervention. The M2M standard defined by the European Telecommunications Standards Institute (ETSI) is the only one providing an end-to-end view of the global M2M architecture. Noticeably, it furnishes a standardised framework for inter-operable M2M services that satisfies most of M2M modelling requirements. However, and even though M2M systems usually operate in highly evolving contexts, this standard does not address the issue of system adaptations. It is furthermore unsuitable for building self-managed systems. This paper introduces a multi-model approach for modelling manageable M2M systems. Said approach consists in a formal graph-based model on top of the ETSI M2M standard, alongside bi-directional updates that ensure layer coherency. Its fitness for enforcing self-management properties is demonstrated by designing high-level reconfiguration rules. Finally, its applicability is illustrated and evaluated using a smart-metering application

    Design and Implementation of a True Decentralized Autonomous Control Architecture for Microgrids

    Get PDF
    Microgrids can serve as an integral part of the future power distribution systems. Most microgrids are currently managed by centralized controllers. There are two major concerns associated with the centralized controllers. One is that the single controller can become performance and reliability bottleneck for the entire system and its failure can bring the entire system down. The second concern is the communication delays that can degrade the system performance. As a solution, a true decentralized control architecture for microgrids is developed and presented. Distributing the control functions to local agents decreases the possibility of network congestion, and leads to the mitigation of long distance transmission of critical commands. Decentralization will also enhance the reliability of the system since the single point of failure is eliminated. In the proposed architecture, primary and secondary microgrid controls layers are combined into one physical layer. Tertiary control is performed by the controller located at the grid point of connection. Each decentralized controller is responsible of multicasting its status and local measurements, creating a general awareness of the microgrid status among all decentralized controllers. The proof-of concept implementation provides a practical evidence of the successful mitigation of the drawback of control command transmission over the network. A Failure Management Unit comprises failure detection mechanisms and a recovery algorithm is proposed and applied to a microgrid case study. Coordination between controllers during the recovery period requires low-bandwidth communications, which has no significant overhead on the communication infrastructure. The proof-of-concept of the true decentralization of microgrid control architecture is implemented using Hardware-in-the-Loop platform. The test results show a robust detection and recovery outcome during a system failure. System test results show the robustness of the proposed architecture for microgrid energy management and control scenarios

    Design and Construction of a New Metering Hot Box for the In Situ Hygrothermal Measurement in Dynamic Conditions of Historic Masonries

    Get PDF
    The main purpose of the HeLLo project is to contribute to data available on the literature on the real hygrothermal behavior of historic walls and the suitability of insulation technologies. Furthermore, it also aims at minimizing the energy simulation errors at the design phase and at improving their conservation features. In this framework, one of the preliminary activities of the study is the creation of a real in situ hot box to measure and analyze different insulation technologies applied to a real historic wall, to quantify the hygrothermal performance of a masonry building. Inside this box, ‘traditional’ experiments can be carried out: recording heat flux, surface temperature, and air temperatures, as well as relative humidity values through the use of a new sensing system (composed of thermocouples and temperature/relative humidity combined sensors). Within this paper, the process of development, construction, and validation of this new metering box is exhibited. The new hot box, specifically studied for historic case studies, when compared to other boxes, presents other advantages compared to previous examples, widely exemplified
    • 

    corecore