220 research outputs found

    Efficient Actor Recovery Paradigm For Wireless Sensor And Actor Networks

    Get PDF
    Wireless sensor networks (WSNs) are becoming widely used worldwide. Wireless Sensor and Actor Networks (WSANs) represent a special category of WSNs wherein actors and sensors collaborate to perform specific tasks. WSANs have become one of the most preeminent emerging type of WSNs. Sensors with nodes having limited power resources are responsible for sensing and transmitting events to actor nodes. Actors are high-performance nodes equipped with rich resources that have the ability to collect, process, transmit data and perform various actions. WSANs have a unique architecture that distinguishes them from WSNs. Due to the characteristics of WSANs, numerous challenges arise. Determining the importance of factors usually depends on the application requirements. The actor nodes are the spine of WSANs that collaborate to perform the specific tasks in an unsubstantiated and uneven environment. Thus, there is a possibility of high failure rate in such unfriendly scenarios due to several factors such as power fatigue of devices, electronic circuit failure, software errors in nodes or physical impairment of the actor nodes and inter-actor connectivity problem. It is essential to keep inter-actor connectivity in order to insure network connectivity. Thus, it is extremely important to discover the failure of a cut-vertex actor and network-disjoint in order to improve the Quality-of-Service (QoS). For network recovery process from actor node failure, optimal re-localization and coordination techniques should take place. In this work, we propose an efficient actor recovery (EAR) paradigm to guarantee the contention-free traffic-forwarding capacity. The EAR paradigm consists of Node Monitoring and Critical Node Detection (NMCND) algorithm that monitors the activities of the nodes to determine the critical node. In addition, it replaces the critical node with backup node prior to complete node-failure which helps balances the network performance. The packet is handled using Network Integration and Message Forwarding (NIMF) algorithm that determines the source of forwarding the packets (Either from actor or sensor). This decision-making capability of the algorithm controls the packet forwarding rate to maintain the network for longer time. Furthermore, for handling the proper routing strategy, Priority-Based Routing for Node Failure Avoidance (PRNFA) algorithm is deployed to decide the priority of the packets to be forwarded based on the significance of information available in the packet. To validate the effectiveness of the proposed EAR paradigm, we compare the performance of our proposed work with state-of the art localization algorithms. Our experimental results show superior performance in regards to network life, residual energy, reliability, sensor/ actor recovery time and data recovery

    Energy-aware distributed routing algorithm to tolerate network failure in wireless sensor networks

    Get PDF
    Wireless Sensor Networks are prone to link/node failures due to various environmental hazards such as interference and internal faults in deployed sensor nodes. Such failures can result in a disconnection in part of the network and the sensed data being unable to obtain a route to the sink(s), i.e. a network failure. Network failures potentially degrade the Quality of Service (QoS) of Wireless Sensor Networks (WSNs). It is very difficult to monitor network failures using a manual operator in a harsh or hostile environment. In such environments, communication links can easy fail because of node unequal energy depletion and hardware failure or invasion. Thus it is desirable that deployed sensor nodes are capable of overcoming network failures. In this paper, we consider the problem of tolerating network failures seen by deployed sensor nodes in a WSN. We first propose a novel clustering algorithm for WSNs, termed Distributed Energy Efficient Heterogeneous Clustering (DEEHC) that selects cluster heads according to the residual energy of deployed sensor nodes with the aid of a secondary timer. During the clustering phase, each sensor node finds k-vertex disjoint paths to cluster heads depending on the energy level of its neighbor sensor nodes. We then present a k-Vertex Disjoint Path Routing (kVDPR) algorithm where each cluster head finds k-vertex disjoint paths to the base station and relays their aggregate data to the base station. Furthermore, we also propose a novel Route Maintenance Mechanism (RMM) that can repair k-vertex disjoint paths throughout the monitoring session. The resulting WSNs become tolerant to k-1 failures in the worst case. The proposed scheme has been extensively tested using various network scenarios and compared to the existing state of the art approaches to show the effectiveness of the proposed scheme

    A RELIABLE ROUTING MECHANISM WITH ENERGY-EFFICIENT NODE SELECTION FOR DATA TRANSMISSION USING A GENETIC ALGORITHM IN WIRELESS SENSOR NETWORK

    Get PDF
    Energy-efficient and reliable data routing is critical in Wireless Sensor Networks (WSNs) application scenarios. Due to oscillations in wireless links in adverse environmental conditions, sensed data may not be sent to a sink node. As a result of wireless connectivity fluctuations, packet loss may occur. However, retransmission-based approaches are used to improve reliable data delivery. These approaches need a high quantity of data transfers for reliable data collection. Energy usage and packet delivery delays increase as a result of an increase in data transmissions. An energy-efficient data collection approach based on a genetic algorithm has been suggested in this paper to determine the most energy-efficient and reliable data routing in wireless sensor networks. The proposed algorithm reduced the number of data transmissions, energy consumption, and delay in network packet delivery. However, increased network lifetime. Furthermore, simulation results demonstrated the efficacy of the proposed method, considering the parameters energy consumption, network lifetime, number of data transmissions, and average delivery delay

    LOCALIZED MOVEMENT CONTROL CONNECTIVITY RESTORATION ALGORITHMS FOR WIRELESS SENSOR AND ACTOR NETWORKS

    Get PDF
    Wireless Sensor and Actor Networks (WSANs) are gaining an increased interest because of their suitability for mission-critical applications that require autonomous and intelligent interaction with the environment. Hazardous application environments such as forest fire monitoring, disaster management, search and rescue, homeland security, battlefield reconnaissance, etc. make actors susceptible to physical damage. Failure of a critical (i.e. cut-vertex) actor partitions the inter-actor network into disjointed segments while leaving a coverage hole. Maintaining inter-actor connectivity is extremely important in mission-critical applications of WSANs where actors have to quickly plan an optimal coordinated response to detected events. Some proactive approaches pursued in the literature deploy redundant nodes to provide fault tolerance; however, this necessitates a large actor count that leads to higher cost and becomes impractical. On the other hand, the harsh environment strictly prohibits an external intervention to replace a failed node. Meanwhile, reactive approaches might not be suitable for time-sensitive applications. The autonomous and unattended nature of WSANs necessitates a self-healing and agile recovery process that involves existing actors to mend the severed inter-actor connectivity by reconfiguring the topology. Moreover, though the possibility of simultaneous multiple actor failure is rare, it may be precipitated by a hostile environment and disastrous events. With only localized information, recovery from such failures is extremely challenging. Furthermore, some applications may impose application-level constraints while recovering from a node failure. In this dissertation, we address the challenging connectivity restoration problem while maintaining minimal network state information. We have exploited the controlled movement of existing (internal) actors to restore the lost connectivity while minimizing the impact on coverage. We have pursued distributed greedy heuristics. This dissertation presents four novel approaches for recovering from node failure. In the first approach, volunteer actors exploit their partially utilized transmission power and reposition themselves in such a way that the connectivity is restored. The second approach identifies critical actors in advance, designates them preferably as noncritical backup nodes that replace the failed primary if such contingency arises in the future. In the third approach, we design a distributed algorithm that recovers from a special case of multiple simultaneous failures. The fourth approach factors in application-level constraints on the mobility of actors while recovering from node failure and strives to minimize the impact of critical node failure on coverage and connectivity. The performance of proposed approaches is analyzed and validated through extensive simulations. Simulation results confirm the effectiveness of proposed approaches that outperform the best contemporary schemes found in literature

    Dagstuhl News January - December 2008

    Get PDF
    "Dagstuhl News" is a publication edited especially for the members of the Foundation "Informatikzentrum Schloss Dagstuhl" to thank them for their support. The News give a summary of the scientific work being done in Dagstuhl. Each Dagstuhl Seminar is presented by a small abstract describing the contents and scientific highlights of the seminar as well as the perspectives or challenges of the research topic

    Energy-aware medium access control protocols for wireless sensors network applications

    Get PDF
    The main purpose of this thesis was to investigate energy efficient Medium Access Control (MAC) protocols designed to extend the lifetime of a wireless sensor network application, such as tracking, environment monitoring, home security, patient monitoring, e.g., foetal monitoring in the last weeks of pregnancy. From the perspective of communication protocols, energy efficiency is one of the most important issues, and can be addressed at each layer of the protocol stack; however, our research only focuses on the medium access control (MAC) layer. An energy efficient MAC protocol was designed based on modifications and optimisations for a synchronized power saving Sensor MAC (SMAC) protocol, which has three important components: periodic listen and sleep, collision and overhearing avoidance and message passing. The Sensor Block Acknowledgement (SBACK) MAC protocol is proposed, which combines contention-based, scheduling-based and block acknowledgement-based schemes to achieve energy efficiency. In SBACK, the use of ACK control packets is reduced since it will not have an ACK packet for every DATA packet sent; instead, one special packet called Block ACK Response will be used at the end of the transmission of all data packets. This packet informs the sender of how many packets were received by the receiver, reducing the number of ACK control packets we intended to reduce the power consumption for the nodes. Hence more useful data packets can be transmitted. A comparison study between SBACK and SMAC protocol is also performed. Considering 0% of packet losses, SBACK decreases the energy consumption when directly compared with S-MAC, we will have always a decrease of energy consumption. Three different transceivers will be used and considering a packet loss of 10% we will have a decrease of energy consumption between 10% and 0.1% depending on the transceiver. When there are no retransmissions of packets, SBACK only achieve worst performance when the number of fragments is less than 12, after that the decrease of average delay increases with the increase of the fragments sent. When 10% of the packets need retransmission only for the TR1000 transceiver worst results occurs in terms of energy waste, all other transceivers (CC2420 and AT86RF230) achieve better results. In terms of delay if we need to retransmit more than 10 packets the SBACK protocol always achieves better performance when comparing with the other MAC protocols that uses ACK

    The role of communication systems in smart grids: Architectures, technical solutions and research challenges

    Get PDF
    The purpose of this survey is to present a critical overview of smart grid concepts, with a special focus on the role that communication, networking and middleware technologies will have in the transformation of existing electric power systems into smart grids. First of all we elaborate on the key technological, economical and societal drivers for the development of smart grids. By adopting a data-centric perspective we present a conceptual model of communication systems for smart grids, and we identify functional components, technologies, network topologies and communication services that are needed to support smart grid communications. Then, we introduce the fundamental research challenges in this field including communication reliability and timeliness, QoS support, data management services, and autonomic behaviors. Finally, we discuss the main solutions proposed in the literature for each of them, and we identify possible future research directions

    Planetary Scale Data Storage

    Get PDF
    The success of virtualization and container-based application deployment has fundamentally changed computing infrastructure from dedicated hardware provisioning to on-demand, shared clouds of computational resources. One of the most interesting effects of this shift is the opportunity to localize applications in multiple geographies and support mobile users around the globe. With relatively few steps, an application and its data systems can be deployed and scaled across continents and oceans, leveraging the existing data centers of much larger cloud providers. The novelty and ease of a global computing context means that we are closer to the advent of an Oceanstore, an Internet-like revolution in personalized, persistent data that securely travels with its users. At a global scale, however, data systems suffer from physical limitations that significantly impact its consistency and performance. Even with modern telecommunications technology, the latency in communication from Brazil to Japan results in noticeable synchronization delays that violate user expectations. Moreover, the required scale of such systems means that failure is routine. To address these issues, we explore consistency in the implementation of distributed logs, key/value databases and file systems that are replicated across wide areas. At the core of our system is hierarchical consensus, a geographically-distributed consensus algorithm that provides strong consistency, fault tolerance, durability, and adaptability to varying user access patterns. Using hierarchical consensus as a backbone, we further extend our system from data centers to edge regions using federated consistency, an adaptive consistency model that gives satellite replicas high availability at a stronger global consistency than existing weak consistency models. In a deployment of 105 replicas in 15 geographic regions across 5 continents, we show that our implementation provides high throughput, strong consistency, and resiliency in the face of failure. From our experimental validation, we conclude that planetary-scale data storage systems can be implemented algorithmically without sacrificing consistency or performance

    A wireless sensor network system for border security and crossing detection

    Get PDF
    The protection of long stretches of countries’ borders has posed a number of challenges. Effective and continuous monitoring of a border requires the implementation of multi-surveillance technologies, such as Wireless Sensor Networks (WSN), that work as an integrated unit to meet the desired goals. The research presented in this thesis investigates the application of topologically Linear WSN (LWSNs) to international border monitoring and surveillance. The main research questions studied here are: What is the best form of node deployment and hierarchy? What is the minimum number of sensor nodes to achieve k− barrier coverage in a given belt region? iven an appropriate network density, how do we determine if a region is indeed k−barrier covered? What are the factors that affect barrier coverage? How to organise nodes into logical segments to perform in-network processing of data? How to transfer information from the networks to the end users while maintaining critical QoS measures such as timeliness and accuracy. To address these questions, we propose an architecture that specifies a mechanism to assign nodes to various network levels depending on their location. These levels are used by a cross-layer communication protocol to achieve data delivery at the lowest possible cost and minimal delivery delay. Building on this levelled architecture, we study the formation of weak and strong barriers and how they determine border crossing detection probability. We propose new method to calculate the required node density to provide higher intruder detection rate. Then, we study the effect of people movement models on the border crossing detection probability. At the data link layer, new energy balancing along with shifted MAC protocol are introduced to further increase the network lifetime and delivery speed. In addition, at network layer, a routing protocol called Level Division raph (LD ) is developed. LD utilises a complex link cost measurement to insure best QoS data delivery to the sink node at the lowest possible cost. The proposed system has the ability to work independently or cooperatively with other monitoring technologies, such as drowns and mobile monitoring stations. The performance of the proposed work is extensively evaluated analytically and in simulation using real-life conditions and parameters. The simulation results show significant performance gains when comparing LD to its best rivals in the literature Dynamic Source Routing. Compared to DSR, LD achieves higher performance in terms of average end-to-end delays by up to 95%, packet delivery ratio by up to 20%, and throughput by up to 60%, while maintaining similar performance in terms of normalised routing load and energy consumption
    corecore