4,058 research outputs found

    Wireless model-based predictive networked control system over cooperative wireless network

    Get PDF
    Owing to their distributed architecture, networked control systems (NCSs) are proven to be feasible in scenarios where a spatially distributed feedback control system is required. Traditionally, such NCSs operate over real-time wired networks. Recently, in order to achieve the utmost flexibility, scalability, ease of deployment, and maintainability, wireless networks such as IEEE 802.11 wireless local area networks (LANs) are being preferred over dedicated wired networks. However, conventional NCSs with event-triggered controllers and actuators cannot operate over such general purpose wireless networks since the stability of the system is compromised due to unbounded delays and unpredictable packet losses that are typical in the wireless medium. Approaching the wireless networked control problem from two perspectives, this work introduces a practical wireless NCS and an implementation of a cooperative medium access control protocol that work jointly to achieve decent control under severe impairments, such as unbounded delay, bursts of packet loss and ambient wireless traffic. The proposed system is evaluated on a dedicated test platform under numerous scenarios and significant performance gains are observed, making cooperative communications a strong candidate for improving the reliability of industrial wireless networks

    An efficient data transmission policy in an integrated voice-data ds-cdma network

    Get PDF
    CDMA schemes appear to be promising access techniques for coping with the requirements of third-generation mobile systems, mainly because of their flexibility. This paper proposes an adaptive S-ALOHA DS-CDMA access scheme as a method for integrating non-real time (i.e. Internet applications) and real-time (i.e. voice) services, by exploiting the potentials of CDMA under time-varying conditions. The adaptive component terminals autonomously change their transmission rate according to the total (voice+data) channel occupancy, so that the minimum possible data delay is almost always achieved.Peer ReviewedPostprint (published version

    Controlling Network Latency in Mixed Hadoop Clusters: Do We Need Active Queue Management?

    Get PDF
    With the advent of big data, data center applications are processing vast amounts of unstructured and semi-structured data, in parallel on large clusters, across hundreds to thousands of nodes. The highest performance for these batch big data workloads is achieved using expensive network equipment with large buffers, which accommodate bursts in network traffic and allocate bandwidth fairly even when the network is congested. Throughput-sensitive big data applications are, however, often executed in the same data center as latency-sensitive workloads. For both workloads to be supported well, the network must provide both maximum throughput and low latency. Progress has been made in this direction, as modern network switches support Active Queue Management (AQM) and Explicit Congestion Notifications (ECN), both mechanisms to control the level of queue occupancy, reducing the total network latency. This paper is the first study of the effect of Active Queue Management on both throughput and latency, in the context of Hadoop and the MapReduce programming model. We give a quantitative comparison of four different approaches for controlling buffer occupancy and latency: RED and CoDel, both standalone and also combined with ECN and DCTCP network protocol, and identify the AQM configurations that maintain Hadoop execution time gains from larger buffers within 5%, while reducing network packet latency caused by bufferbloat by up to 85%. Finally, we provide recommendations to administrators of Hadoop clusters as to how to improve latency without degrading the throughput of batch big data workloads.The research leading to these results has received funding from the European Unions Seventh Framework Programme (FP7/2007–2013) under grant agreement number 610456 (Euroserver). The research was also supported by the Ministry of Economy and Competitiveness of Spain under the contracts TIN2012-34557 and TIN2015-65316-P, Generalitat de Catalunya (contracts 2014-SGR-1051 and 2014-SGR-1272), HiPEAC-3 Network of Excellence (ICT- 287759), and the Severo Ochoa Program (SEV-2011-00067) of the Spanish Government.Peer ReviewedPostprint (author's final draft

    JiTS: Just-in-Time Scheduling for Real-Time Sensor Data Dissemination

    Full text link
    We consider the problem of real-time data dissemination in wireless sensor networks, in which data are associated with deadlines and it is desired for data to reach the sink(s) by their deadlines. To this end, existing real-time data dissemination work have developed packet scheduling schemes that prioritize packets according to their deadlines. In this paper, we first demonstrate that not only the scheduling discipline but also the routing protocol has a significant impact on the success of real-time sensor data dissemination. We show that the shortest path routing using the minimum number of hops leads to considerably better performance than Geographical Forwarding, which has often been used in existing real-time data dissemination work. We also observe that packet prioritization by itself is not enough for real-time data dissemination, since many high priority packets may simultaneously contend for network resources, deteriorating the network performance. Instead, real-time packets could be judiciously delayed to avoid severe contention as long as their deadlines can be met. Based on this observation, we propose a Just-in-Time Scheduling (JiTS) algorithm for scheduling data transmissions to alleviate the shortcomings of the existing solutions. We explore several policies for non-uniformly delaying data at different intermediate nodes to account for the higher expected contention as the packet gets closer to the sink(s). By an extensive simulation study, we demonstrate that JiTS can significantly improve the deadline miss ratio and packet drop ratio compared to existing approaches in various situations. Notably, JiTS improves the performance requiring neither lower layer support nor synchronization among the sensor nodes

    LPDQ: a self-scheduled TDMA MAC protocol for one-hop dynamic lowpower wireless networks

    Get PDF
    Current Medium Access Control (MAC) protocols for data collection scenarios with a large number of nodes that generate bursty traffic are based on Low-Power Listening (LPL) for network synchronization and Frame Slotted ALOHA (FSA) as the channel access mechanism. However, FSA has an efficiency bounded to 36.8% due to contention effects, which reduces packet throughput and increases energy consumption. In this paper, we target such scenarios by presenting Low-Power Distributed Queuing (LPDQ), a highly efficient and low-power MAC protocol. LPDQ is able to self-schedule data transmissions, acting as a FSA MAC under light traffic and seamlessly converging to a Time Division Multiple Access (TDMA) MAC under congestion. The paper presents the design principles and the implementation details of LPDQ using low-power commercial radio transceivers. Experiments demonstrate an efficiency close to 99% that is independent of the number of nodes and is fair in terms of resource allocation.Peer ReviewedPostprint (author’s final draft

    Effect of Energy Harvesting on Stable Throughput in Cooperative Relay Systems

    Full text link
    In this paper, the impact of energy constraints on a two-hop network with a source, a relay and a destination under random medium access is studied. A collision channel with erasures is considered, and the source and the relay nodes have energy harvesting capabilities and an unlimited battery to store the harvested energy. Additionally, the source and the relay node have external traffic arrivals and the relay forwards a fraction of the source node's traffic to the destination; the cooperation is performed at the network level. An inner and an outer bound of the stability region for a given transmission probability vector are obtained. Then, the closure of the inner and the outer bound is obtained separately and they turn out to be identical. This work is not only a step in connecting information theory and networking, by studying the maximum stable throughput region metric but also it taps the relatively unexplored and important domain of energy harvesting and assesses the effect of that on this important measure.Comment: 20 pages, 4 figure

    Deep Reinforcement Learning for Real-Time Optimization in NB-IoT Networks

    Get PDF
    NarrowBand-Internet of Things (NB-IoT) is an emerging cellular-based technology that offers a range of flexible configurations for massive IoT radio access from groups of devices with heterogeneous requirements. A configuration specifies the amount of radio resource allocated to each group of devices for random access and for data transmission. Assuming no knowledge of the traffic statistics, there exists an important challenge in "how to determine the configuration that maximizes the long-term average number of served IoT devices at each Transmission Time Interval (TTI) in an online fashion". Given the complexity of searching for optimal configuration, we first develop real-time configuration selection based on the tabular Q-learning (tabular-Q), the Linear Approximation based Q-learning (LA-Q), and the Deep Neural Network based Q-learning (DQN) in the single-parameter single-group scenario. Our results show that the proposed reinforcement learning based approaches considerably outperform the conventional heuristic approaches based on load estimation (LE-URC) in terms of the number of served IoT devices. This result also indicates that LA-Q and DQN can be good alternatives for tabular-Q to achieve almost the same performance with much less training time. We further advance LA-Q and DQN via Actions Aggregation (AA-LA-Q and AA-DQN) and via Cooperative Multi-Agent learning (CMA-DQN) for the multi-parameter multi-group scenario, thereby solve the problem that Q-learning agents do not converge in high-dimensional configurations. In this scenario, the superiority of the proposed Q-learning approaches over the conventional LE-URC approach significantly improves with the increase of configuration dimensions, and the CMA-DQN approach outperforms the other approaches in both throughput and training efficiency
    • …
    corecore