14,337 research outputs found

    Data Transmission with Reduced Delay for Distributed Acoustic Sensors

    Full text link
    This paper proposes a channel access control scheme fit to dense acoustic sensor nodes in a sensor network. In the considered scenario, multiple acoustic sensor nodes within communication range of a cluster head are grouped into clusters. Acoustic sensor nodes in a cluster detect acoustic signals and convert them into electric signals (packets). Detection by acoustic sensors can be executed periodically or randomly and random detection by acoustic sensors is event driven. As a result, each acoustic sensor generates their packets (50bytes each) periodically or randomly over short time intervals (400ms~4seconds) and transmits directly to a cluster head (coordinator node). Our approach proposes to use a slotted carrier sense multiple access. All acoustic sensor nodes in a cluster are allocated to time slots and the number of allocated sensor nodes to each time slot is uniform. All sensor nodes allocated to a time slot listen for packet transmission from the beginning of the time slot for a duration proportional to their priority. The first node that detect the channel to be free for its whole window is allowed to transmit. The order of packet transmissions with the acoustic sensor nodes in the time slot is autonomously adjusted according to the history of packet transmissions in the time slot. In simulations, performances of the proposed scheme are demonstrated by the comparisons with other low rate wireless channel access schemes.Comment: Accepted to IJDSN, final preprinted versio

    Handling Tradeoffs between Performance and Query-Result Quality in Data Stream Processing

    Get PDF
    Data streams in the form of potentially unbounded sequences of tuples arise naturally in a large variety of domains including finance markets, sensor networks, social media, and network traffic management. The increasing number of applications that require processing data streams with high throughput and low latency have promoted the development of data stream processing systems (DSPS). A DSPS processes data streams with continuous queries, which are issued once and return query results to users continuously as new tuples arrive. For stream-based applications, both the query-execution performance (in terms of, e.g., throughput and end-to-end latency) and the quality of produced query results (in terms of, e.g., accuracy and completeness) are important. However, a DSPS often needs to make tradeoffs between these two requirements, either because of the data imperfection within the streams, or because of the limited computation capacity of the DSPS itself. Performance versus result-quality tradeoffs caused by data imperfection are inevitable, because the quality of the incoming data is beyond the control of a DSPS, whereas tradeoffs caused by system limitations can be alleviated—even erased—by enhancing the DSPS itself. This dissertation seeks to advance the state of the art on handling the performance versus result-quality tradeoffs in data stream processing caused by the above two aspects of reasons. For tradeoffs caused by data imperfection, this dissertation focuses on the typical data-imperfection problem of stream disorder and proposes the concept of quality-driven disorder handling (QDDH). QDDH enables a DSPS to make flexible and user-configurable tradeoffs between the end-to-end latency and the query-result quality when dealing with stream disorder. Moreover, compared to existing disorder handling approaches, QDDH can significantly reduce the end-to-end latency, and at the same time provide users with desired query-result quality. In this dissertation, a generic buffer-based QDDH framework and three instantiations of the generic framework for distinct query types are presented. For tradeoffs caused by system limitations, this dissertation proposes a system-enhancement approach that combines the row-oriented and the column-oriented data layout and processing techniques in data stream processing to improve the throughput. To fully exploit the potential of such hybrid execution of continuous queries, a static, cost-based query optimizer is introduced. The optimizer works at the operator level and takes the unique property of execution plans of continuous queries—feasibility—into account

    A phased array-based method for damage detection and localization in thin plates

    Get PDF
    A method for damage localization based on the phased array idea has been developed. Four arrays oftransducers are used to perform a beam-forming procedure. Each array consists of nine transducersplaced along a line, which are able to excite and register elastic waves. The A0 Lamb wave mode hasbeen chosen for the localization method. The arrays are placed in such a way that the angulardifference between them is 458 and the rotation point is the middle transducer, which is common for allthe arrays. The idea has been tested on a square aluminium plate modeled by the Spectral Element Method. Two types of damage were considered, namely distributed damage, which was modeled asstiffness reduction, and cracks, modeled as separation of nodes between selected spectral elements.The plate is excited by a wave packet. The whole array system is placed in the middle of the plate.Each linear phased array in the system acts independently and produces maps of a scanned fieldbased on the beam-forming procedure. These maps are made of time signals (transferred to spacedomain) that represent the difference between the damaged plate signals and those from the intactplate. An algorithm was developed to join all four maps. The final map is modified by proposed signal processing algorithm to indicate the damaged area of the plate more precisely. The problem fordamage localization was investigated and exemplary maps confirming the effectiveness of theproposed system were obtained. It was also shown that the response of the introduced configurationremoves the ambiguity of damage localization normally present when a linear phased array is utilized.The investigation is based exclusively on numerical data

    Learning Scheduling Algorithms for Data Processing Clusters

    Full text link
    Efficiently scheduling data processing jobs on distributed compute clusters requires complex algorithms. Current systems, however, use simple generalized heuristics and ignore workload characteristics, since developing and tuning a scheduling policy for each workload is infeasible. In this paper, we show that modern machine learning techniques can generate highly-efficient policies automatically. Decima uses reinforcement learning (RL) and neural networks to learn workload-specific scheduling algorithms without any human instruction beyond a high-level objective such as minimizing average job completion time. Off-the-shelf RL techniques, however, cannot handle the complexity and scale of the scheduling problem. To build Decima, we had to develop new representations for jobs' dependency graphs, design scalable RL models, and invent RL training methods for dealing with continuous stochastic job arrivals. Our prototype integration with Spark on a 25-node cluster shows that Decima improves the average job completion time over hand-tuned scheduling heuristics by at least 21%, achieving up to 2x improvement during periods of high cluster load

    Streaming Similarity Self-Join

    Full text link
    We introduce and study the problem of computing the similarity self-join in a streaming context (SSSJ), where the input is an unbounded stream of items arriving continuously. The goal is to find all pairs of items in the stream whose similarity is greater than a given threshold. The simplest formulation of the problem requires unbounded memory, and thus, it is intractable. To make the problem feasible, we introduce the notion of time-dependent similarity: the similarity of two items decreases with the difference in their arrival time. By leveraging the properties of this time-dependent similarity function, we design two algorithmic frameworks to solve the sssj problem. The first one, MiniBatch (MB), uses existing index-based filtering techniques for the static version of the problem, and combines them in a pipeline. The second framework, Streaming (STR), adds time filtering to the existing indexes, and integrates new time-based bounds deeply in the working of the algorithms. We also introduce a new indexing technique (L2), which is based on an existing state-of-the-art indexing technique (L2AP), but is optimized for the streaming case. Extensive experiments show that the STR algorithm, when instantiated with the L2 index, is the most scalable option across a wide array of datasets and parameters

    FAST TCP: Motivation, Architecture, Algorithms, Performance

    Get PDF
    We describe FAST TCP, a new TCP congestion control algorithm for high-speed long-latency networks, from design to implementation. We highlight the approach taken by FAST TCP to address the four difficulties which the current TCP implementation has at large windows. We describe the architecture and summarize some of the algorithms implemented in our prototype. We characterize its equilibrium and stability properties. We evaluate it experimentally in terms of throughput, fairness, stability, and responsiveness
    • …
    corecore