358 research outputs found

    Design and Implementation of a Data Stream Management System

    Get PDF
    The amount of data stored by companies has grown exponentially over the last decade. Of late, data is being continuously collected for various purposes - click stream analysis, credit card transactions for fraud detection, weather monitoring, stock tickers in nancial services, link statistics in networking, user logins and web surng statistics, highway trac congestion analysis and so on. The data that is being collected is in the form of a stream - arrives continuously, at a variable rate, and can occupy potentially innite storage. As organizations have realized that fast and ecient processing of this data can help in protable predictions, there exists a need for developing systems to handle this collected data eectively. We present in this thesis, the architecture of a generic Database Stream Management System (DBSMS) to handle streaming data. While literature has provided insights into Data Stream Management Systems (DSMS), the DBSMS is a dierent approach that tries to integrate a DSMS with the traditional Database Management Systems (DBMS). We discuss the need for such a generic DBSMS and present the system that we have implemented using the discussed architecture. We also present the performance of our system, in terms of space taken, time taken to answer a query and the accuracy of the result compared to a DBMS. Finally, we conclude with brief discussion on certain goals and open challenges that are of interest and which still need to be addressed by the system

    Automatic parallelization of array-oriented programs for a multi-core machine

    Get PDF
    Abstract We present the work on automatic parallelization of array-oriented programs for multi-core machines. Source programs written in standard APL are translated by a parallelizing APL-to-C compiler into parallelized C code, i.e. C mixed with OpenMP directives. We describe techniques such as virtual operations and datapartitioning used to effectively exploit parallelism structured around array-primitives. We present runtime performance data, showing the speedup of the resulting parallelized code, using different numbers of threads and different problem sizes, on a 4-core machine, for several examples

    Inferring Symbolic Automata

    Get PDF
    We study the learnability of symbolic finite state automata, a model shown useful in many applications in software verification. The state-of-the-art literature on this topic follows the query learning paradigm, and so far all obtained results are positive. We provide a necessary condition for efficient learnability of SFAs in this paradigm, from which we obtain the first negative result. The main focus of our work lies in the learnability of SFAs under the paradigm of identification in the limit using polynomial time and data. We provide a necessary condition and a sufficient condition for efficient learnability of SFAs in this paradigm, from which we derive a positive and a negative result

    FPT: a Fixed-Point Accelerator for Torus Fully Homomorphic Encryption

    Full text link
    Fully Homomorphic Encryption is a technique that allows computation on encrypted data. It has the potential to change privacy considerations in the cloud, but computational and memory overheads are preventing its adoption. TFHE is a promising Torus-based FHE scheme that relies on bootstrapping, the noise-removal tool invoked after each encrypted logical/arithmetical operation. We present FPT, a Fixed-Point FPGA accelerator for TFHE bootstrapping. FPT is the first hardware accelerator to exploit the inherent noise present in FHE calculations. Instead of double or single-precision floating-point arithmetic, it implements TFHE bootstrapping entirely with approximate fixed-point arithmetic. Using an in-depth analysis of noise propagation in bootstrapping FFT computations, FPT is able to use noise-trimmed fixed-point representations that are up to 50% smaller than prior implementations. FPT is built as a streaming processor inspired by traditional streaming DSPs: it instantiates directly cascaded high-throughput computational stages, with minimal control logic and routing networks. We explore throughput-balanced compositions of streaming kernels with a user-configurable streaming width in order to construct a full bootstrapping pipeline. Our approach allows 100% utilization of arithmetic units and requires only a small bootstrapping key cache, enabling an entirely compute-bound bootstrapping throughput of 1 BS / 35us. This is in stark contrast to the classical CPU approach to FHE bootstrapping acceleration, which is typically constrained by memory and bandwidth. FPT is implemented and evaluated as a bootstrapping FPGA kernel for an Alveo U280 datacenter accelerator card. FPT achieves two to three orders of magnitude higher bootstrapping throughput than existing CPU-based implementations, and 2.5x higher throughput compared to recent ASIC emulation experiments.Comment: ACM CCS 202

    A Synchronisation Facility for a Stream Processing Coordination Language

    Get PDF
    In this thesis we present the AstraKahn project that aims to provide environment for stream processing applications with an automatic resource and concurrency management based on communication demand. At the moment the work on the thesis started, the project was at an early stage and there existed no software implementation. The aim of my work is to implement a stream synchronisation facility called synchroniser which is a part of the AstraKahn concept. Synchronisers are programmed in a dedicated language. The thesis focuses on the implementation of the language compiler to be integrated into the runtime system prototype being developed in parallel. AstraKahn structures streaming networks using a xed set of wiring patterns, including the serial replication. This pattern dynamically replicates its operand network conceptually in nitely many times and wires the replicas in a chain. AstraKahn provides an approach to extract messages from the chain based on the concept of xed point. The thesis explores the role of synchronisers in forming from a serial replication pipeline

    Point cloud data compression

    Get PDF
    The rapid growth in the popularity of Augmented Reality (AR), Virtual Reality (VR), and Mixed Reality (MR) experiences have resulted in an exponential surge of three-dimensional data. Point clouds have emerged as a commonly employed representation for capturing and visualizing three-dimensional data in these environments. Consequently, there has been a substantial research effort dedicated to developing efficient compression algorithms for point cloud data. This Master's thesis aims to investigate the current state-of-the-art lossless point cloud geometry compression techniques, explore some of these techniques in more detail and then propose improvements and/or extensions to enhance them and provide directions for future work on this topic
    corecore