6 research outputs found

    Affirmative sampling: theory and applications

    Get PDF
    Affirmative Sampling is a practical and efficient novel algorithm to obtain random samples of distinct elements from a data stream. Its most salient feature is that the size S of the sample will, on expectation, grow with the (unknown) number n of distinct elements in the data stream. As any distinct element has the same probability to be sampled, and the sample size is greater when the “diversity” (the number of distinct elements) is greater, the samples that Affirmative Sampling delivers are more representative than those produced by any scheme where the sample size is fixed a priori - hence its name. Our algorithm is straightforward to implement, and several implementations already exist.This work has been supported by funds from the MOTION Project (Project PID2020-112581GB-C21) of the Spanish Ministry of Science & Innovation MCIN/AEI/10.13039/501100011033, and by Princeton University, and its Department of Computer Science.Peer ReviewedPostprint (published version

    An evaluation of two new inference control methods

    Get PDF
    [[abstract]]An evaluation method is developed to measure the cost-effectiveness of two inference methods. The factors of the evaluation function consist of: preparation cost for the control method; query complexity; and security level under various attacks. The first control method is based on restriction, and the second on perturbation. Simulation results indicate that both methods have higher preparation cost, better security, and faster response time than L.H. Cox's method (1980) and L.L. Beck's method (1980). Finally, these two methods are compared to each other. In general, the control methods based on restriction have higher preparation cost and better security, and the control methods based on perturbation have fast response time for a query, but more information leak[[fileno]]2030204010032[[department]]資訊工程學

    Random sampling with a reservoir

    Full text link

    Ανίχνευση συμβάντων και δειγματοληψία δεξαμενής σε ασύρματα δίκτυα αισθητήρων

    Get PDF
    Στόχος της παρούσας πτυχιακής εργασίας είναι η ανάλυση των τιμών θερμοκρασίας που αποστέλλει ένα ασύρματο δίκτυο αισθητήρων, χρησιμοποιώντας δύο διαφορετικές υλοποιήσεις. Η πρώτη, αφορά στην εφαρμογή του αλγορίθμου Shewhart για κάθε απεσταλμένη τιμή, σε πραγματικό χρόνο, ενώ η δεύτερη αφορά στη συσσώρευση τιμών και την εξαγωγή δείγματος, μέσω ενός αλγορίθμου δειγματοληψίας δεξαμενής και έπειτα την τροφοδότηση του μεγίστου (ή μέσου όρου) αυτού του δείγματος, ως είσοδο στον αλγόριθμο Shewhart. Τέλος, η αποτελεσματικότητα των δύο υλοποιήσεων συγκρίνεται με τη βοήθεια της απόστασης Hamming.The main aim of this thesis is to analyze the temperature values sent over a wireless sensor network using two different implementations. The first one implements Shewhart algorithm processing every value sent in real time, whereas the second one, accumulates the values and exports a sample using a reservoir algorithm. From that sample, called reservoir, the max or the average value is extracted and then forwarded as input in the Shewhart algorithm. Finally, the efficiency of the two outputs is compared by means of Hamming distance

    Faster methods for random sampling

    Full text link

    Sampling Algorithms for Evolving Datasets

    Get PDF
    Perhaps the most flexible synopsis of a database is a uniform random sample of the data; such samples are widely used to speed up the processing of analytic queries and data-mining tasks, to enhance query optimization, and to facilitate information integration. Most of the existing work on database sampling focuses on how to create or exploit a random sample of a static database, that is, a database that does not change over time. The assumption of a static database, however, severely limits the applicability of these techniques in practice, where data is often not static but continuously evolving. In order to maintain the statistical validity of the sample, any changes to the database have to be appropriately reflected in the sample. In this thesis, we study efficient methods for incrementally maintaining a uniform random sample of the items in a dataset in the presence of an arbitrary sequence of insertions, updates, and deletions. We consider instances of the maintenance problem that arise when sampling from an evolving set, from an evolving multiset, from the distinct items in an evolving multiset, or from a sliding window over a data stream. Our algorithms completely avoid any accesses to the base data and can be several orders of magnitude faster than algorithms that do rely on such expensive accesses. The improved efficiency of our algorithms comes at virtually no cost: the resulting samples are provably uniform and only a small amount of auxiliary information is associated with the sample. We show that the auxiliary information not only facilitates efficient maintenance, but it can also be exploited to derive unbiased, low-variance estimators for counts, sums, averages, and the number of distinct items in the underlying dataset. In addition to sample maintenance, we discuss methods that greatly improve the flexibility of random sampling from a system's point of view. More specifically, we initiate the study of algorithms that resize a random sample upwards or downwards. Our resizing algorithms can be exploited to dynamically control the size of the sample when the dataset grows or shrinks; they facilitate resource management and help to avoid under- or oversized samples. Furthermore, in large-scale databases with data being distributed across several remote locations, it is usually infeasible to reconstruct the entire dataset for the purpose of sampling. To address this problem, we provide efficient algorithms that directly combine the local samples maintained at each location into a sample of the global dataset. We also consider a more general problem, where the global dataset is defined as an arbitrary set or multiset expression involving the local datasets, and provide efficient solutions based on hashing
    corecore