9,410 research outputs found

    A Linear-time Algorithm for Sparsification of Unweighted Graphs

    Full text link
    Given an undirected graph GG and an error parameter Ļµ>0\epsilon > 0, the {\em graph sparsification} problem requires sampling edges in GG and giving the sampled edges appropriate weights to obtain a sparse graph GĻµG_{\epsilon} with the following property: the weight of every cut in GĻµG_{\epsilon} is within a factor of (1Ā±Ļµ)(1\pm \epsilon) of the weight of the corresponding cut in GG. If GG is unweighted, an O(mlogā”n)O(m\log n)-time algorithm for constructing GĻµG_{\epsilon} with O(nlogā”n/Ļµ2)O(n\log n/\epsilon^2) edges in expectation, and an O(m)O(m)-time algorithm for constructing GĻµG_{\epsilon} with O(nlogā”2n/Ļµ2)O(n\log^2 n/\epsilon^2) edges in expectation have recently been developed (Hariharan-Panigrahi, 2010). In this paper, we improve these results by giving an O(m)O(m)-time algorithm for constructing GĻµG_{\epsilon} with O(nlogā”n/Ļµ2)O(n\log n/\epsilon^2) edges in expectation, for unweighted graphs. Our algorithm is optimal in terms of its time complexity; further, no efficient algorithm is known for constructing a sparser GĻµG_{\epsilon}. Our algorithm is Monte-Carlo, i.e. it produces the correct output with high probability, as are all efficient graph sparsification algorithms

    Almost-Smooth Histograms and Sliding-Window Graph Algorithms

    Full text link
    We study algorithms for the sliding-window model, an important variant of the data-stream model, in which the goal is to compute some function of a fixed-length suffix of the stream. We extend the smooth-histogram framework of Braverman and Ostrovsky (FOCS 2007) to almost-smooth functions, which includes all subadditive functions. Specifically, we show that if a subadditive function can be (1+Ļµ)(1+\epsilon)-approximated in the insertion-only streaming model, then it can be (2+Ļµ)(2+\epsilon)-approximated also in the sliding-window model with space complexity larger by factor O(Ļµāˆ’1logā”w)O(\epsilon^{-1}\log w), where ww is the window size. We demonstrate how our framework yields new approximation algorithms with relatively little effort for a variety of problems that do not admit the smooth-histogram technique. For example, in the frequency-vector model, a symmetric norm is subadditive and thus we obtain a sliding-window (2+Ļµ)(2+\epsilon)-approximation algorithm for it. Another example is for streaming matrices, where we derive a new sliding-window (2+Ļµ)(\sqrt{2}+\epsilon)-approximation algorithm for Schatten 44-norm. We then consider graph streams and show that many graph problems are subadditive, including maximum submodular matching, minimum vertex-cover, and maximum kk-cover, thereby deriving sliding-window O(1)O(1)-approximation algorithms for them almost for free (using known insertion-only algorithms). Finally, we design for every dāˆˆ(1,2]d\in (1,2] an artificial function, based on the maximum-matching size, whose almost-smoothness parameter is exactly dd

    Adaptive random forests for evolving data stream classification

    Get PDF
    Random forests is currently one of the most used machine learning algorithms in the non-streaming (batch) setting. This preference is attributable to its high learning performance and low demands with respect to input preparation and hyper-parameter tuning. However, in the challenging context of evolving data streams, there is no random forests algorithm that can be considered state-of-the-art in comparison to bagging and boosting based algorithms. In this work, we present the adaptive random forest (ARF) algorithm for classification of evolving data streams. In contrast to previous attempts of replicating random forests for data stream learning, ARF includes an effective resampling method and adaptive operators that can cope with different types of concept drifts without complex optimizations for different data sets. We present experiments with a parallel implementation of ARF which has no degradation in terms of classification performance in comparison to a serial implementation, since trees and adaptive operators are independent from one another. Finally, we compare ARF with state-of-the-art algorithms in a traditional test-then-train evaluation and a novel delayed labelling evaluation, and show that ARF is accurate and uses a feasible amount of resources
    • ā€¦
    corecore