270,059 research outputs found

    Sparse Hopsets in Congested Clique

    Get PDF
    We give the first Congested Clique algorithm that computes a sparse hopset with polylogarithmic hopbound in polylogarithmic time. Given a graph G=(V,E)G=(V,E), a (β,ϵ)(\beta,\epsilon)-hopset HH with "hopbound" β\beta, is a set of edges added to GG such that for any pair of nodes uu and vv in GG there is a path with at most β\beta hops in GHG \cup H with length within (1+ϵ)(1+\epsilon) of the shortest path between uu and vv in GG. Our hopsets are significantly sparser than the recent construction of Censor-Hillel et al. [6], that constructs a hopset of size O~(n3/2)\tilde{O}(n^{3/2}), but with a smaller polylogarithmic hopbound. On the other hand, the previously known constructions of sparse hopsets with polylogarithmic hopbound in the Congested Clique model, proposed by Elkin and Neiman [10],[11],[12], all require polynomial rounds. One tool that we use is an efficient algorithm that constructs an \ell-limited neighborhood cover, that may be of independent interest. Finally, as a side result, we also give a hopset construction in a variant of the low-memory Massively Parallel Computation model, with improved running time over existing algorithms

    Search algorithms for regression test case prioritization

    Get PDF
    Regression testing is an expensive, but important, process. Unfortunately, there may be insufficient resources to allow for the re-execution of all test cases during regression testing. In this situation, test case prioritisation techniques aim to improve the effectiveness of regression testing, by ordering the test cases so that the most beneficial are executed first. Previous work on regression test case prioritisation has focused on Greedy Algorithms. However, it is known that these algorithms may produce sub-optimal results, because they may construct results that denote only local minima within the search space. By contrast, meta-heuristic and evolutionary search algorithms aim to avoid such problems. This paper presents results from an empirical study of the application of several greedy, meta-heuristic and evolutionary search algorithms to six programs, ranging from 374 to 11,148 lines of code for 3 choices of fitness metric. The paper addresses the problems of choice of fitness metric, characterisation of landscape modality and determination of the most suitable search technique to apply. The empirical results replicate previous results concerning Greedy Algorithms. They shed light on the nature of the regression testing search space, indicating that it is multi-modal. The results also show that Genetic Algorithms perform well, although Greedy approaches are surprisingly effective, given the multi-modal nature of the landscape

    Discovering Reliable Dependencies from Data: Hardness and Improved Algorithms

    Get PDF
    The reliable fraction of information is an attractive score for quantifying (functional) dependencies in high-dimensional data. In this paper, we systematically explore the algorithmic implications of using this measure for optimization. We show that the problem is NP-hard, which justifies the usage of worst-case exponential-time as well as heuristic search methods. We then substantially improve the practical performance for both optimization styles by deriving a novel admissible bounding function that has an unbounded potential for additional pruning over the previously proposed one. Finally, we empirically investigate the approximation ratio of the greedy algorithm and show that it produces highly competitive results in a fraction of time needed for complete branch-and-bound style search.Comment: Accepted to Proceedings of the IEEE International Conference on Data Mining (ICDM'18

    An effective simulation model to predict and optimize the performance of single and double glaze flat-plate solar collector designs

    Get PDF
    This paper outlines and formulates a compact and effective simulation model, which predicts the performance of single and double glaze flat-plate collector. The model uses an elaborated iterative simulation algorithm and provides the collector top losses, the glass covers temperatures, the collector absorber temperature, the collector fluid outlet temperature, the system efficiency, and the thermal gain for any operational and environmental conditions. It is a numerical approach based on simultaneous guesses for the three temperatures, Tp plate collector temperature and the temperatures of the two glass covers Tg1, Tg2. A set of energy balance equations is developed which allows for structured iteration modes whose results converge very fast and provide the values of any quantity which concerns the steady state performance profile of any flat-plate collector design. Comparison of the results obtained by this model for flat-plate collectors, single or double glaze, with those obtained by using the Klein formula, as well as the results provided by other researchers, is presented
    corecore