9 research outputs found

    An Efficient Algorithm for Computing Entropic Measures of Feature Subsets

    No full text
    International audienceEntropic measures such as conditional entropy or mutual information have been used numerous times in pattern mining, for instance to characterize valuable itemsets or approximate functional dependencies. Strangely enough the fundamental problem of designing efficient algorithms to compute entropy of subsets of features (or mutual information of feature subsets relatively to some target feature) has received little attention compared to the analog problem of computing frequency of itemsets. The present article proposes to fill this gap: it introduces a fast and scalable method that computes entropy and mutual information for a large number of feature subsets by adopting the divide and conquer strategy used by FP-growth-one of the most efficient frequent itemset mining algorithm. In order to illustrate its practical interest, the algorithm is then used to solve the recently introduced problem of mining reliable approximate functional dependencies. It finally provides empirical evidences that in the context of non-redundant pattern extraction, the proposed method outperforms existing algorithms for both speed and scalability

    Discovering Approximate Functional Dependencies using Smoothed Mutual Information

    No full text

    On the Calibration of Epistemic Uncertainty: Principles, Paradoxes and Conflictual Loss

    No full text
    International audienceThe calibration of predictive distributions has been widely studied in deep learning, but the same cannot be said about the more specific epistemic uncertainty as produced by Deep Ensembles, Bayesian Deep Networks, or Evidential Deep Networks. Although measurable, this form of uncertainty is difficult to calibrate on an objective basis as it depends on the prior for which a variety of choices exist. Nevertheless, epistemic uncertainty must in all cases satisfy two formal requirements: firstly, it must decrease when the training dataset gets larger and, secondly, it must increase when the model expressiveness grows. Despite these expectations, our experimental study shows that on several reference datasets and models, measures of epistemic uncertainty violate these requirements, sometimes presenting trends completely opposite to those expected. These paradoxes between expectation and reality raise the question of the true utility of epistemic uncertainty as estimated by these models. A formal argument suggests that this disagreement is due to a poor approximation of the posterior distribution rather than to a flaw in the measure itself. Based on this observation, we propose a regularization function for deep ensembles, called conflictual loss in line with the above requirements. We emphasize its strengths by showing experimentally that it fulfills both requirements of epistemic uncertainty, without sacrificing either the performance nor the calibration of the deep ensembles

    The ATLAS DAQ and Event Filter prototype "-1" project

    No full text
    A project has been approved by the ATLAS Collaboration for the design and implementation of a Data Acquisition and Event Filter prototype, based on the functional architecture described in the ATLAS Technical Proposal. The prototype consists of a full "vertical" slice of the ATLAS Data Acquisition and Event Filter architecture, including all the hardware and software elements of the data flow, its control and monitoring as well as all the elements of a complete on-line system. This paper outlines the project, its goals, structure, schedule and current status and describes details of the system architecture and its components. (C) 1998 Elsevier Science B.V
    corecore