60 research outputs found

    THE EXPECTED CONVEX HULL TRIMMED REGIONS OF A SAMPLE

    Get PDF
    Given a data set in the multivariate Euclidean space, we study regions of central points built by averaging all their subsets with a fixed number of elements. The averaging of these sets is performed by appropriately scaling the Minkowski or elementwise summation of their convex hulls. The volume of such central regions is proposed as a multivariate scatter estimate and a circular sequence algorithm to compute the central regions of a bivariate data set is described.

    The expected convex hull trimmed regions of a sample

    Get PDF
    Given a data set in the multivariate Euclidean space, we study regions of central points built by averaging all their subsets with a fixed number of elements. The averaging of these sets is performed by appropriately scaling the Minkowski or elementwise summation of their convex hulls. The volume of such central regions is proposed as a multivariate scatter estimate and a circular sequence algorithm to compute the central regions of a bivariate data set is described

    Fast implementation of the Tukey depth

    Full text link
    Tukey depth function is one of the most famous multivariate tools serving robust purposes. It is also very well known for its computability problems in dimensions p3p \ge 3. In this paper, we address this computing issue by presenting two combinatorial algorithms. The first is naive and calculates the Tukey depth of a single point with complexity O(np1log(n))O\left(n^{p-1}\log(n)\right), while the second further utilizes the quasiconcave of the Tukey depth function and hence is more efficient than the first. Both require very minimal memory and run much faster than the existing ones. All experiments indicate that they compute the exact Tukey depth.Comment: 16 pages, 13 figure

    An exact algorithm for weighted-mean trimmed regions in any dimension

    Get PDF
    Trimmed regions are a powerful tool of multivariate data analysis. They describe a probability distribution in Euclidean d-space regarding location, dispersion, and shape, and they order multivariate data with respect to their centrality. Dyckerhoff and Mosler (201x) have introduced the class of weighted-mean trimmed regions, which possess attractive properties regarding continuity, subadditivity, and monotonicity. We present an exact algorithm to compute the weighted-mean trimmed regions of a given data cloud in arbitrary dimension d. These trimmed regions are convex polytopes in Rd. To calculate them, the algorithm builds on methods from computational geometry. A characterization of a region's facets is used, and information about the adjacency of the facets is extracted from the data. A key problem consists in ordering the facets. It is solved by the introduction of a tree-based order. The algorithm has been programmed in C++ and is available as an R package. --central regions,data depth,multivariate data analysis,convex polytope,computational geometry,algorithm,C++, R

    Depth and Depth-Based Classification with R Package ddalpha

    Get PDF
    Following the seminal idea of Tukey (1975), data depth is a function that measures how close an arbitrary point of the space is located to an implicitly defined center of a data cloud. Having undergone theoretical and computational developments, it is now employed in numerous applications with classification being the most popular one. The R package ddalpha is a software directed to fuse experience of the applicant with recent achievements in the area of data depth and depth-based classification. ddalpha provides an implementation for exact and approximate computation of most reasonable and widely applied notions of data depth. These can be further used in the depth-based multivariate and functional classifiers implemented in the package, where the DDα-procedure is in the main focus. The package is expandable with user-defined custom depth methods and separators. The implemented functions for depth visualization and the built-in benchmark procedures may also serve to provide insights into the geometry of the data and the quality of pattern recognition

    An optimal randomized algorithm for d-variate zonoid depth

    Get PDF
    AbstractA randomized linear expected-time algorithm for computing the zonoid depth [R. Dyckerhoff, G. Koshevoy, K. Mosler, Zonoid data depth: Theory and computation, in: A. Prat (Ed.), COMPSTAT 1996—Proceedings in Computational Statistics, Physica-Verlag, Heidelberg, 1996, pp. 235–240; K. Mosler, Multivariate Dispersion, Central Regions and Depth. The Lift Zonoid Approach, Lecture Notes in Statistics, vol. 165, Springer-Verlag, New York, 2002] of a point with respect to a fixed dimensional point set is presented
    corecore