3,865 research outputs found
Methods of Hierarchical Clustering
We survey agglomerative hierarchical clustering algorithms and discuss
efficient implementations that are available in R and other software
environments. We look at hierarchical self-organizing maps, and mixture models.
We review grid-based clustering, focusing on hierarchical density-based
approaches. Finally we describe a recently developed very efficient (linear
time) hierarchical clustering algorithm, which can also be viewed as a
hierarchical grid-based algorithm.Comment: 21 pages, 2 figures, 1 table, 69 reference
Fast Hierarchical Clustering and Other Applications of Dynamic Closest Pairs
We develop data structures for dynamic closest pair problems with arbitrary
distance functions, that do not necessarily come from any geometric structure
on the objects. Based on a technique previously used by the author for
Euclidean closest pairs, we show how to insert and delete objects from an
n-object set, maintaining the closest pair, in O(n log^2 n) time per update and
O(n) space. With quadratic space, we can instead use a quadtree-like structure
to achieve an optimal time bound, O(n) per update. We apply these data
structures to hierarchical clustering, greedy matching, and TSP heuristics, and
discuss other potential applications in machine learning, Groebner bases, and
local improvement algorithms for partition and placement problems. Experiments
show our new methods to be faster in practice than previously used heuristics.Comment: 20 pages, 9 figures. A preliminary version of this paper appeared at
the 9th ACM-SIAM Symp. on Discrete Algorithms, San Francisco, 1998, pp.
619-628. For source code and experimental results, see
http://www.ics.uci.edu/~eppstein/projects/pairs
Reverse engineering of CAD models via clustering and approximate implicitization
In applications like computer aided design, geometric models are often
represented numerically as polynomial splines or NURBS, even when they
originate from primitive geometry. For purposes such as redesign and
isogeometric analysis, it is of interest to extract information about the
underlying geometry through reverse engineering. In this work we develop a
novel method to determine these primitive shapes by combining clustering
analysis with approximate implicitization. The proposed method is automatic and
can recover algebraic hypersurfaces of any degree in any dimension. In exact
arithmetic, the algorithm returns exact results. All the required parameters,
such as the implicit degree of the patches and the number of clusters of the
model, are inferred using numerical approaches in order to obtain an algorithm
that requires as little manual input as possible. The effectiveness, efficiency
and robustness of the method are shown both in a theoretical analysis and in
numerical examples implemented in Python
Ward's Hierarchical Clustering Method: Clustering Criterion and Agglomerative Algorithm
The Ward error sum of squares hierarchical clustering method has been very
widely used since its first description by Ward in a 1963 publication. It has
also been generalized in various ways. However there are different
interpretations in the literature and there are different implementations of
the Ward agglomerative algorithm in commonly used software systems, including
differing expressions of the agglomerative criterion. Our survey work and case
studies will be useful for all those involved in developing software for data
analysis using Ward's hierarchical clustering method.Comment: 20 pages, 21 citations, 4 figure
- …