6 research outputs found
Adaptive Hybrid Blood Cell Image Segmentation
Image segmentation is an important phase in the image recognition system. In medical imaging such as blood cell analysis, it becomes a crucial step in quantitative cytophotometry. Currently, blood cell images become predominantly valuable in medical diagnostics tools. In this paper, we present an adaptive hybrid analysis based on selected segmentation algorithms. Three designates common approaches, that are Fuzzy c-means, K-means and Mean-shift are adapted. Blood cell images that are infected with malaria parasites at various stages were tested. The most suitable method will be selected based on the lowest number of regions. The selected approach will be enhanced by applying Median-cut algorithm to further expand the segmentation process. The proposed adaptive hybrid method has shown a significant improvement in the number of region
A Bibliographic View on Constrained Clustering
A keyword search on constrained clustering on Web-of-Science returned just
under 3,000 documents. We ran automatic analyses of those, and compiled our own
bibliography of 183 papers which we analysed in more detail based on their
topic and experimental study, if any. This paper presents general trends of the
area and its sub-topics by Pareto analysis, using citation count and year of
publication. We list available software and analyse the experimental sections
of our reference collection. We found a notable lack of large comparison
experiments. Among the topics we reviewed, applications studies were most
abundant recently, alongside deep learning, active learning and ensemble
learning.Comment: 18 pages, 11 figures, 177 reference
Kernel Methods for Weakly Supervised Mean Shift Clustering
Mean shift clustering is a powerful unsupervised data analysis technique which does not require prior knowledge of the number of clusters, and does not constrain the shape of the clusters. The data association criteria is based on the underlying probability distribution of the data points which is defined in advance via the employed distance metric. In many problem domains, the initially designed distance metric fails to resolve the ambiguities in the clustering process. We present a novel semi-supervised kernel mean shift algorithm where the inherent structure of the data points is learned with a few user supplied constraints in addition to the original metric. The constraints we consider are the pairs of points that should be clustered together. The data points are implicitly mapped to a higher dimensional space induced by the kernel function where the constraints can be effectively enforced. The mode seeking is then performed on the embedded space and the approach preserves all the advantages of the original mean shift algorithm. Experiments on challenging synthetic and real data clearly demonstrate that significant improvements in clustering accuracy can be achieved by employing only a few constraints. 1
Kernel Methods for Weakly Supervised Mean Shift Clustering
Mean shift clustering is a powerful unsupervised data analysis technique which does not require prior knowledge of the number of clusters, and does not constrain the shape of the clusters. The data association criteria is based on the underlying probability distribution of the data points which is defined in advance via the employed distance metric. In many problem domains, the initially designed distance metric fails to resolve the ambiguities in the clustering process. We present a novel semi-supervised kernel mean shift algorithm where the inherent structure of the data points is learned with a few user supplied constraints in addition to the original metric. The constraints we consider are the pairs of points that should be clustered together. The data points are implicitly mapped to a higher dimensional space induced by the kernel function where the constraints can be effectively enforced. The mode seeking is then performed on the embedded space and the approach preserves all the advantages of the original mean shift algorithm. Experiments on challenging synthetic and real data clearly demonstrate that significant improvements in clustering accuracy can be achieved by employing only a few constraints