41,084 research outputs found
Optimal Clustering under Uncertainty
Classical clustering algorithms typically either lack an underlying
probability framework to make them predictive or focus on parameter estimation
rather than defining and minimizing a notion of error. Recent work addresses
these issues by developing a probabilistic framework based on the theory of
random labeled point processes and characterizing a Bayes clusterer that
minimizes the number of misclustered points. The Bayes clusterer is analogous
to the Bayes classifier. Whereas determining a Bayes classifier requires full
knowledge of the feature-label distribution, deriving a Bayes clusterer
requires full knowledge of the point process. When uncertain of the point
process, one would like to find a robust clusterer that is optimal over the
uncertainty, just as one may find optimal robust classifiers with uncertain
feature-label distributions. Herein, we derive an optimal robust clusterer by
first finding an effective random point process that incorporates all
randomness within its own probabilistic structure and from which a Bayes
clusterer can be derived that provides an optimal robust clusterer relative to
the uncertainty. This is analogous to the use of effective class-conditional
distributions in robust classification. After evaluating the performance of
robust clusterers in synthetic mixtures of Gaussians models, we apply the
framework to granular imaging, where we make use of the asymptotic
granulometric moment theory for granular images to relate robust clustering
theory to the application.Comment: 19 pages, 5 eps figures, 1 tabl
Robust Gaussian Filtering using a Pseudo Measurement
Many sensors, such as range, sonar, radar, GPS and visual devices, produce
measurements which are contaminated by outliers. This problem can be addressed
by using fat-tailed sensor models, which account for the possibility of
outliers. Unfortunately, all estimation algorithms belonging to the family of
Gaussian filters (such as the widely-used extended Kalman filter and unscented
Kalman filter) are inherently incompatible with such fat-tailed sensor models.
The contribution of this paper is to show that any Gaussian filter can be made
compatible with fat-tailed sensor models by applying one simple change: Instead
of filtering with the physical measurement, we propose to filter with a pseudo
measurement obtained by applying a feature function to the physical
measurement. We derive such a feature function which is optimal under some
conditions. Simulation results show that the proposed method can effectively
handle measurement outliers and allows for robust filtering in both linear and
nonlinear systems
Quasi-Optimal Filtering in Inverse Problems
A way of constructing a nonlinear filter close to the optimal Kolmogorov -
Wiener filter is proposed within the framework of the statistical approach to
inverse problems. Quasi-optimal filtering, which has no Bayesian assumptions,
produces stable and efficient solutions by relying solely on the internal
resources of the inverse theory. The exact representation is given of the
Feasible Region for inverse solutions that follows from the statistical
consideration.Comment: 9 pages, 240 K
- …