14,273 research outputs found
Latent Fisher Discriminant Analysis
Linear Discriminant Analysis (LDA) is a well-known method for dimensionality
reduction and classification. Previous studies have also extended the
binary-class case into multi-classes. However, many applications, such as
object detection and keyframe extraction cannot provide consistent
instance-label pairs, while LDA requires labels on instance level for training.
Thus it cannot be directly applied for semi-supervised classification problem.
In this paper, we overcome this limitation and propose a latent variable Fisher
discriminant analysis model. We relax the instance-level labeling into
bag-level, is a kind of semi-supervised (video-level labels of event type are
required for semantic frame extraction) and incorporates a data-driven prior
over the latent variables. Hence, our method combines the latent variable
inference and dimension reduction in an unified bayesian framework. We test our
method on MUSK and Corel data sets and yield competitive results compared to
the baseline approach. We also demonstrate its capacity on the challenging
TRECVID MED11 dataset for semantic keyframe extraction and conduct a
human-factors ranking-based experimental evaluation, which clearly demonstrates
our proposed method consistently extracts more semantically meaningful
keyframes than challenging baselines.Comment: 12 page
Mixtures of Skew-t Factor Analyzers
In this paper, we introduce a mixture of skew-t factor analyzers as well as a
family of mixture models based thereon. The mixture of skew-t distributions
model that we use arises as a limiting case of the mixture of generalized
hyperbolic distributions. Like their Gaussian and t-distribution analogues, our
mixture of skew-t factor analyzers are very well-suited to the model-based
clustering of high-dimensional data. Imposing constraints on components of the
decomposed covariance parameter results in the development of eight flexible
models. The alternating expectation-conditional maximization algorithm is used
for model parameter estimation and the Bayesian information criterion is used
for model selection. The models are applied to both real and simulated data,
giving superior clustering results compared to a well-established family of
Gaussian mixture models
- …