45 research outputs found

    Classification using distance nearest neighbours

    Get PDF
    This paper proposes a new probabilistic classification algorithm using a Markov random field approach. The joint distribution of class labels is explicitly modelled using the distances between feature vectors. Intuitively, a class label should depend more on class labels which are closer in the feature space, than those which are further away. Our approach builds on previous work by Holmes and Adams (2002, 2003) and Cucala et al. (2008). Our work shares many of the advantages of these approaches in providing a probabilistic basis for the statistical inference. In comparison to previous work, we present a more efficient computational algorithm to overcome the intractability of the Markov random field model. The results of our algorithm are encouraging in comparison to the k-nearest neighbour algorithm.Comment: 12 pages, 2 figures. To appear in Statistics and Computin

    Image labeling and grouping by minimizing linear functionals over cones

    Full text link
    We consider energy minimization problems related to image labeling, partitioning, and grouping, which typically show up at mid-level stages of computer vision systems. A common feature of these problems is their intrinsic combinatorial complexity from an optimization pointof-view. Rather than trying to compute the global minimum - a goal we consider as elusive in these cases - we wish to design optimization approaches which exhibit two relevant properties: First, in each application a solution with guaranteed degree of suboptimality can be computed. Secondly, the computations are based on clearly defined algorithms which do not comprise any (hidden) tuning parameters. In this paper, we focus on the second property and introduce a novel and general optimization technique to the field of computer vision which amounts to compute a sub optimal solution by just solving a convex optimization problem. As representative examples, we consider two binary quadratic energy functionals related to image labeling and perceptual grouping. Both problems can be considered as instances of a general quadratic functional in binary variables, which is embedded into a higher-dimensional space such that sub optimal solutions can be computed as minima of linear functionals over cones in that space (semidefinite programs). Extensive numerical results reveal that, on the average, sub optimal solutions can be computed which yield a gap below 5% with respect to the global optimum in case where this is known

    Some Aspects of Latent Structure Analysis

    Get PDF
    Latent structure models involve real, potentially observable variables and latent, unobservable variables. The framework includes various particular types of model, such as factor analysis, latent class analysis, latent trait analysis, latent profile models, mixtures of factor analysers, state-space models and others. The simplest scenario, of a single discrete latent variable, includes finite mixture models, hidden Markov chain models and hidden Markov random field models. The paper gives a brief tutorial of the application of maximum likelihood and Bayesian approaches to the estimation of parameters within these models, emphasising especially the fact that computational complexity varies greatly among the different scenarios. In the case of a single discrete latent variable, the issue of assessing its cardinality is discussed. Techniques such as the EM algorithm, Markov chain Monte Carlo methods and variational approximations are mentioned

    L-function of geographical flows

    No full text
    corecore