7 research outputs found

    Missing data imputation through generative topographic mapping as a mixture of t-distributions: Theoretical developments

    Get PDF
    The Generative Topographic Mapping (GTM) was originally conceived as a probabilistic alternative to the well-known, neural network-inspired, Self-Organizing Map (SOM). The GTM can also be interpreted as a constrained mixture of distributions model. In recent years, much attention has been directed towards Student t-distributions as an alternative to Gaussians in mixture models due to their robustness towards outliers. In this report, the GTM is redefined as a constrained mixture of t-distributions: the t-GTM, and the Expectation-Maximization algorithm that is used to fit the model to the data is modified to provide missing data imputation.Postprint (published version

    Multiple Resolution Nonparametric Classifiers

    Get PDF
    Bayesian discriminant functions provide optimal classification decision boundaries in the sense of minimizing the average error rate. An operational assumption is that the probability density functions for the individual classes are either known a priori or can be estimated from the data through the use of estimating techniques. The use of Parzen- windows is a popular and theoretically sound choice for such estimation. However, while the minimal average error rate can be achieved when combining Bayes Rule with Parzen-window density estimation, the latter is computationally costly to the point where it may lead to unacceptable run-time performance. We present the Multiple Resolution Nonparametric (MRN) classifier as a new approach for significantly reducing the computational cost of using Parzen-window density estimates without sacrificing the virtues of Bayesian discriminant functions. Performance is evaluated against a standard Parzen-window classifier on several common datasets

    Connectionist-Symbolic Machine Intelligence using Cellular Automata based Reservoir-Hyperdimensional Computing

    Full text link
    We introduce a novel framework of reservoir computing, that is capable of both connectionist machine intelligence and symbolic computation. Cellular automaton is used as the reservoir of dynamical systems. Input is randomly projected onto the initial conditions of automaton cells and nonlinear computation is performed on the input via application of a rule in the automaton for a period of time. The evolution of the automaton creates a space-time volume of the automaton state space, and it is used as the reservoir. The proposed framework is capable of long short-term memory and it requires orders of magnitude less computation compared to Echo State Networks. We prove that cellular automaton reservoir holds a distributed representation of attribute statistics, which provides a more effective computation than local representation. It is possible to estimate the kernel for linear cellular automata via metric learning, that enables a much more efficient distance computation in support vector machine framework. Also, binary reservoir feature vectors can be combined using Boolean operations as in hyperdimensional computing, paving a direct way for concept building and symbolic processing.Comment: Corrected Typos. Responded some comments on section 8. Added appendix for details. Recurrent architecture emphasize

    Stochastic model reduction for robust dynamical characterization of structures with random parameters

    Get PDF
    International audienceIn this paper, we characterize random eigenspaces with a non-intrusive method based on the decoupling of random eigenvalues from their corresponding random eigenvectors. This method allows us to estimate the first statistical moments of the random eigenvalues of the system with a reduced number of deterministic finite element computations. The originality of this work is to adapt the method used to estimate each random eigenvalue depending on a global accuracy requirement. This allows us to ensure a minimal computational cost. The stochastic model of the structure is thus reduced by exploiting specific properties of random eigenvectors associated with the random eigenfrequencies being sought. An indicator with no additional computation cost is proposed to identify when the method needs to be enhanced. Finally, a simple three-beam frame and an industrial structure illustrate the proposed approach

    Flexible and Robust Bayesian Classification by Finite Mixture Models

    No full text
    The regularized Mahalanobis distance is proposed in the framework of nite mixture models to avoid commonly faced numerical difficulties encountered with EM. Its principle is applied to Gaussian and Student-t mixtures, resulting in reliable density estimates, the model complexity being kept low. Besides, the regularized models are robust to various noise types. Finally, it is shown that the quality of the associated Bayesian classication is near optimal on Ripley's synthetic data set

    Flexible and Robust Bayesian Classification by Finite Mixture Models

    No full text
    The regularized Mahalanobis distance is proposed in the framework of finite mixture models to avoid commonly faced numerical difficulties encountered with EM. Its principle is applied to Gaussian and Student-t mixtures, resulting in reliable density estimates, the model complexity being kept low. Besides, the regularized models are robust to various noise types. Finally, it is shown..
    corecore