4,992 research outputs found

    The Design and Implementation of a Bayesian CAD Modeler for Robotic Applications

    Get PDF
    We present a Bayesian CAD modeler for robotic applications. We address the problem of taking into account the propagation of geometric uncertainties when solving inverse geometric problems. The proposed method may be seen as a generalization of constraint-based approaches in which we explicitly model geometric uncertainties. Using our methodology, a geometric constraint is expressed as a probability distribution on the system parameters and the sensor measurements, instead of a simple equality or inequality. To solve geometric problems in this framework, we propose an original resolution method able to adapt to problem complexity. Using two examples, we show how to apply our approach by providing simulation results using our modeler

    A Robotic CAD System using a Bayesian Framework

    Get PDF
    We present in this paper a Bayesian CAD system for robotic applications. We address the problem of the propagation of geometric uncertainties and how esian CAD system for robotic applications. We address the problem of the propagation of geometric uncertainties and how to take this propagation into account when solving inverse problems. We describe the methodology we use to represent and handle uncertainties using probability distributions on the system's parameters and sensor measurements. It may be seen as a generalization of constraint-based approaches where we express a constraint as a probability distribution instead of a simple equality or inequality. Appropriate numerical algorithms used to apply this methodology are also described. Using an example, we show how to apply our approach by providing simulation results using our CAD system

    Parameter Solving of Probability Integral Method Based on Improved Genetic Algorithm

    Get PDF
    The probability integral method (PIM) is the main method for mining subsidence prediction in China. Parameter errors and model errors are the main sources of error in the application of the probability integral method. There are many surface subsidence problems caused by coal mining. In order to improve the accuracy and operating efficiency of the genetic algorithm (GA) in calculating the parameters of the PIM, this paper proposes an improved genetic algorithm (IGA) by adding the dynamic crossover and mutation rate to the traditional GA. Made improvements to the shortcomings of random crossover and mutation rate of all individuals in the population in the original algorithm.Through simulation experiments, it is confirmed that the IGA improves the calculation efficiency and accuracy of the traditional GA under the same conditions.The IGA has higher accuracy, reliability, resistance to gross interference and resistance to missing observation points. This method is obviously superior to direct inversion and conventional optimization inversion algorithms, and effectively avoids the dependence on the initial value of the probabilistic integral method parameter

    A robust statistical estimation of the basic parameters of single stellar populations. I. Method

    Full text link
    The colour-magnitude diagrams of resolved single stellar populations, such as open and globular clusters, have provided the best natural laboratories to test stellar evolution theory. Whilst a variety of techniques have been used to infer the basic properties of these simple populations, systematic uncertainties arise from the purely geometrical degeneracy produced by the similar shape of isochrones of different ages and metallicities. Here we present an objective and robust statistical technique which lifts this degeneracy to a great extent through the use of a key observable: the number of stars along the isochrone. Through extensive Monte Carlo simulations we show that, for instance, we can infer the four main parameters (age, metallicity, distance and reddening) in an objective way, along with robust confidence intervals and their full covariance matrix. We show that systematic uncertainties due to field contamination, unresolved binaries, initial or present-day stellar mass function are either negligible or well under control. This technique provides, for the first time, a proper way to infer with unprecedented accuracy the fundamental properties of simple stellar populations, in an easy-to-implement algorithm.Comment: 17 pages, 12 figures, MNRAS, in pres

    Solving, Estimating and Selecting Nonlinear Dynamic Economic Models without the Curse of Dimensionality

    Get PDF
    A welfare analysis of a risky policy is impossible within a linear or linearized model and its certainty equivalence property. The presented algorithms are designed as a toolbox for a general model class. The computational challenges are considerable and I concentrate on the numerics and statistics for a simple model of dynamic consumption and labor choice. I calculate the optimal policy and estimate the posterior density of structural parameters and the marginal likelihood within a nonlinear state space model. My approach is even in an interpreted language twenty time faster than the only alternative compiled approach. The model is estimated on simulated data in order to test the routines against known true parameters. The policy function is approximated by Smolyak Chebyshev polynomials and the rational expectation integral by Smolyak Gaussian quadrature. The Smolyak operator is used to extend univariate approximation and integration operators to many dimensions. It reduces the curse of dimensionality from exponential to polynomial growth. The likelihood integrals are evaluated by a Gaussian quadrature and Gaussian quadrature particle filter. The bootstrap or sequential importance resampling particle filter is used as an accuracy benchmark. The posterior is estimated by the Gaussian filter and a Metropolis- Hastings algorithm. I propose a genetic extension of the standard Metropolis-Hastings algorithm by parallel random walk sequences. This improves the robustness of start values and the global maximization properties. Moreover it simplifies a cluster implementation and the random walk variances decision is reduced to only two parameters so that almost no trial sequences are needed. Finally the marginal likelihood is calculated as a criterion for nonnested and quasi-true models in order to select between the nonlinear estimates and a first order perturbation solution combined with the Kalman filter.stochastic dynamic general equilibrium model, Chebyshev polynomials, Smolyak operator, nonlinear state space filter, Curse of Dimensionality, posterior of structural parameters, marginal likelihood

    Multi-scale uncertainty quantification in geostatistical seismic inversion

    Full text link
    Geostatistical seismic inversion is commonly used to infer the spatial distribution of the subsurface petro-elastic properties by perturbing the model parameter space through iterative stochastic sequential simulations/co-simulations. The spatial uncertainty of the inferred petro-elastic properties is represented with the updated a posteriori variance from an ensemble of the simulated realizations. Within this setting, the large-scale geological (metaparameters) used to generate the petro-elastic realizations, such as the spatial correlation model and the global a priori distribution of the properties of interest, are assumed to be known and stationary for the entire inversion domain. This assumption leads to underestimation of the uncertainty associated with the inverted models. We propose a practical framework to quantify uncertainty of the large-scale geological parameters in seismic inversion. The framework couples geostatistical seismic inversion with a stochastic adaptive sampling and Bayesian inference of the metaparameters to provide a more accurate and realistic prediction of uncertainty not restricted by heavy assumptions on large-scale geological parameters. The proposed framework is illustrated with both synthetic and real case studies. The results show the ability retrieve more reliable acoustic impedance models with a more adequate uncertainty spread when compared with conventional geostatistical seismic inversion techniques. The proposed approach separately account for geological uncertainty at large-scale (metaparameters) and local scale (trace-by-trace inversion)

    Neural network determination of parton distributions: the nonsinglet case

    Get PDF
    We provide a determination of the isotriplet quark distribution from available deep--inelastic data using neural networks. We give a general introduction to the neural network approach to parton distributions, which provides a solution to the problem of constructing a faithful and unbiased probability distribution of parton densities based on available experimental information. We discuss in detail the techniques which are necessary in order to construct a Monte Carlo representation of the data, to construct and evolve neural parton distributions, and to train them in such a way that the correct statistical features of the data are reproduced. We present the results of the application of this method to the determination of the nonsinglet quark distribution up to next--to--next--to--leading order, and compare them with those obtained using other approaches.Comment: 46 pages, 18 figures, LaTeX with JHEP3 clas

    Infinite Latent Feature Selection: A Probabilistic Latent Graph-Based Ranking Approach

    Get PDF
    Feature selection is playing an increasingly significant role with respect to many computer vision applications spanning from object recognition to visual object tracking. However, most of the recent solutions in feature selection are not robust across different and heterogeneous set of data. In this paper, we address this issue proposing a robust probabilistic latent graph-based feature selection algorithm that performs the ranking step while considering all the possible subsets of features, as paths on a graph, bypassing the combinatorial problem analytically. An appealing characteristic of the approach is that it aims to discover an abstraction behind low-level sensory data, that is, relevancy. Relevancy is modelled as a latent variable in a PLSA-inspired generative process that allows the investigation of the importance of a feature when injected into an arbitrary set of cues. The proposed method has been tested on ten diverse benchmarks, and compared against eleven state of the art feature selection methods. Results show that the proposed approach attains the highest performance levels across many different scenarios and difficulties, thereby confirming its strong robustness while setting a new state of the art in feature selection domain.Comment: Accepted at the IEEE International Conference on Computer Vision (ICCV), 2017, Venice. Preprint cop

    Intelligent flight control systems

    Get PDF
    The capabilities of flight control systems can be enhanced by designing them to emulate functions of natural intelligence. Intelligent control functions fall in three categories. Declarative actions involve decision-making, providing models for system monitoring, goal planning, and system/scenario identification. Procedural actions concern skilled behavior and have parallels in guidance, navigation, and adaptation. Reflexive actions are spontaneous, inner-loop responses for control and estimation. Intelligent flight control systems learn knowledge of the aircraft and its mission and adapt to changes in the flight environment. Cognitive models form an efficient basis for integrating 'outer-loop/inner-loop' control functions and for developing robust parallel-processing algorithms
    corecore