864,629 research outputs found

    Statistical identification of geometric parameters for high speed train catenary

    Get PDF
    Pantograph/catenary interaction is known to be strongly dependent on the static geometry of the catenary, this research thus seeks to build a statistical model of this geometry. Sensitivity analyses provide a selection of relevant parameters affecting the geometry. After correction for the dynamic nature of the measurement, provide a database of measurements. One then seeks to solve the statistical inverse problem using the maximum entropy principle and the maximum likelihood method. Two methods of multivariate density estimations are presented, the Gaussian kernel density estimation method and the Gaussian parametric method. The results provide statistical information on the significant parameters and show that the messenger wire tension of the catenary hides sources of variability that are not yet taken into account in the model

    Bayesian model selection in logistic regression for the detection of adverse drug reactions

    Full text link
    Motivation: Spontaneous adverse event reports have a high potential for detecting adverse drug reactions. However, due to their dimension, exploring such databases requires statistical methods. In this context, disproportionality measures are used. However, by projecting the data onto contingency tables, these methods become sensitive to the problem of co-prescriptions and masking effects. Recently, logistic regressions have been used with a Lasso type penalty to perform the detection of associations between drugs and adverse events. However, the choice of the penalty value is open to criticism while it strongly influences the results. Results: In this paper, we propose to use a logistic regression whose sparsity is viewed as a model selection challenge. Since the model space is huge, a Metropolis-Hastings algorithm carries out the model selection by maximizing the BIC criterion. Thus, we avoid the calibration of penalty or threshold. During our application on the French pharmacovigilance database, the proposed method is compared to well established approaches on a reference data set, and obtains better rates of positive and negative controls. However, many signals are not detected by the proposed method. So, we conclude that this method should be used in parallel to existing measures in pharmacovigilance.Comment: 7 pages, 3 figures, submitted to Biometrical Journa

    Missing ordinal covariates with informative selection

    Get PDF
    This paper considers the problem of parameter estimation in a model for a continuous response variable y when an important ordinal explanatory variable x is missing for a large proportion of the sample. Non-missingness of x, or sample selection, is correlated with the response variable and/or with the unobserved values the ordinal explanatory variable takes when missing. We suggest solving the endogenous selection, or 'not missing at random' (NMAR), problem by modelling the informative selection mechanism, the ordinal explanatory variable, and the response variable together. The use of the method is illustrated by re-examining the problem of the ethnic gap in school achievement at age 16 in England using linked data from the National Pupil database (NPD), the Longitudinal Study of Young People in England (LSYPE), and the Census 2001.Missing covariate, sample selection, latent class models, ordinal variables, NMAR

    Das Lehrstück Kemit

    Get PDF
    This thesis project has been carried out at Linköpings universitet at the Department of Mechanical Engineering. The emphasis of the project lies in the exploration of automatic selection of components for a propulsion kit. Specifically for this project, propulsion based on electric power and meeting the requirements for use in a Micro Aerial Vehicle (MAV). The key features include a systematic selection method based on user criterias and a model for evaluating propeller performance. These are implemented in a program written as a part of the project. The conclusion is that it is possible to make a program capable of a component selection and that the programs usability is mainly reliant on three factors: model for propeller evaluation, method of selection and the quality of the component database

    A knowledge-based decision support system for roofing materials selection and cost estimating: a conceptual framework and data modelling

    Get PDF
    A plethora of materials is available to the modern day house designer but selecting the appropriate material is a complex task. It requires synthesising a multitude of performance criteria such as initial cost, maintenance cost, thermal performance and sustainability among others. This research aims to develop a Knowledge-based Decision support System for Material Selection (KDSMS) that facilitates the selection of optimal material for different sub elements of a roof design. The proposed system also has a facility for estimating roof cost based on the identified criteria. This paper presents the data modelling conceptual framework for the proposed system. The roof sub elements are modelled on the Building Cost Information Service (BCIS) Standard Form of Cost Analysis. This model consists of a knowledge base and a database to store different types of roofing materials with their corresponding performance characteristics and rankings. The system s knowledge is elicited from an extensive review of literature and the use of a domain expert forum. The proposed system employs the multi criteria decision method of TOPSIS (Technique of ranking Preferences by Similarity to the Ideal Solution), to resolve the materials selection and optimisation problem. The KDSMS is currently being developed for the housing sector of Northern Ireland

    A MOM-based ensemble method for robustness, subsampling and hyperparameter tuning

    Full text link
    Hyperparameters tuning and model selection are important steps in machine learning. Unfortunately, classical hyperparameter calibration and model selection procedures are sensitive to outliers and heavy-tailed data. In this work, we construct a selection procedure which can be seen as a robust alternative to cross-validation and is based on a median-of-means principle. Using this procedure, we also build an ensemble method which, trained with algorithms and corrupted heavy-tailed data, selects an algorithm, trains it with a large uncorrupted subsample and automatically tune its hyperparameters. The construction relies on a divide-and-conquer methodology, making this method easily scalable for autoML given a corrupted database. This method is tested with the LASSO which is known to be highly sensitive to outliers.Comment: 17 pages, 3 figure
    corecore