275 research outputs found

    Quality management in heavy duty manufacturing industry: TQM vs. Six Sigma

    Get PDF
    ‘Is TQM a management fad?’ This question has been extensively documented in the quality management literature; and will be tackled in this research though a critical literature review on the area. ‘TQM versus Six-Sigma’ debate, which has also been a fundamental challenge in this research filed, is addressed by a thematic and chronological review on the peer papers. To evaluate this challenge in practice, a primary research in heavy duty machinery production industry have been conducted using a case-study on, J C Bamford Excavators Ltd (JCB), the largest European construction machinery producer. The result highlights that TQM is a natural foundation to build up Six-Sigma upon; and not surprisingly the quality yield in a TQM approach complemented by Six-sigma is far higher and more stable than when TQM with no Six-Sigma focus is being put in place; thus presenting the overall finding that TQM and Six Sigma are compliments, not substitutes. The study will be concluded with an overview on quality management approaches in the heavy duty manufacturing industry to highlight the way forward for the industry

    On the combination of omics data for prediction of binary outcomes

    Full text link
    Enrichment of predictive models with new biomolecular markers is an important task in high-dimensional omic applications. Increasingly, clinical studies include several sets of such omics markers available for each patient, measuring different levels of biological variation. As a result, one of the main challenges in predictive research is the integration of different sources of omic biomarkers for the prediction of health traits. We review several approaches for the combination of omic markers in the context of binary outcome prediction, all based on double cross-validation and regularized regression models. We evaluate their performance in terms of calibration and discrimination and we compare their performance with respect to single-omic source predictions. We illustrate the methods through the analysis of two real datasets. On the one hand, we consider the combination of two fractions of proteomic mass spectrometry for the calibration of a diagnostic rule for the detection of early-stage breast cancer. On the other hand, we consider transcriptomics and metabolomics as predictors of obesity using data from the Dietary, Lifestyle, and Genetic determinants of Obesity and Metabolic syndrome (DILGOM) study, a population-based cohort, from Finland

    Generating Explainable and Effective Data Descriptors Using Relational Learning: Application to Cancer Biology

    Get PDF
    The key to success in machine learning is the use of effective data representations. The success of deep neural networks (DNNs) is based on their ability to utilize multiple neural network layers, and big data, to learn how to convert simple input representations into richer internal representations that are effective for learning. However, these internal representations are sub-symbolic and difficult to explain. In many scientific problems explainable models are required, and the input data is semantically complex and unsuitable for DNNs. This is true in the fundamental problem of understanding the mechanism of cancer drugs, which requires complex background knowledge about the functions of genes/proteins, their cells, and the molecular structure of the drugs. This background knowledge cannot be compactly expressed propositionally, and requires at least the expressive power of Datalog. Here we demonstrate the use of relational learning to generate new data descriptors in such semantically complex background knowledge. These new descriptors are effective: adding them to standard propositional learning methods significantly improves prediction accuracy. They are also explainable, and add to our understanding of cancer. Our approach can readily be expanded to include other complex forms of background knowledge, and combines the generality of relational learning with the efficiency of standard propositional learning

    Statistical Estimation of Correlated Genome Associations to a Quantitative Trait Network

    Get PDF
    Many complex disease syndromes, such as asthma, consist of a large number of highly related, rather than independent, clinical or molecular phenotypes. This raises a new technical challenge in identifying genetic variations associated simultaneously with correlated traits. In this study, we propose a new statistical framework called graph-guided fused lasso (GFlasso) to directly and effectively incorporate the correlation structure of multiple quantitative traits such as clinical metrics and gene expressions in association analysis. Our approach represents correlation information explicitly among the quantitative traits as a quantitative trait network (QTN) and then leverages this network to encode structured regularization functions in a multivariate regression model over the genotypes and traits. The result is that the genetic markers that jointly influence subgroups of highly correlated traits can be detected jointly with high sensitivity and specificity. While most of the traditional methods examined each phenotype independently and combined the results afterwards, our approach analyzes all of the traits jointly in a single statistical framework. This allows our method to borrow information across correlated phenotypes to discover the genetic markers that perturb a subset of the correlated traits synergistically. Using simulated datasets based on the HapMap consortium and an asthma dataset, we compared the performance of our method with other methods based on single-marker analysis and regression-based methods that do not use any of the relational information in the traits. We found that our method showed an increased power in detecting causal variants affecting correlated traits. Our results showed that, when correlation patterns among traits in a QTN are considered explicitly and directly during a structured multivariate genome association analysis using our proposed methods, the power of detecting true causal SNPs with possibly pleiotropic effects increased significantly without compromising performance on non-pleiotropic SNPs

    Geographically weighted elastic net logistic regression

    Get PDF
    This paper develops a localized approach to elastic net logistic regression, extending previous research describing a localized elastic net as an extension to a localized ridge regression or a localized lasso. All such models have the objective to capture data relationships that vary across space. Geographically weighted elastic net logistic regression is first evaluated through a simulation experiment and shown to provide a robust approach for local model selection and alleviating local collinearity, before application to two case studies: county-level voting patterns in the 2016 USA presidential election, examining the spatial structure of socio-economic factors associated with voting for Trump, and a species presence–absence data set linked to explanatory environmental and climatic factors at gridded locations covering mainland USA. The approach is compared with other logistic regressions. It improves prediction for the election case study only which exhibits much greater spatial heterogeneity in the binary response than the species case study. Model comparisons show that standard geographically weighted logistic regression over-estimated relationship non-stationarity because it fails to adequately deal with collinearity and model selection. Results are discussed in the context of predictor variable collinearity and selection and the heterogeneities that were observed. Ongoing work is investigating locally derived elastic net parameters

    Expression QTLs Mapping and Analysis: A Bayesian Perspective.

    Get PDF
    The aim of expression Quantitative Trait Locus (eQTL) mapping is the identification of DNA sequence variants that explain variation in gene expression. Given the recent yield of trait-associated genetic variants identified by large-scale genome-wide association analyses (GWAS), eQTL mapping has become a useful tool to understand the functional context where these variants operate and eventually narrow down functional gene targets for disease. Despite its extensive application to complex (polygenic) traits and disease, the majority of eQTL studies still rely on univariate data modeling strategies, i.e., testing for association of all transcript-marker pairs. However these "one at-a-time" strategies are (1) unable to control the number of false-positives when an intricate Linkage Disequilibrium structure is present and (2) are often underpowered to detect the full spectrum of trans-acting regulatory effects. Here we present our viewpoint on the most recent advances on eQTL mapping approaches, with a focus on Bayesian methodology. We review the advantages of the Bayesian approach over frequentist methods and provide an empirical example of polygenic eQTL mapping to illustrate the different properties of frequentist and Bayesian methods. Finally, we discuss how multivariate eQTL mapping approaches have distinctive features with respect to detection of polygenic effects, accuracy, and interpretability of the results
    • …
    corecore