11 research outputs found

    Scalable Feature Selection Applications for Genome-Wide Association Studies of Complex Diseases

    Get PDF
    Personalized medicine will revolutionize our capabilities to combat disease. Working toward this goal, a fundamental task is the deciphering of geneticvariants that are predictive of complex diseases. Modern studies, in the formof genome-wide association studies (GWAS) have afforded researchers with the opportunity to reveal new genotype-phenotype relationships through the extensive scanning of genetic variants. These studies typically contain over half a million genetic features for thousands of individuals. Examining this with methods other than univariate statistics is a challenging task requiring advanced algorithms that are scalable to the genome-wide level. In the future, next-generation sequencing studies (NGS) will contain an even larger number of common and rare variants. Machine learning-based feature selection algorithms have been shown to have the ability to effectively create predictive models for various genotype-phenotype relationships. This work explores the problem of selecting genetic variant subsets that are the most predictive of complex disease phenotypes through various feature selection methodologies, including filter, wrapper and embedded algorithms. The examined machine learning algorithms were demonstrated to not only be effective at predicting the disease phenotypes, but also doing so efficiently through the use of computational shortcuts. While much of the work was able to be run on high-end desktops, some work was further extended so that it could be implemented on parallel computers helping to assure that they will also scale to the NGS data sets. Further, these studies analyzed the relationships between various feature selection methods and demonstrated the need for careful testing when selecting an algorithm. It was shown that there is no universally optimal algorithm for variant selection in GWAS, but rather methodologies need to be selected based on the desired outcome, such as the number of features to be included in the prediction model. It was also demonstrated that without proper model validation, for example using nested cross-validation, the models can result in overly-optimistic prediction accuracies and decreased generalization ability. It is through the implementation and application of machine learning methods that one can extract predictive genotype–phenotype relationships and biological insights from genetic data sets.Siirretty Doriast

    Regularized Machine Learning in the Genetic Prediction of Complex Traits

    Get PDF
    Compared to univariate analysis of genome-wide association (GWA) studies, machine learning&ndash;based models have been shown to provide improved means of learning such multilocus panels of genetic variants and their interactions that are most predictive of complex phenotypic traits. Many applications of predictive modeling rely on effective variable selection, often implemented through model regularization, which penalizes the model complexity and enables predictions in individuals outside of the training dataset. However, the different regularization approaches may also lead to considerable differences, especially in the number of genetic variants needed for maximal predictive accuracy, as illustrated here in examples from both disease classification and quantitative trait prediction. We also highlight the potential pitfalls of the regularized machine learning models, related to issues such as model overfitting to the training data, which may lead to over-optimistic prediction results, as well as identifiability of the predictive variants, which is important in many medical applications. While genetic risk prediction for human diseases is used as a motivating use case, we argue that these models are also widely applicable in nonhuman applications, such as animal and plant breeding, where accurate genotype-to-phenotype modeling is needed. Finally, we discuss some key future advances, open questions and challenges in this developing field, when moving toward low-frequency variants and cross-phenotype interactions.</p

    Penalty terms and loss functions.

    No full text
    <p>(A) Penalty terms: <i>L</i><sub>0</sub>-norm imposes the most explicit constraint on the model complexity as it effectively counts the number of nonzero entries in the model parameter vector. While it is possible to train prediction models with <i>L</i><sub>0</sub>-penalty using, e.g., greedy or other types of discrete optimization methods, the problem becomes mathematically challenging due to the nonconvexity of the constraint, especially when other than the squared loss function is used. The convexity of the <i>L</i><sub>1</sub> and <i>L</i><sub>2</sub> norms makes them easier for the optimization. While the <i>L</i><sub>2</sub> norm has good regularization properties, it must be used together with either <i>L</i><sub>0</sub> or <i>L</i><sub>1</sub> norms to perform feature selection. (B) Loss functions: The plain classification error is difficult to minimize due to its nonconvex and discontinuous nature, and therefore one often resorts to its better behaving surrogates, including the hinge loss used with SVMs, the cross-entropy used with logistic regression, or the squared error used with regularized least-squares classification and regression. These surrogates in turn differ both in their quality of approximating the classification error and in terms of the optimization machinery they can be minimized with (<a href="http://www.plosgenetics.org/article/info:doi/10.1371/journal.pgen.1004754#pgen.1004754.s001" target="_blank">Text S1</a>).</p

    Performance of regularized machine learning models.

    No full text
    <p>Upper panel: Behavior of the learning approaches in terms of their predictive accuracy (<i>y</i>-axis) as a function of the number of selected variants (<i>x</i>-axis). Differences can be attributed to the genotypic and phenotypic heterogeneity as well as genotyping density and quality. (A) The area under the receiver operating characteristic curve (AUC) for the prediction of Type 1 diabetes (T1D) cases in SNP data from WTCCC <a href="http://www.plosgenetics.org/article/info:doi/10.1371/journal.pgen.1004754#pgen.1004754-Wellcome1" target="_blank">[118]</a>, representing ca. one million genetic features and ca. 5,000 individuals in a case-control setup. (B) Coefficient of determination (<i>R</i><sup>2</sup>) for the prediction of a continuous trait (Tunicamycin) in SNP data from a cross between two yeast strains (Y2C) <a href="http://www.plosgenetics.org/article/info:doi/10.1371/journal.pgen.1004754#pgen.1004754-Bloom1" target="_blank">[44]</a>, representing ca. 12,000 variants and ca. 1,000 segregants in a controlled laboratory setup. The peak prediction accuracy/number of most predictive variants are listed in the legend. The model validation was implemented using nested 3-fold cross-validation (CV) <a href="http://www.plosgenetics.org/article/info:doi/10.1371/journal.pgen.1004754#pgen.1004754-Okser2" target="_blank">[5]</a>. Prior to any analysis being done, the data was split into three folds. On each outer round of CV, two of the folds were combined forming a training set, and the remaining one was used as an independent test set. On each round, all feature and parameter selection was done using a further internal 3-fold CV on the training set, and the predictive performance of the learned models was evaluated on the independent test set. The final performance estimates were calculated as the average over these three iterations of the experiment. In learning approaches where internal CV was not needed to select model parameters (e.g., log odds), this is equivalent to a standard 3-fold CV. T1D data: the <i>L</i><sub>2</sub>-regularized (ridge) regression was based on selecting the top 500 variants according to the <i>χ</i><sup>2</sup> filter. For wrappers, we used our greedy <i>L</i><sub>2</sub>-regularized least squares (RLS) implementation <a href="http://www.plosgenetics.org/article/info:doi/10.1371/journal.pgen.1004754#pgen.1004754-Pahikkala1" target="_blank">[30]</a>, while the embedded methods, Lasso, Elastic Net and <i>L</i><sub>1</sub>-logistic regression, were implemented through the Scikit-Learn <a href="http://www.plosgenetics.org/article/info:doi/10.1371/journal.pgen.1004754#pgen.1004754-Pedregosa1" target="_blank">[119]</a>, interpolated across various regularization parameters up to the maximal number of variants (500 or 1,000). As a baseline model, we implemented a log odds-ratio weighted sum of the minor allele dosage in the 500 selected variants within each individual <a href="http://www.plosgenetics.org/article/info:doi/10.1371/journal.pgen.1004754#pgen.1004754-Evans1" target="_blank">[25]</a>. Y2C: the filter method was based on the top 1,000 variants selected according to <i>R</i><sup>2</sup>, followed by <i>L</i><sub>2</sub>-regularization within greedy RLS using nested CV. As a baseline model, we implemented a greedy version of least squares (LS), which is similar to the stepwise forward regression used in the original work <a href="http://www.plosgenetics.org/article/info:doi/10.1371/journal.pgen.1004754#pgen.1004754-Bloom1" target="_blank">[44]</a>; the greedy LS differs from the greedy RLS in terms that it implements regularization through optimization of <i>L</i><sub>0</sub> norm instead of <i>L</i><sub>2</sub>. It was noted that the greedy LS method drops around the point where the number of selected variants exceeds the number training examples (here, 400). Lower panel: Overlap in the genetic features selected by the different approaches. (C) The numbers of selected variants within the major histocompatibility complex (MHC) are shown in parentheses for the T1D data. (D) The overlap among then maximally predictive variants in the Y2C data. Note: these results should be considered merely as illustrative examples. Differing results may be obtained when other prediction models are implemented in other genetic datasets or other prediction applications.</p
    corecore