148 research outputs found

    On the combination of omics data for prediction of binary outcomes

    Full text link
    Enrichment of predictive models with new biomolecular markers is an important task in high-dimensional omic applications. Increasingly, clinical studies include several sets of such omics markers available for each patient, measuring different levels of biological variation. As a result, one of the main challenges in predictive research is the integration of different sources of omic biomarkers for the prediction of health traits. We review several approaches for the combination of omic markers in the context of binary outcome prediction, all based on double cross-validation and regularized regression models. We evaluate their performance in terms of calibration and discrimination and we compare their performance with respect to single-omic source predictions. We illustrate the methods through the analysis of two real datasets. On the one hand, we consider the combination of two fractions of proteomic mass spectrometry for the calibration of a diagnostic rule for the detection of early-stage breast cancer. On the other hand, we consider transcriptomics and metabolomics as predictors of obesity using data from the Dietary, Lifestyle, and Genetic determinants of Obesity and Metabolic syndrome (DILGOM) study, a population-based cohort, from Finland

    A Comparison of Machine Learning Methods for Cross-Domain Few-Shot Learning

    Get PDF
    We present an empirical evaluation of machine learning algorithms in cross-domain few-shot learning based on a fixed pre-trained feature extractor. Experiments were performed in five target domains (CropDisease, EuroSAT, Food101, ISIC and ChestX) and using two feature extractors: a ResNet10 model trained on a subset of ImageNet known as miniImageNet and a ResNet152 model trained on the ILSVRC 2012 subset of ImageNet. Commonly used machine learning algorithms including logistic regression, support vector machines, random forests, nearest neighbour classification, naĂŻve Bayes, and linear and quadratic discriminant analysis were evaluated on the extracted feature vectors. We also evaluated classification accuracy when subjecting the feature vectors to normalisation using p-norms. Algorithms originally developed for the classification of gene expression data—the nearest shrunken centroid algorithm and LDA ensembles obtained with random projections—were also included in the experiments, in addition to a cosine similarity classifier that has recently proved popular in few-shot learning. The results enable us to identify algorithms, normalisation methods and pre-trained feature extractors that perform well in cross-domain few-shot learning. We show that the cosine similarity classifier and ℓÂČ -regularised 1-vs-rest logistic regression are generally the best-performing algorithms. We also show that algorithms such as LDA yield consistently higher accuracy when applied to ℓÂČ -normalised feature vectors. In addition, all classifiers generally perform better when extracting feature vectors using the ResNet152 model instead of the ResNet10 model

    Does pathway analysis make it easier for common variants to tag rare ones?

    Get PDF
    Analyzing sequencing data is difficult because of the low frequency of rare variants, which may result in low power to detect associations. We consider pathway analysis to detect multiple common and rare variants jointly and to investigate whether analysis at the pathway level provides an alternative strategy for identifying susceptibility genes. Available pathway analysis methods for data from genome-wide association studies might not be efficient because these methods are designed to detect common variants. Here, we investigate the performance of several existing pathway analysis methods for sequencing data. In particular, we consider the global test, which does not consider linkage disequilibrium between the variants in a gene. We improve the performance of the global test by assigning larger weights to rare variants, as proposed in the weighted-sum approach. Our conclusion is that straightforward application of pathway analysis is not satisfactory; hence, when common and rare variants are jointly analyzed, larger weights should be assigned to rare variants

    Prediction meets causal inference: the role of treatment in clinical prediction models

    Get PDF
    In this paper we study approaches for dealing with treatment when developing a clinical prediction model. Analogous to the estimand framework recently proposed by the European Medicines Agency for clinical trials, we propose a ‘predictimand’ framework of different questions that may be of interest when predicting risk in relation to treatment started after baseline. We provide a formal definition of the estimands matching these questions, give examples of settings in which each is useful and discuss appropriate estimators including their assumptions. We illustrate the impact of the predictimand choice in a dataset of patients with end-stage kidney disease. We argue that clearly defining the estimand is equally important in prediction research as in causal inference

    Evaluating the Effectiveness and Cost-Effectiveness of Seizure Dogs in Persons With Medically Refractory Epilepsy in the Netherlands

    Get PDF
    __Background:__ Epilepsy is associated with a high disease burden, impacting the lives of people with epilepsy and their caregivers and family. Persons with medically refractory epilepsy experience the greatest burden, suffering from profound physical, psychological, and social consequences. Anecdotal evidence suggests these persons may benefit from a seizure dog. As the training of a seizure dog is a substantial investment, their accessibility is limited in the absence of collective reimbursement as is seen in the Netherlands. Despite sustained interest in seizure dogs, scientific knowledge on their benefits and costs remains scarce. To substantiate reimbursement decisions stronger evidence is required. The EPISODE study aims to provide this evidence by evaluating the effectiveness and cost-effectiveness of seizure dogs in adults with medically refractory epilepsy. __Methods:__ The study is designed as a stepped wedge randomized controlled trial that compares the use of seizure dogs in addition to usual care, with usual care alone. The study includes adults with epilepsy for whom current treatment options failed to achieve seizure freedom. Seizure frequency of participants should be at least two seizures per week, and the seizures should be associated with a high risk of injury or dysfunction. During the 3 year follow-up period, participants receive a seizure dog in a randomized order. Outcome measures are taken at multiple time points both before and after receiving the seizure dog. Seizure frequency is the primary outcome of the study and will be recorded continuously using a seizure diary. Questionnaires measuring seizure severity, quality of life, well-being, resource use, productivity, social participation, and caregiver burden will be completed at baseline and every 3 months thereafter. The study is designed to include a minimum of 25 participants. __Discussion:__ This protocol describes the first randomized controlled trial on seizure dogs. The study will provide comprehensive data on the effectiveness and cost effectiveness of seizure dogs in adults with medically refractory epilepsy. Broader benefits of seizure dogs for persons with epilepsy and their caregivers are taken into account, as well as the welfare of the dogs. The findings of the study can be used to inform decision-makers on the reimbursement of seizure dogs

    Perceptron Connectives in Knowledge Representation

    Get PDF
    We discuss the role of perceptron (or threshold) connectives in the context of Description Logic, and in particular their possible use as a bridge between statistical learning of models from data and logical reasoning over knowledge bases. We prove that such connectives can be added to the language of most forms of Description Logic without increasing the complexity of the corresponding inference problem. We show, with a practical example over the Gene Ontology, how even simple instances of perceptron connectives are expressive enough to represent learned, complex concepts derived from real use cases. This opens up the possibility to import concepts learnt from data into existing ontologies

    Evaluating the role of paternal factors in aetiology and prognosis of recurrent pregnancy loss: Study protocol for a hospital-based multicentre case-control study and cohort study (REMI III project)

    Get PDF
    Introduction Recurrent pregnancy loss (RPL) is defined as the spontaneous demise of two or more pregnancies before the fetus reaches viability. Despite investigation of multiple known maternal risk factors, in more than 50% of couples, this condition remains unexplained. Studies focusing on paternal factors in RPL are scarce, and therefore, paternal evaluation in RPL is currently very limited. However, regarding single miscarriage, there are multiple publications suggesting a contributive role of paternal factors. In this project, we aim to identify paternal factors associated with RPL and to improve couple-specific prediction of future pregnancy outcomes by developing a prediction model containing both maternal and paternal factors. Methods and analysis In a case-control design, the relation between unexplained RPL and paternal age, lifestyle factors, sperm DNA damage and immunomodulatory factors in peripheral blood and semen will be studied. Prospectively, 135 couples with naturally conceived unexplained RPL (cases) and 135 fertile couples without a history of pregnancy loss (controls) will be included, with collection of paternal blood and semen samples and documentation of clinical and lifestyle characteristics. In addition, 600 couples from both groups will be included retrospectively. To adjust for confounders, multivariate logistic regression will be used. The predictive value of paternal and maternal factors will be studied in the total RPL cohort consisting of approximately 735 couples. The primary outcome of the cohort study is live birth within 5 years after initial visit of the clinic. Secondary outcomes are ongoing pregnancy, time interval until next pregnancy and pregnancy complications. Ethics and dissemination This project is approved by the Medical Research Ethics Committee

    Centralization of Esophageal Cancer Surgery: Does It Improve Clinical Outcome?

    Get PDF
    Background: The volume-outcome relationship for complex surgical procedures has been extensively studied. Most studies are based on administrative data and use in-hospital mortality as the sole outcome measure. It is still unknown if concentration of these procedures leads to improvement of clinical outcome. The aim of our study was to audit the process and effect of centralizing oesophageal resections for cancer by using detailed clinical data. Methods: From January 1990 until December 2004, 555 esophagectomies for cancer were performed in 11 hospitals in the region of the Comprehensive Cancer Center West (CCCW); 342 patients were operated on before and 213 patients after the introduction of a centralization project. In this project patients were referred to the hospitals which showed superior outcomes in a regional audit. In this audit patient, tumor, and operative details as well as clinical outcome were compared between hospitals. The outcome of both cohorts, patients operated on before and after the start of the project, were evaluated. Results: Despite the more severe comorbidity of the patient group, outcome improved after centralizing esophageal resections. Along with a reduction in postoperative morbidity and length of stay, mortality fell from 12% to 4% and survival improved significantly (P = 0.001). The hospitals with the highest procedural volume showed the biggest improvement in outcome. Conclusion: Volume is an important determinant of quality of care in esophageal cancer surgery. Referral of patients with esophageal cancer to surgical units with adequate experience and superior outcomes (outcome-based referral) improves quality of care
    • 

    corecore