17 research outputs found

    A compendium of multi-omics data illuminating host responses to lethal human virus infections

    Get PDF
    Human infections caused by viral pathogens trigger a complex gamut of host responses that limit disease, resolve infection, generate immunity, and contribute to severe disease or death. Here, we present experimental methods and multi-omics data capture approaches representing the global host response to infection generated from 45 individual experiments involving human viruses from the Orthomyxoviridae, Filoviridae, Flaviviridae, and Coronaviridae families. Analogous experimental designs were implemented across human or mouse host model systems, longitudinal samples were collected over defined time courses, and global multi-omics data (transcriptomics, proteomics, metabolomics, and lipidomics) were acquired by microarray, RNA sequencing, or mass spectrometry analyses. For comparison, we have included transcriptomics datasets from cells treated with type I and type II human interferon. Raw multi-omics data and metadata were deposited in public repositories, and we provide a central location linking the raw data with experimental metadata and ready-to-use, quality-controlled, statistically processed multi-omics datasets not previously available in any public repository. This compendium of infection-induced host response data for reuse will be useful for those endeavouring to understand viral disease pathophysiology and network biology

    Integration of infant metabolite, genetic and islet autoimmunity signatures to predict type 1 diabetes by 6 years of age.

    No full text
    CONTEXT: Biomarkers that can accurately predict risk of type 1 diabetes (T1D) in genetically predisposed children can facilitate interventions to delay or prevent the disease. OBJECTIVE: Determine if a combination of genetic, immunologic, and metabolic features, measured at infancy, can be utilized to predict the likelihood that a child will develop T1D by the age of 6 years. DESIGN: Newborns with HLA typing enrolled in the prospective birth cohort of The Environmental Determinants of Diabetes in the Young (TEDDY). SETTING: TEDDY ascertained children in Finland, Germany, Sweden, and the United States. PATIENTS: TEDDY children were either from the general population or from families with T1D with an HLA genotype associated with T1D specific to TEDDY eligibility criteria. From the TEDDY cohort there were 702 children will all data sources measured at 3, 6 and 9 months of age, 11.4% of which progressed to T1D by the age of 6. INTERVENTIONS: None. MAIN OUTCOME MEASURES: Diagnosis of T1D as diagnosed by American Diabetes Association criteria. RESULTS: Machine learning-based feature selection yielded classifiers based on disparate demographic, immunologic, genetic and metabolite features. The accuracy of the model utilizing all available data evaluated by the Area Under a Receiver Operating Characteristic Curve is 0.84. Reducing to only 3- and 9-month measurements did not reduce the AUC significantly. Metabolomics had the largest value when evaluating the accuracy at a low false positive rate. CONCLUSIONS: The metabolite features identified as important for progression to T1D by age 6 point to altered sugar metabolism in infancy. Integrating this information with classic risk factors improves prediction of the progression to T1D in early childhood

    Quality control analysis in real-time (QC-ART): A tool for real-time quality control assessment of mass spectrometry-based proteomics data.

    No full text
    Liquid chromatography-mass spectrometry (LC-MS)-based proteomics studies of large sample cohorts can easily require from months to years to complete. Acquiring consistent, high-quality data in such large-scale studies is challenging because of normal variations in instrumentation performance over time, as well as artifacts introduced by the samples themselves, such as those because of collection, storage and processing. Existing quality control methods for proteomics data primarily focus on post-hoc analysis to remove low-quality data that would degrade downstream statistics; they are not designed to evaluate the data in near real-time, which would allow for interventions as soon as deviations in data quality are detected. In addition to flagging analyses that demonstrate outlier behavior, evaluating how the data structure changes over time can aide in understanding typical instrument performance or identify issues such as a degradation in data quality because of the need for instrument cleaning and/or re-calibration. To address this gap for proteomics, we developed Quality Control Analysis in Real-Time (QC-ART), a tool for evaluating data as they are acquired to dynamically flag potential issues with instrument performance or sample quality. QC-ART has similar accuracy as standard post-hoc analysis methods with the additional benefit of real-time analysis. We demonstrate the utility and performance of QC-ART in identifying deviations in data quality because of both instrument and sample issues in near real-time for LC-MS-based plasma proteomics analyses of a sample subset of The Environmental Determinants of Diabetes in the Young cohort. We also present a case where QC-ART facilitated the identification of oxidative modifications, which are often underappreciated in proteomic experiments
    corecore