179 research outputs found

    MEDIC: a practical disease vocabulary used at the Comparative Toxicogenomics Database

    Get PDF
    The Comparative Toxicogenomics Database (CTD) is a public resource that promotes understanding about the effects of environmental chemicals on human health. CTD biocurators manually curate a triad of chemical–gene, chemical–disease and gene–disease relationships from the scientific literature. The CTD curation paradigm uses controlled vocabularies for chemicals, genes and diseases. To curate disease information, CTD first had to identify a source of controlled terms. Two resources seemed to be good candidates: the Online Mendelian Inheritance in Man (OMIM) and the ‘Diseases’ branch of the National Library of Medicine's Medical Subject Headers (MeSH). To maximize the advantages of both, CTD biocurators undertook a novel initiative to map the flat list of OMIM disease terms into the hierarchical nature of the MeSH vocabulary. The result is CTD’s ‘merged disease vocabulary’ (MEDIC), a unique resource that integrates OMIM terms, synonyms and identifiers with MeSH terms, synonyms, definitions, identifiers and hierarchical relationships. MEDIC is both a deep and broad vocabulary, composed of 9700 unique diseases described by more than 67 000 terms (including synonyms). It is freely available to download in various formats from CTD. While neither a true ontology nor a perfect solution, this vocabulary has nonetheless proved to be extremely successful and practical for our biocurators in generating over 2.5 million disease-associated toxicogenomic relationships in CTD. Other external databases have also begun to adopt MEDIC for their disease vocabulary. Here, we describe the construction, implementation, maintenance and use of MEDIC to raise awareness of this resource and to offer it as a putative scaffold in the formal construction of an official disease ontology

    Recent advances in biocuration: Meeting report from the Fifth International Biocuration Conference

    Get PDF
    The 5th International Biocuration Conference brought together over 300 scientists to exchange on their work, as well as discuss issues relevant to the International Society for Biocuration’s (ISB) mission. Recurring themes this year included the creation and promotion of gold standards, the need for more ontologies, and more formal interactions with journals. The conference is an essential part of the ISB\u27s goal to support exchanges among members of the biocuration community. Next year\u27s conference will be held in Cambridge, UK, from 7 to 10 April 2013. In the meanwhile, the ISB website provides information about the society\u27s activities (http://biocurator.org), as well as related events of interest

    BioGraph: unsupervised biomedical knowledge discovery via automated hypothesis generation

    Get PDF
    We present BioGraph, a data integration and data mining platform for the exploration and discovery of biomedical information. The platform offers prioritizations of putative disease genes, supported by functional hypotheses. We show that BioGraph can retrospectively confirm recently discovered disease genes and identify potential susceptibility genes, outperforming existing technologies, without requiring prior domain knowledge. Additionally, BioGraph allows for generic biomedical applications beyond gene discovery. BioGraph is accessible at http://www.biograph.be

    NanoSolveIT project: driving nanoinformatics research to develop innovative and integrated tools for in silico nanosafety assessment

    Get PDF
    Nanotechnology has enabled the discovery of a multitude of novel materials exhibiting unique physicochemical (PChem) properties compared to their bulk analogues. These properties have led to a rapidly increasing range of commercial applications; this, however, may come at a cost, if an association to long-term health and environmental risks is discovered or even just perceived. Many nanomaterials (NMs) have not yet had their potential adverse biological effects fully assessed, due to costs and time constraints associated with the experimental assessment, frequently involving animals. Here, the available NM libraries are analyzed for their suitability for integration with novel nanoinformatics approaches and for the development of NM specific Integrated Approaches to Testing and Assessment (IATA) for human and environmental risk assessment, all within the NanoSolveIT cloud-platform. These established and well-characterized NM libraries (e.g. NanoMILE, NanoSolutions, NANoREG, NanoFASE, caLIBRAte, NanoTEST and the Nanomaterial Registry (>2000 NMs)) contain physicochemical characterization data as well as data for several relevant biological endpoints, assessed in part using harmonized Organisation for Economic Co-operation and Development (OECD) methods and test guidelines. Integration of such extensive NM information sources with the latest nanoinformatics methods will allow NanoSolveIT to model the relationships between NM structure (morphology), properties and their adverse effects and to predict the effects of other NMs for which less data is available. The project specifically addresses the needs of regulatory agencies and industry to effectively and rapidly evaluate the exposure, NM hazard and risk from nanomaterials and nano-enabled products, enabling implementation of computational ‘safe-by-design’ approaches to facilitate NM commercialization

    Applications of Genome-Wide Screening and Systems Biology Approaches in Drug Repositioning

    Get PDF
    Simple Summary Drug repurposing is an accelerated route for drug development and a promising approach for finding medications for orphan and common diseases. Here, we compiled databases that comprise both computationally- or experimentally-derived data, and categorized them based on quiddity and origin of data, further focusing on those that present high throughput omic data or drug screens. These databases were then contextualized with genome-wide screening methods such as CRISPR/Cas9 and RNA interference, as well as state of art systems biology approaches that enable systematic characterizations of multi-omic data to find new indications for approved drugs or those that reached the latest phases of clinical trials. Modern drug discovery through de novo drug discovery entails high financial costs, low success rates, and lengthy trial periods. Drug repositioning presents a suitable approach for overcoming these issues by re-evaluating biological targets and modes of action of approved drugs. Coupling high-throughput technologies with genome-wide essentiality screens, network analysis, genome-scale metabolic modeling, and machine learning techniques enables the proposal of new drug-target signatures and uncovers unanticipated modes of action for available drugs. Here, we discuss the current issues associated with drug repositioning in light of curated high-throughput multi-omic databases, genome-wide screening technologies, and their application in systems biology/medicine approaches

    Predicting Adverse Drug Effects from Literature Assertions that Link Drugs to Targets and Targets to Effects

    Get PDF
    Adverse drug effects (ADEs) are a major reason for drug candidate failure in clinical trials; thus, it is critical to predict possible ADEs in the early stages of drug discovery. In this study, cheminformatics, bioinformatics, and data mining approaches were employed to integrate and analyze publicly-available pharmacological and clinical data with the goal of inferring novel associations between drugs, targets, and ADEs. A new database was created that integrated experimental drug-target binding data and known associations between drugs (7448 unique instances), targets (1280), and ADEs (4492) expressed as assertions found in the literature. Unreported associations between drugs, targets, and ADEs were inferred, and inferences were prioritized as testable hypotheses. As a proof of concept, an association was identified between paroxetine and thrombocytopenic purpura using a focused subset of ~47K top-ranked inferences published prior to the first reports confirming this association in 2013. Given the increasing costs of bringing new drug entities to market, there is a strong need for cost-effective methods of identifying potential adverse effects of a drug candidate early on in the development process. The workflow presented here, based on free-access databases and an association-based inference scheme, has provided chemical-ADE inferences that have been validated post-hoc in literature case reports. With refinement of prioritization schemes for the generated chemical-ADE inferences, this workflow may provide an effective computational method for the early detection of drug candidate ADEs.Doctor of Pharmac

    Contextual Analysis of Large-Scale Biomedical Associations for the Elucidation and Prioritization of Genes and their Roles in Complex Disease

    Get PDF
    Vast amounts of biomedical associations are easily accessible in public resources, spanning gene-disease associations, tissue-specific gene expression, gene function and pathway annotations, and many other data types. Despite this mass of data, information most relevant to the study of a particular disease remains loosely coupled and difficult to incorporate into ongoing research. Current public databases are difficult to navigate and do not interoperate well due to the plethora of interfaces and varying biomedical concept identifiers used. Because no coherent display of data within a specific problem domain is available, finding the latent relationships associated with a disease of interest is impractical. This research describes a method for extracting the contextual relationships embedded within associations relevant to a disease of interest. After applying the method to a small test data set, a large-scale integrated association network is constructed for application of a network propagation technique that helps uncover more distant latent relationships. Together these methods are adept at uncovering highly relevant relationships without any a priori knowledge of the disease of interest. The combined contextual search and relevance methods power a tool which makes pertinent biomedical associations easier to find, easier to assimilate into ongoing work, and more prominent than currently available databases. Increasing the accessibility of current information is an important component to understanding high-throughput experimental results and surviving the data deluge
    • …
    corecore