9,707 research outputs found

    Computational Strategies for Scalable Genomics Analysis.

    Get PDF
    The revolution in next-generation DNA sequencing technologies is leading to explosive data growth in genomics, posing a significant challenge to the computing infrastructure and software algorithms for genomics analysis. Various big data technologies have been explored to scale up/out current bioinformatics solutions to mine the big genomics data. In this review, we survey some of these exciting developments in the applications of parallel distributed computing and special hardware to genomics. We comment on the pros and cons of each strategy in the context of ease of development, robustness, scalability, and efficiency. Although this review is written for an audience from the genomics and bioinformatics fields, it may also be informative for the audience of computer science with interests in genomics applications

    ¿Quién mató a Elaine? Autos robot y toma de decisiones

    Get PDF
    Este artículo discute sobre vehículos autónomos y cuáles deben ser los criterios para la configuración de los algoritmos en la toma de decisiones de los autos robot. Analiza el caso HWY18MH010 de la NTSB sobre el primer peatón que fue atropellado y muerto por un taxi de uber conducido por un sistema computarizado autónomo. — This article discusses autonomous vehicles and what should be the criteria for the configuration of algorithms in the decision making of robot cars. Analyze the case HWY18MH010 of the NTSB about the first pedestrian who was run over and killed by a Uber taxi driven by an autonomous computerized system

    Deep learning extends de novo protein modelling coverage of genomes using iteratively predicted structural constraints

    Get PDF
    The inapplicability of amino acid covariation methods to small protein families has limited their use for structural annotation of whole genomes. Recently, deep learning has shown promise in allowing accurate residue-residue contact prediction even for shallow sequence alignments. Here we introduce DMPfold, which uses deep learning to predict inter-atomic distance bounds, the main chain hydrogen bond network, and torsion angles, which it uses to build models in an iterative fashion. DMPfold produces more accurate models than two popular methods for a test set of CASP12 domains, and works just as well for transmembrane proteins. Applied to all Pfam domains without known structures, confident models for 25% of these so-called dark families were produced in under a week on a small 200 core cluster. DMPfold provides models for 16% of human proteome UniProt entries without structures, generates accurate models with fewer than 100 sequences in some cases, and is freely available.Comment: JGG and SMK contributed equally to the wor

    The first 100 days: modeling the evolution of the COVID-19 pandemic

    Full text link
    A simple analytical model for modeling the evolution of the 2020 COVID-19 pandemic is presented. The model is based on the numerical solution of the widely used Susceptible-Infectious-Removed (SIR) populations model for describing epidemics. We consider an expanded version of the original Kermack-McKendrick model, which includes a decaying value of the parameter β\beta (the effective contact rate) due to externally imposed conditions, to which we refer as the forced-SIR (FSIR) model. We introduce an approximate analytical solution to the differential equations that represent the FSIR model which gives very reasonable fits to real data for a number of countries over a period of 100 days (from the first onset of exponential increase, in China). The proposed model contains 3 adjustable parameters which are obtained by fitting actual data (up to April 28, 2020). We analyze these results to infer the physical meaning of the parameters involved. We use the model to make predictions about the total expected number of infections in each country as well as the date when the number of infections will have reached 99% of this total. We also compare key findings of the model with recently reported results on the high contagiousness and rapid spread of the disease.Comment: 17 pages, 5 figure

    Graph Neural Networks Meet Neural-Symbolic Computing: A Survey and Perspective

    Get PDF
    Neural-symbolic computing has now become the subject of interest of both academic and industry research laboratories. Graph Neural Networks (GNN) have been widely used in relational and symbolic domains, with widespread application of GNNs in combinatorial optimization, constraint satisfaction, relational reasoning and other scientific domains. The need for improved explainability, interpretability and trust of AI systems in general demands principled methodologies, as suggested by neural-symbolic computing. In this paper, we review the state-of-the-art on the use of GNNs as a model of neural-symbolic computing. This includes the application of GNNs in several domains as well as its relationship to current developments in neural-symbolic computing.Comment: Updated version, draft of accepted IJCAI2020 Survey Pape
    corecore