72 research outputs found

    GPU accelerated maximum cardinality matching algorithms for bipartite graphs

    Get PDF
    We design, implement, and evaluate GPU-based algorithms for the maximum cardinality matching problem in bipartite graphs. Such algorithms have a variety of applications in computer science, scientific computing, bioinformatics, and other areas. To the best of our knowledge, ours is the first study which focuses on GPU implementation of the maximum cardinality matching algorithms. We compare the proposed algorithms with serial and multicore implementations from the literature on a large set of real-life problems where in majority of the cases one of our GPU-accelerated algorithms is demonstrated to be faster than both the sequential and multicore implementations.Comment: 14 pages, 5 figure

    Computing Maximum Cardinality Matchings in Parallel on Bipartite Graphs via Tree-Grafting

    Full text link

    eXplainable data processing

    Get PDF
    Seminario realizado en U & P U Patel Department of Computer Engineering, Chandubhai S. Patel Institute of Technology, Charotar University of Science And Technology (CHARUSAT), Changa-388421, Gujarat, India 2021[EN]Deep Learning y has created many new opportunities, it has unfortunately also become a means for achieving ill-intentioned goals. Fake news, disinformation campaigns, and manipulated images and videos have plagued the internet which has had serious consequences on our society. The myriad of information available online means that it may be difficult to distinguish between true and fake news, leading many users to unknowingly share fake news, contributing to the spread of misinformation. The use of Deep Learning to create fake images and videos has become known as deepfake. This means that there are ever more effective and realistic forms of deception on the internet, making it more difficult for internet users to distinguish reality from fictio

    The 4th Conference of PhD Students in Computer Science

    Get PDF

    Optimizing de novo transcriptome assembly from short-read RNA-Seq data: a comparative study

    Get PDF
    With the fast advances in nextgen sequencing technology, high-throughput RNA sequencing has emerged as a powerful and cost-effective way for transcriptome study. De novo assembly of transcripts provides an important solution to transcriptome analysis for organisms with no reference genome. However, there lacked understanding on how the different variables affected assembly outcomes, and there was no consensus on how to approach an optimal solution by selecting software tool and suitable strategy based on the properties of RNA-Seq data. To reveal the performance of different programs for transcriptome assembly, this work analyzed some important factors, including k-mer values, genome complexity, coverage depth, directional reads, etc. Seven program conditions, four single k-mer assemblers (SK: SOAPdenovo, ABySS, Oases and Trinity) and three multiple k-mer methods (MK: SOAPdenovo-MK, trans-ABySS and Oases-MK) were tested. While small and large k-mer values performed better for reconstructing lowly and highly expressed transcripts, respectively, MK strategy worked well for almost all ranges of expression quintiles. Among SK tools, Trinity performed well across various conditions but took the longest running time. Oases consumed the most memory whereas SOAPdenovo required the shortest runtime but worked poorly to reconstruct full-length CDS. ABySS showed some good balance between resource usage and quality of assemblies. Our work compared the performance of publicly available transcriptome assemblers, and analyzed important factors affecting de novo assembly. Some practical guidelines for transcript reconstruction from short-read RNA-Seq data were proposed. De novo assembly of C. sinensis transcriptome was greatly improved using some optimized methods

    Modeling and Simulation in Engineering

    Get PDF
    This book provides an open platform to establish and share knowledge developed by scholars, scientists, and engineers from all over the world, about various applications of the modeling and simulation in the design process of products, in various engineering fields. The book consists of 12 chapters arranged in two sections (3D Modeling and Virtual Prototyping), reflecting the multidimensionality of applications related to modeling and simulation. Some of the most recent modeling and simulation techniques, as well as some of the most accurate and sophisticated software in treating complex systems, are applied. All the original contributions in this book are jointed by the basic principle of a successful modeling and simulation process: as complex as necessary, and as simple as possible. The idea is to manipulate the simplifying assumptions in a way that reduces the complexity of the model (in order to make a real-time simulation), but without altering the precision of the results

    NASA Tech Briefs, July 1999

    Get PDF
    Topics: Test and Measurement; Electronic Components and Circuits; Electronic Systems; Physical Sciences; Materials; Computer Software; Mechanics; Machinery/Automation; Bio-Medical; Books and Reports; Semiconductors/ICs

    Design automation in synthetic biology : a dual evolutionary strategy

    Get PDF
    PhD ThesisSynthetic biology o ers a new horizon in designing complex systems. However, unprecedented complexity hinders the development of biological systems to its full potential. Mitigating complexity via adopting design principles from engineering and computer science elds has resulted in some success. For example, modularisation to foster reuse of design elements, and using computer assisted design tools have helped contain complexity to an extent. Nevertheless, these design practices are still limited, due to their heavy dependence on rational decision making by human designers. The issue with rational design approaches here arises from the challenging nature of dealing with highly complex biological systems of which we currently do not have complete understanding. Systematic processes that can algorithmically nd design solutions would be better able to cope with uncertainties posed by high levels of design complexity. A new framework for enabling design automation in synthetic biology was investigated. The framework works by projecting design problems into search problems, and by searching for design solutions based on the dual-evolutionary approach to combine the respective power of design domains in vivo and in silico. Proof-of-concept ideas, software, and hardware were developed to exemplify key technologies necessary in realising the dual evolutionary approach. Some of the areas investigated as part of this research included single-cell-level micro uidics, programmatic data collection, processing and analysis, molecular devices supporting solution search in vivo, and mathematical modelling. These somewhat eclectic collection of research themes were shown to work together to provide necessary means with which to design and characterise biological systems in a systematic fashion
    • …
    corecore