838 research outputs found
Homology sequence analysis using GPU acceleration
A number of problems in bioinformatics, systems biology and computational biology field require abstracting physical entities to mathematical or computational models. In such studies, the computational paradigms often involve algorithms that can be solved by the Central Processing Unit (CPU). Historically, those algorithms benefit from the advancements of computing power in the serial processing capabilities of individual CPU cores. However, the growth has slowed down over recent years, as scaling out CPU has been shown to be both cost-prohibitive and insecure. To overcome this problem, parallel computing approaches that employ the Graphics Processing Unit (GPU) have gained attention as complementing or replacing traditional CPU approaches. The premise of this research is to investigate the applicability of various parallel computing platforms to several problems in the detection and analysis of homology in biological sequence. I hypothesize that by exploiting the sheer amount of computation power and sequencing data, it is possible to deduce information from raw sequences without supplying the underlying prior knowledge to come up with an answer. I have developed such tools to perform analysis at scales that are traditionally unattainable with general-purpose CPU platforms. I have developed a method to accelerate sequence alignment on the GPU, and I used the method to investigate whether the Operational Taxonomic Unit (OTU) classification problem can be improved with such sheer amount of computational power. I have developed a method to accelerate pairwise k-mer comparison on the GPU, and I used the method to further develop PolyHomology, a framework to scaffold shared sequence motifs across large numbers of genomes to illuminate the structure of the regulatory network in yeasts. The results suggest that such approach to heterogeneous computing could help to answer questions in biology and is a viable path to new discoveries in the present and the future.Includes bibliographical reference
Annual Report 2005 - Institute of Nuclear and Hadron Physics
Preface The Forschungszentrum Rossendorf (FZR) at Dresden is a multidisciplinary research center within the Wissenschafts-Gemeinschaft G. W. Leibniz (WGL), one of the German agencies for extra-university research. The center is active in investigations on the structure of matter as well as in the life sciences and in environmental research. The Institute of Nuclear and Hadron Physics (IKH) within the FZR avails for its research the coupling of radiation to matter in subatomic dimensions as well as to tissue, to cells, and to their components. Its research in the field of Subatomic Physics is part of the FZR-program Structure of Matter and its investigations concerning the interaction of Biostructures and Radiation contribute to the bf Life Science program of the FZR. In this field the IKH exploits possibilities for transfer and introduction of experimental and theoretical techniques from particle and nuclear physics to projects in radiobiology and biophysics. Much of this kind of interdisciplinary transfer is connected to the Radiation Source ELBE at the FZR. With its superconducting accelerator for relativistic electrons this large installation provides photons in the wide wavelength range from fm to mm - i.e. bremsstrahlung for the investigation of photonuclear processes, hard X-rays for radiobiological and other studies and infrared light for research on the structural dynamics of biomolecules. The investigation of radiation-induced processes not only dominates the projects in nuclear astrophysics as pursued at ELBE, it also is a central theme of the experimental and theoretical research performed by the IKH in close connection to the heavy ion synchrotron SIS and the upcoming FAIR facility at Darmstadt. ELBE also will deliver compact bunches of secondary neutrons and fission fragments; both offer new possibilities in laboratory studies related to the cosmic breeding of the chemical elements thus complementing the astrophysics-motivated studies with bremsstrahlung photons..
Computational Processing and Biological Interpretation of Genomes and Metagenomes of Marine Bacteria
MR-CUDASW - GPU accelerated Smith-Waterman algorithm for medium-length (meta)genomic data
The idea of using a graphics processing unit (GPU) for more than simply graphic output purposes has been around for quite some time in scientific communities. However, it is only recently that its benefits for a range
of bioinformatics and life sciences compute-intensive tasks has been recognized. This thesis investigates the possibility of improving the performance of the overlap determination stage of an Overlap Layout Consensus
(OLC)-based assembler by using a GPU-based implementation of the Smith-Waterman algorithm.
In this thesis an existing GPU-accelerated sequence alignment algorithm is adapted and expanded to reduce its completion time. A number of improvements and changes are made to the original software. Workload distribution, query profile construction, and thread scheduling techniques implemented by the original program are replaced by custom methods specifically designed to handle medium-length reads.
Accordingly, this algorithm is the first highly parallel solution that has been specifically optimized to process medium-length nucleotide reads (DNA/RNA) from modern sequencing machines (i.e. Ion Torrent).
Results show that the software reaches up to 82 GCUPS (Giga Cell Updates Per Second) on a single-GPU graphic card running on a commodity desktop hardware. As a result it is the fastest GPU-based implemen-
tation of the Smith-Waterman algorithm tailored for processing medium-length nucleotide reads. Despite being designed for performing the Smith-Waterman algorithm on medium-length nucleotide sequences, this
program also presents great potential for improving heterogeneous computing with CUDA-enabled GPUs in general and is expected to make contributions to other research problems that require sensitive pairwise alignment to be applied to a large number of reads. Our results show that it is possible to improve the performance of bioinformatics algorithms by taking full advantage of the compute resources of the underlying commodity hardware and further, these results are especially encouraging since GPU performance grows faster than multi-core CPUs
Recommended from our members
Nevada Test Site-Directed Research and Development, FY 2007 Report
The Nevada Test Site-Directed Research and Development (SDRD) program completed a very successful year of research and development activities in FY 2007. Twenty-nine new projects were selected for funding this year, and eight projects started in FY 2006 were brought to conclusion. The total funds expended by the SDRD program were 153 thousand. An external audit conducted in September 2007 verified that appropriate accounting practices were applied to the SDRD program. Highlights for the year included: programmatic adoption of 8 SDRD-developed technologies; the filing of 9 invention disclosures for innovation evolving from SDRD projects; participation in the tri-Lab Laboratory Directed Research and Development (LDRD) and SDRD Symposium that was broadly attended by Nevada Test Site (NTS), National Nuclear Security Administration (NNSA), LDRD, U.S. Department of Homeland Security (DHS), and U.S. Department of Defense (DoD) representatives; peer reviews of all FY 2007 projects; and the successful completion of 37 R&D projects, as presented in this report. In response to a company-wide call, authors throughout the NTS complex submitted 182 proposals for FY 2007 SDRD projects. The SDRD program has seen a dramatic increase in the yearly total of submitted proposals--from 69 in FY 2002 to 182 this year--while the number of projects funded has actually decreased from a program high of 57 in FY 2004. The overall effect of this trend has helped ensure an increasingly competitive program that benefited from a broader set of innovative ideas, making project selection both challenging and rewarding. Proposals were evaluated for technical merit, including such factors as innovation, probability of success, potential benefit, and mission applicability. Authors and reviewers benefited from the use of a shortfalls list entitled the 'NTS Technology Needs Assessment' that was compiled from NTS, National Weapons Laboratory (NWL), and NNSA sources. This tool continues to be of considerable value in aligning the SDRD program with mission priorities, and was expanded in FY 2007 to include technology development needs from the DHS and other agencies with missions closely aligned to that of the NTS
- …