5,379 research outputs found
Efficient maintenance and update of nonbonded lists in macromolecular simulations
Molecular mechanics and dynamics simulations use distance based cutoff approximations for faster computation of pairwise van der Waals and electrostatic energy terms. These approximations traditionally use a precalculated and periodically updated list of interacting atom pairs, known as the “nonbonded neighborhood lists” or nblists, in order to reduce the overhead of finding atom pairs that are within distance cutoff. The size of nblists grows linearly with the number of atoms in the system and superlinearly with the distance cutoff, and as a result, they require significant amount of memory for large molecular systems. The high space usage leads to poor cache performance, which slows computation for large distance cutoffs. Also, the high cost of updates means that one cannot afford to keep the data structure always synchronized with the configuration of the molecules when efficiency is at stake. We propose a dynamic octree data structure for implicit maintenance of nblists using space linear in the number of atoms but independent of the distance cutoff. The list can be updated very efficiently as the coordinates of atoms change during the simulation. Unlike explicit nblists, a single octree works for all distance cutoffs. In addition, octree is a cache-friendly data structure, and hence, it is less prone to cache miss slowdowns on modern memory hierarchies than nblists. Octrees use almost 2 orders of magnitude less memory, which is crucial for simulation of large systems, and while they are comparable in performance to nblists when the distance cutoff is small, they outperform nblists for larger systems and large cutoffs. Our tests show that octree implementation is approximately 1.5 times faster in practical use case scenarios as compared to nblists
The Role of the Dielectric Barrier in Narrow Biological Channels: a Novel Composite Approach to Modeling Single-channel Currents
A composite continuum theory for calculating ion current through a protein channel of known structure is proposed, which incorporates information about the channel dynamics. The approach is utilized to predict current through the Gramicidin A ion channel, a narrow pore in which the applicability of conventional continuum theories is questionable. The proposed approach utilizes a modified version of Poisson-Nernst-Planck (PNP) theory, termed Potential-of-Mean-Force-Poisson-Nernst-Planck theory (PMFPNP), to compute ion currents. As in standard PNP, ion permeation is modeled as a continuum drift-diffusion process in a self-consistent electrostatic potential. In PMFPNP, however, information about the dynamic relaxation of the protein and the surrounding medium is incorporated into the model of ion permeation by including the free energy of inserting a single ion into the channel, i.e., the potential of mean force along the permeation pathway. In this way the dynamic flexibility of the channel environment is approximately accounted for. The PMF profile of the ion along the Gramicidin A channel is obtained by combining an equilibrium molecular dynamics (MD) simulation that samples dynamic protein configurations when an ion resides at a particular location in the channel with a continuum electrostatics calculation of the free energy. The diffusion coefficient of a potassium ion within the channel is also calculated using the MD trajectory. Therefore, except for a reasonable choice of dielectric constants, no direct fitting parameters enter into this model. The results of our study reveal that the channel response to the permeating ion produces significant electrostatic stabilization of the ion inside the channel. The dielectric self-energy of the ion remains essentially unchanged in the course of the MD simulation, indicating that no substantial changes in the protein geometry occur as the ion passes through it. Also, the model accounts for the experimentally observed saturation of ion current with increase of the electrolyte concentration, in contrast to the predictions of standard PNP theory
Many-Task Computing and Blue Waters
This report discusses many-task computing (MTC) generically and in the
context of the proposed Blue Waters systems, which is planned to be the largest
NSF-funded supercomputer when it begins production use in 2012. The aim of this
report is to inform the BW project about MTC, including understanding aspects
of MTC applications that can be used to characterize the domain and
understanding the implications of these aspects to middleware and policies.
Many MTC applications do not neatly fit the stereotypes of high-performance
computing (HPC) or high-throughput computing (HTC) applications. Like HTC
applications, by definition MTC applications are structured as graphs of
discrete tasks, with explicit input and output dependencies forming the graph
edges. However, MTC applications have significant features that distinguish
them from typical HTC applications. In particular, different engineering
constraints for hardware and software must be met in order to support these
applications. HTC applications have traditionally run on platforms such as
grids and clusters, through either workflow systems or parallel programming
systems. MTC applications, in contrast, will often demand a short time to
solution, may be communication intensive or data intensive, and may comprise
very short tasks. Therefore, hardware and software for MTC must be engineered
to support the additional communication and I/O and must minimize task dispatch
overheads. The hardware of large-scale HPC systems, with its high degree of
parallelism and support for intensive communication, is well suited for MTC
applications. However, HPC systems often lack a dynamic resource-provisioning
feature, are not ideal for task communication via the file system, and have an
I/O system that is not optimized for MTC-style applications. Hence, additional
software support is likely to be required to gain full benefit from the HPC
hardware
Recommended from our members
Improvements in Molecular Mechanics Sampling and Energy Models
The process of bringing drugs to market continues to be a slow and expensive affair. And despite recent advances in technology, the cost both in monetary terms and in terms of time between target identification and arrival of a new drug on the market continues to increase. High throughput screening is a first step towards testing a large number of possible bioactive compounds very quickly. However, the space of possible small molecules is limitless, and high throughput screening is limited both by the size of available libraries and the cost of running such a large number of experiments. Therefore, advancements in computational drug screening are necessary in order to maintain the current rate of progress in modern medicine.
Computational drug design, or computer assisted drug design, offers a possible way of addressing some of the shortfalls of conventional high throughput screening. Using computational methods, it is possible to estimate parameters such as binding affinity of any small molecule, even those not currently present in any small molecule library, without having to first invest in the often slow and expensive process of finding a synthetic pathway. Computational methods can be used to screen similar molecules, or mutations in small molecule space, seeking to increase binding affinity to the protein target, and thereby efficacy, while simultaneously minimizing binding affinity to other proteins, decreasing cross reactivity, and reducing toxicity and harmful side effects.Computational biology methods of drug research can be broadly classified in a number of different ways.
However, one of the most common classifications is according to the methods used to identify possible drug compounds and later optimize those leads. The first broad category is informatics or artificial intelligence based approaches. In these approaches, artificial intelligence methods such as neural networks, support vector machines, and qualitative structure-activity relationships (QSAR) are used to identify chemical or structural properties that contribute heavily to binding affinity.
The next category, ligand based approaches, is very useful when there are a large number of known binders for a specific family of proteins. In this approach, the ligands are clustered using a metric of chemical similarity and new compounds which occupy a similar chemical space are likely to also bind strongly with the protein of interest.
The final class of methods of computational drug design, and the method explored in this thesis, is the diverse class known as structural methods. These approaches in the most general sense make use of a sampling method to sample a number of protein, or protein-small-molecule interaction conformations and an energy model or scoring function to measure dimensions which would be very difficult and or expensive to measure experimentally. In this thesis, a number of different sampling methods that are applicable to different questions in computational biology are presented. Additionally, an improved algorithm for evaluating implicit solvent effects is presented, and a number of improvements in performance, reliability and utility of the molecular mechanics program used are discussed
Recent Trends in In-silico Drug Discovery
A Drug designing is a process in which new leads (potential drugs) are discovered which have therapeutic benefits in diseased condition. With development of various computational tools and availability of databases (having information about 3D structure of various molecules) discovery of drugs became comparatively, a faster process. The two major drug development methods are structure based drug designing and ligand based drug designing. Structure based methods try to make predictions based on three dimensional structure of the target molecules. The major approach of structure based drug designing is Molecular docking, a method based on several sampling algorithms and scoring functions. Docking can be performed in several ways depending upon whether ligand and receptors are rigid or flexible. Hotspot grafting, is another method of drug designing. It is preferred when the structure of a native binding protein and target protein complex is available and the hotspots on the interface are known. In absence of information of three Dimensional structure of target molecule, Ligand based methods are used. Two common methods used in ligand based drug designing are Pharmacophore modelling and QSAR. Pharmacophore modelling explains only essential features of an active ligand whereas QSAR model determines effect of certain property on activity of ligand. Fragment based drug designing is a de novo approach of building new lead compounds using fragments within the active site of the protein. All the candidate leads obtained by various drug designing method need to satisfy ADMET properties for its development as a drug. In-silico ADMET prediction tools have made ADMET profiling an easier and faster process. In this review, various softwares available for drug designing and ADMET property predictions have also been listed
Geometric modeling, simulation, and visualization methods for plasmid DNA molecules
Plasmid DNA molecules are a special type of DNA molecules that are used, among other applications,
in DNA vaccination and gene therapy. These molecules are characterized by, when in
their natural state, presenting a closed-circular conformation and by being supercoiled. The
production of plasmid DNA using bacteria as hosts implies a purification step where the plasmid
DNA molecules are separated from the DNA of the host and other contaminants. This purification
process, and all the physical and chemical variations involved, such as temperature
changes, may affect the plasmid DNA molecules conformation by uncoiling or even by open
them, which makes them useless for therapeutic applications. Because of that, researchers
are always searching for new purification techniques that maximize the amount of supercoiled
plasmid DNA that is produced. Computer simulations and 3D visualization of plasmid DNA can
bring many advantages because they allow researchers to actually see what can happen to the
molecules under certain conditions. In this sense, it was necessary to develop reliable and accurate
geometric models specific for plasmid DNA simulations. This dissertation presents a new
assembling algorithm for B-DNA specifically developed for plasmid DNA assembling. This new
assembling algorithm is completely adaptive in the sense that it allows researchers to assemble
any plasmid DNA base-pair sequence along any arbitrary conformation that fits the length
of the plasmid DNA molecule. This is specially suitable for plasmid DNA simulations, where
conformations are generated by simulation procedures and there is the need to assemble the
given base-pair sequence over that conformation, what can not be done by conventional predictive
DNA assembling methods. Unlike traditional molecular visualization methods that are
based on the atomic structure, this new assembling algorithm uses color coded 3D molecular
surfaces of the nucleotides as the building blocks for DNA assembling. This new approach, not
only reduces the amount of graphical objects and, consequently, makes the rendering faster,
but also makes it easier to visually identify the nucleotides in the DNA strands. The algorithm
used to triangulate the molecular surfaces of the nucleotides building blocks is also a novelty
presented as part of this dissertation. This new triangulation algorithm for Gaussian molecular
surfaces introduces a new mechanism that divides the atomic structure of molecules into boxes
and spheres. This new space division method is faster because it confines the local calculation
of the molecular surface to a specific region of influence of the atomic structure, not taking into
account atoms that do not influence the triangulation of the molecular surface in that region.
This new method also guarantees the continuity of the molecular surface. Having in mind that
the aim of this dissertation is to present a complete set of methods for plasmid DNA visualization
and simulation, it is also proposed a new deformation algorithm to be used for plasmid
DNA Monte Carlo simulations. This new deformation algorithm uses a 3D polyline to represent
the plasmid DNA conformation and performs small deformations on that polyline, keeping the
segments length and connectivity. Experiments have been performed in order to compare this
new deformation method with deformation methods traditionally used by Monte Carlo plasmid
DNA simulations These experiments shown that the new method is more efficient in the sense
that its trial acceptance ratio is higher and it converges sooner and faster to the elastic energy
equilibrium state of the plasmid DNA molecule. In sum, this dissertation successfully presents
an end-to-end set of models and algorithms for plasmid DNA geometric modelling, visualization
and simulation
Conformational Preferences of a 14-Residue Fibrillogenic Peptide from Acetylcholinesterase†
A 14-residue fragment from near the C-terminus of the enzyme acetylcholinesterase (AChE) is believed to have a neurotoxic/neurotrophic effect acting via an unknown pathway. While the peptide is α-helical in the full-length enzyme, the structure and association mechanism of the fragment are unknown. Using multiple molecular dynamics simulations, starting from a tetrameric complex of the association domain of AChE and systematicall disassembled subsets that include the peptide fragment, we show that the fragment is incapable of retaining its helicity in solution. Extensive replica exchange Monte Carlo folding and unfolding simulations in implicit solvent with capped and uncappted termini failed to converge to any consistent cluster of structures, suggesting that the fragment remains largely unstructured in solution under the conditions considered. Furthermore, extended molecular dynamics simulations of two steric zipper models show that the peptide is likely to form a zipper with antiparallel sheets and that peptides with mutations known to prevent fibril formation likely do so by interfering with this packing. The results demonstrate how the local environment of a peptide can stabilize a particular conformation
MOLNs: A cloud platform for interactive, reproducible and scalable spatial stochastic computational experiments in systems biology using PyURDME
Computational experiments using spatial stochastic simulations have led to
important new biological insights, but they require specialized tools, a
complex software stack, as well as large and scalable compute and data analysis
resources due to the large computational cost associated with Monte Carlo
computational workflows. The complexity of setting up and managing a
large-scale distributed computation environment to support productive and
reproducible modeling can be prohibitive for practitioners in systems biology.
This results in a barrier to the adoption of spatial stochastic simulation
tools, effectively limiting the type of biological questions addressed by
quantitative modeling. In this paper, we present PyURDME, a new, user-friendly
spatial modeling and simulation package, and MOLNs, a cloud computing appliance
for distributed simulation of stochastic reaction-diffusion models. MOLNs is
based on IPython and provides an interactive programming platform for
development of sharable and reproducible distributed parallel computational
experiments
- …