868 research outputs found

    Advances in image processing for single-particle analysis by electron cryomicroscopy and challenges ahead

    Get PDF
    Electron cryomicroscopy (cryo-EM) is essential for the study and functional understanding of non-crystalline macromolecules such as proteins. These molecules cannot be imaged using X-ray crystallography or other popular methods. CryoEM has been successfully used to visualize molecules such as ribosomes, viruses, and ion channels, for example. Obtaining structural models of these at various conformational states leads to insight on how these molecules function. Recent advances in imaging technology have given cryo-EM a scientific rebirth. Because of imaging improvements, image processing and analysis of the resultant images have increased the resolution such that molecular structures can be resolved at the atomic level. Cryo-EM is ripe with stimulating image processing challenges. In this article, we will touch on the most essential in order to build an accurate structural three-dimensional model from noisy projection images. Traditional approaches, such as k-means clustering for class averaging, will be provided as background. With this review, however, we will highlight fresh approaches from new and varied angles for each image processing sub-problem, including a 3D reconstruction method for asymmetric molecules using just two projection images and deep learning algorithms for automated particle picking. Keywords: Cryo-electron microscopy, Single Particle Analysis, Image processing algorithms

    Automatic post-picking improves particle image detection from Cryo-EM micrographs

    Full text link
    Cryo-electron microscopy (cryo-EM) studies using single particle reconstruction is extensively used to reveal structural information of macromolecular complexes. Aiming at the highest achievable resolution, state of the art electron microscopes acquire thousands of high-quality images. Having collected these data, each single particle must be detected and windowed out. Several fully- or semi-automated approaches have been developed for the selection of particle images from digitized micrographs. However they still require laborious manual post processing, which will become the major bottleneck for next generation of electron microscopes. Instead of focusing on improvements in automated particle selection from micrographs, we propose a post-picking step for classifying small windowed images, which are output by common picking software. A supervised strategy for the classification of windowed micrograph images into particles and non-particles reduces the manual workload by orders of magnitude. The method builds on new powerful image features, and the proper training of an ensemble classifier. A few hundred training samples are enough to achieve a human-like classification performance.Comment: 14 pages, 5 figure

    Contributions To Automatic Particle Identification In Electron Micrographs: Algorithms, Implementation, And Applications

    Get PDF
    Three dimensional reconstruction of large macromolecules like viruses at resolutions below 8 Ã… - 10 Ã… requires a large set of projection images and the particle identification step becomes a bottleneck. Several automatic and semi-automatic particle detection algorithms have been developed along the years. We present a general technique designed to automatically identify the projection images of particles. The method utilizes Markov random field modelling of the projected images and involves a preprocessing of electron micrographs followed by image segmentation and post processing for boxing of the particle projections. Due to the typically extensive computational requirements for extracting hundreds of thousands of particle projections, parallel processing becomes essential. We present parallel algorithms and load balancing schemes for our algorithms. The lack of a standard benchmark for relative performance analysis of particle identification algorithms has prompted us to develop a benchmark suite. Further, we present a collection of metrics for the relative performance analysis of particle identification algorithms on the micrograph images in the suite, and discuss the design of the benchmark suite

    Novel computational methods for in vitro and in situ cryo-electron microscopy

    Get PDF
    Over the past decade, advances in microscope hardware and image data processing algorithms have made cryo-electron microscopy (cryo-EM) a dominant technique for protein structure determination. Near-atomic resolution can now be obtained for many challenging in vitro samples using single-particle analysis (SPA), while sub-tomogram averaging (STA) can obtain sub-nanometer resolution for large protein complexes in a crowded cellular environment. Reaching high resolution requires large amounts of im-age data. Modern transmission electron microscopes (TEMs) automate the acquisition process and can acquire thousands of micrographs or hundreds of tomographic tilt se-ries over several days without intervention. In a first step, the data must be pre-processed: Micrographs acquired as movies are cor-rected for stage and beam-induced motion. For tilt series, additional alignment of all micrographs in 3D is performed using gold- or patch-based fiducials. Parameters of the contrast-transfer function (CTF) are estimated to enable its reversal during SPA refine-ment. Finally, individual protein particles must be located and extracted from the aligned micrographs. Current pre-processing algorithms, especially those for particle picking, are not robust enough to enable fully unsupervised operation. Thus, pre-processing is start-ed after data collection, and takes several days due to the amount of supervision re-quired. Pre-processing the data in parallel to acquisition with more robust algorithms would save time and allow to discover bad samples and microscope settings early on. Warp is a new software for cryo-EM data pre-processing. It implements new algorithms for motion correction, CTF estimation, tomogram reconstruction, as well as deep learn-ing-based approaches to particle picking and image denoising. The algorithms are more accurate and robust, enabling unsupervised operation. Warp integrates all pre-processing steps into a pipeline that is executed on-the-fly during data collection. Inte-grated with SPA tools, the pipeline can produce 2D and 3D classes less than an hour into data collection for favorable samples. Here I describe the implementation of the new algorithms, and evaluate them on various movie and tilt series data sets. I show that un-supervised pre-processing of a tilted influenza hemagglutinin trimer sample with Warp and refinement in cryoSPARC can improve previously published resolution from 3.9 Å to 3.2 Å. Warp’s algorithms operate in a reference-free manner to improve the image resolution at the pre-processing stage when no high-resolution maps are available for the particles yet. Once 3D maps have been refined, they can be used to go back to the raw data and perform reference-based refinement of sample motion and CTF in movies and tilt series. M is a new tool I developed to solve this task in a multi-particle framework. Instead of following the SPA assumption that every particle is single and independent, M models all particles in a field of view as parts of a large, physically connected multi-particle system. This allows M to optimize hyper-parameters of the system, such as sample motion and deformation, or higher-order aberrations in the CTF. Because M models these effects accurately and optimizes all hyper-parameters simultaneously with particle alignments, it can surpass previous reference-based frame and tilt series alignment tools. Here I de-scribe the implementation of M, evaluate it on several data sets, and demonstrate that the new algorithms achieve equally high resolution with movie and tilt series data of the same sample. Most strikingly, the combination of Warp, RELION and M can resolve 70S ribosomes bound to an antibiotic at 3.5 Å inside vitrified Mycoplasma pneumoniae cells, marking a major advance in resolution for in situ imaging

    Advances in xmipp for cryo-electron microscopy: From xmipp to scipion

    Get PDF
    Xmipp is an open-source software package consisting of multiple programs for processing data originating from electron microscopy and electron tomography, designed and managed by the Biocomputing Unit of the Spanish National Center for Biotechnology, although with contributions from many other developers over the world. During its 25 years of existence, Xmipp underwent multiple changes and updates. While there were many publications related to new programs and functionality added to Xmipp, there is no single publication on the Xmipp as a package since 2013. In this article, we give an overview of the changes and new work since 2013, describe technologies and techniques used during the development, and take a peek at the future of the package

    Statistical Reconstruction Methods for 3D Imaging of Biological Samples with Electron Microscopy

    Get PDF
    Electron microscopy has emerged as the leading method for the in vivo study of biological structures such as cells, organelles, protein molecules and virus like particles. By providing 3D images up to near atomic resolution, it plays a significant role in analyzing complex organizations, understanding physiological functions and developing medicines. The 3D images representing the electrostatic potential distribution are reconstructed by utilizing the 2D projection images of the target acquired by electron microscope. There are two main 3D reconstruction techniques in the field of electron microscopy: electron tomography (ET) and single particle reconstruction (SPR). In ET, the projection images are acquired by rotating the specimen for different angles. In SPR, the projection images are obtained by analyzing the images of multiple objects representing the same structure. Then, the tomographic reconstruction methods are applied in both methods to obtain the 3D image through the 2D projections.Physical and mechanical limitations can prevent to acquire projection images that cover the projection angle space completely and uniformly. Incomplete and non-uniform sampling of the projection angles results in anisotropic resolution in the image plane and generates artifacts. Another problem is that the total applied dose of electrons is limited in order to prevent the radiation damage to the biological target. Therefore, limited number of projection images with low signal to noise ratio can be used in the reconstruction process. This affects the resolution of the reconstructed image significantly. This study presents statistical methods to overcome these major challenges to obtain precise and high resolution images in electron microscopy.Statistical image reconstruction methods have been successful in recovering a signal from imperfect measurements due to their capability of utilizing a priori information. First, we developed a sequential application of a statistical method for ET. Then we extended the method to support projection angles freely distributed in 3D space and applied the method in SPR. In both applications, we observed the strength of the method in projection gap filling, robustness against noise, and resolving the high resolution details in comparison with the conventional reconstruction methods. Afterwards, we improved the method in terms of computation time by incorporating multiresolution reconstruction. Furthermore, we developed an adaptive regularization method to minimize the parameters required to be set by the user. We also proposed the local adaptive Wiener filter for the class averaging step of SPR to improve the averaging accuracy.The qualitative and quantitative analysis of the reconstructions with phantom and experimental datasets has demonstrated that the proposed reconstruction methods outperform the conventional reconstruction methods. These statistical approaches provided better image accuracy and higher resolution compared with the conventional algebraic and transfer domain based reconstruction methods. The methods provided in this study contribute to enhance our understanding of cellular and molecular structures by providing 3D images of those with improved accuracy and resolution

    Aceleración de algoritmos de procesamiento de imágenes para el análisis de partículas individuales con microscopia electrónica

    Full text link
    Tesis Doctoral inédita cotutelada por la Masaryk University (República Checa) y la Universidad Autónoma de Madrid, Escuela Politécnica Superior, Departamento de Ingeniería Informática. Fecha de Lectura: 24-10-2022Cryogenic Electron Microscopy (Cryo-EM) is a vital field in current structural biology. Unlike X-ray crystallography and Nuclear Magnetic Resonance, it can be used to analyze membrane proteins and other samples with overlapping spectral peaks. However, one of the significant limitations of Cryo-EM is the computational complexity. Modern electron microscopes can produce terabytes of data per single session, from which hundreds of thousands of particles must be extracted and processed to obtain a near-atomic resolution of the original sample. Many existing software solutions use high-Performance Computing (HPC) techniques to bring these computations to the realm of practical usability. The common approach to acceleration is parallelization of the processing, but in praxis, we face many complications, such as problem decomposition, data distribution, load scheduling, balancing, and synchronization. Utilization of various accelerators further complicates the situation, as heterogeneous hardware brings additional caveats, for example, limited portability, under-utilization due to synchronization, and sub-optimal code performance due to missing specialization. This dissertation, structured as a compendium of articles, aims to improve the algorithms used in Cryo-EM, esp. the SPA (Single Particle Analysis). We focus on the single-node performance optimizations, using the techniques either available or developed in the HPC field, such as heterogeneous computing or autotuning, which potentially needs the formulation of novel algorithms. The secondary goal of the dissertation is to identify the limitations of state-of-the-art HPC techniques. Since the Cryo-EM pipeline consists of multiple distinct steps targetting different types of data, there is no single bottleneck to be solved. As such, the presented articles show a holistic approach to performance optimization. First, we give details on the GPU acceleration of the specific programs. The achieved speedup is due to the higher performance of the GPU, adjustments of the original algorithm to it, and application of the novel algorithms. More specifically, we provide implementation details of programs for movie alignment, 2D classification, and 3D reconstruction that have been sped up by order of magnitude compared to their original multi-CPU implementation or sufficiently the be used on-the-fly. In addition to these three programs, multiple other programs from an actively used, open-source software package XMIPP have been accelerated and improved. Second, we discuss our contribution to HPC in the form of autotuning. Autotuning is the ability of software to adapt to a changing environment, i.e., input or executing hardware. Towards that goal, we present cuFFTAdvisor, a tool that proposes and, through autotuning, finds the best configuration of the cuFFT library for given constraints of input size and plan settings. We also introduce a benchmark set of ten autotunable kernels for important computational problems implemented in OpenCL or CUDA, together with the introduction of complex dynamic autotuning to the KTT tool. Third, we propose an image processing framework Umpalumpa, which combines a task-based runtime system, data-centric architecture, and dynamic autotuning. The proposed framework allows for writing complex workflows which automatically use available HW resources and adjust to different HW and data but at the same time are easy to maintainThe project that gave rise to these results received the support of a fellowship from the “la Caixa” Foundation (ID 100010434). The fellowship code is LCF/BQ/DI18/11660021. This project has received funding from the European Union’s Horizon 2020 research and innovation programme under the Marie Skłodowska-Curie grant agreement No. 71367

    Single particle 2D Electron crystallography for membrane protein structure determination

    Get PDF
    Proteins embedded into or attached to the cellular membrane perform crucial biological functions. Despite such importance, they remain among the most challenging targets of structural biology. Dedicated methods for membrane protein structure determination have been devised since decades, however with only partial success if compared to soluble proteins. One of these methods is 2D electron crystallography, in which the proteins are periodically arranged into a lipid bilayer. Using transmission electron microscopy to acquire projection images of samples containing such 2D crystals, which are embedded into a thin vitreous ice layer for radiation protection (cryo-EM), computer algorithms can be used to generate a 3D reconstruction of the protein. Unfortunately, in nearly every case, the 2D crystals are not flat and ordered enough to yield high-resolution reconstructions. Single particle analysis, on the other hand, is a technique that aligns projections of proteins isolated in solution in order to obtain a 3D reconstruction with a high success rate in terms of high resolution structures. In this thesis, we couple 2D crystal data processing with single particle analysis algorithms in order to perform a local correction of crystal distortions. We show that this approach not only allows reconstructions of much higher resolution than expected from the diffraction patterns obtained, but also reveals the existence of conformational heterogeneity within the 2D crystals. This structural variability can be linked to protein function, providing novel mechanistic insights and an explanation for why 2D crystals do not diffract to high resolution, in general. We present the computational methods that enable this hybrid approach, as well as other tools that aid several steps of cryo-EM data processing, from storage to postprocessing

    Structural studies of the 26S proteasome and its interaction with Ubp6 by cryo-electron microscopy

    No full text

    Statistical analysis and modeling for biomolecular structures

    Get PDF
    Most of the recent studies on biomolecules address their three dimensional structure since it is closely related to their functions in a biological system. Determination of structure of biomolecules can be done by using various methods, which rely on data from various experimental instruments or on computational approaches to previously obtained data or datasets. Single particle reconstruction using electron microscopic images of macromolecules has proven resource-wise to be useful and affordable for determining their molecular structure in increasing details. The main goal of this thesis is to contribute to the single particle reconstruction methodology, by adding a process of denoising in the analysis of the cryo-electron microscopic images. First, the denoising methods are briefly surveyed and their efficiencies for filtering cryo-electron microscopic images are evaluated. In this thesis, the focus has been set to information theoretic minimum description length (MDL) principle for coding efficiently the essential part of the signal. This approach can also be applied to reduce noise in signals and here it is used to develop a novel denoising method for cryo-electron microscopic images. An existing denoising method has been modified to suit the given problem in single particle reconstruction. In addition, a more general denoising method has been developed, discovering a novel way to find model class by using the MDL principle. This method was then thoroughly tested and compared with co-existing methods in order to evaluate the utility of denoising in single particle reconstruction. A secondary goal in the research for this thesis deals with studying protein oligomerisation, using computational approaches. The focus has been to recognize interacting residues in proteins for oligomerization and to model the interaction site for hantavirus N-protein. In order to unravel the interaction structure, the approach has been to understand the phenomenon of protein folding towards quaternary structure.reviewe
    corecore