21,229 research outputs found
Fast automated cell phenotype image classification
BACKGROUND: The genomic revolution has led to rapid growth in sequencing of genes and proteins, and attention is now turning to the function of the encoded proteins. In this respect, microscope imaging of a protein's sub-cellular localisation is proving invaluable, and recent advances in automated fluorescent microscopy allow protein localisations to be imaged in high throughput. Hence there is a need for large scale automated computational techniques to efficiently quantify, distinguish and classify sub-cellular images. While image statistics have proved highly successful in distinguishing localisation, commonly used measures suffer from being relatively slow to compute, and often require cells to be individually selected from experimental images, thus limiting both throughput and the range of potential applications. Here we introduce threshold adjacency statistics, the essence which is to threshold the image and to count the number of above threshold pixels with a given number of above threshold pixels adjacent. These novel measures are shown to distinguish and classify images of distinct sub-cellular localization with high speed and accuracy without image cropping. RESULTS: Threshold adjacency statistics are applied to classification of protein sub-cellular localization images. They are tested on two image sets (available for download), one for which fluorescently tagged proteins are endogenously expressed in 10 sub-cellular locations, and another for which proteins are transfected into 11 locations. For each image set, a support vector machine was trained and tested. Classification accuracies of 94.4% and 86.6% are obtained on the endogenous and transfected sets, respectively. Threshold adjacency statistics are found to provide comparable or higher accuracy than other commonly used statistics while being an order of magnitude faster to calculate. Further, threshold adjacency statistics in combination with Haralick measures give accuracies of 98.2% and 93.2% on the endogenous and transfected sets, respectively. CONCLUSION: Threshold adjacency statistics have the potential to greatly extend the scale and range of applications of image statistics in computational image analysis. They remove the need for cropping of individual cells from images, and are an order of magnitude faster to calculate than other commonly used statistics while providing comparable or better classification accuracy, both essential requirements for application to large-scale approaches
From Cellular Characteristics to Disease Diagnosis: Uncovering Phenotypes with Supercells
Cell heterogeneity and the inherent complexity due to the interplay of multiple molecular processes within the cell pose difficult challenges for current single-cell biology. We introduce an approach that identifies a disease phenotype from multiparameter single-cell measurements, which is based on the concept of ‘‘supercell statistics’’, a single-cell-based averaging procedure followed by a machine learning classification scheme. We are able to assess the optimal tradeoff between the number of single cells averaged and the number of measurements needed to capture phenotypic differences between healthy and diseased patients, as well as between different diseases that are difficult to diagnose otherwise. We apply our approach to two kinds of single-cell datasets, addressing the diagnosis of a premature aging disorder using images of cell nuclei, as well as the phenotypes of two non-infectious uveitides (the ocular manifestations of Behc¸et’s disease and sarcoidosis) based on multicolor flow cytometry. In the former case, one nuclear shape measurement taken over a group of 30 cells is sufficient to classify samples as healthy or diseased, in agreement with usual laboratory practice. In the latter, our method is able to identify a minimal set of 5 markers that accurately predict Behc¸et’s disease and sarcoidosis. This is the first time that a quantitative phenotypic distinction between these two diseases has been achieved. To obtain this clear phenotypic signature, about one hundred CD8+ T cells need to be measured. Although the molecular markers identified have been reported to be important players in autoimmune disorders, this is the first report pointing out that CD8+ T cells can be used to distinguish two systemic inflammatory diseases. Beyond these specific cases, the approach proposed here is applicable to datasets generated by other kinds of state-of-the-art and forthcoming single-cell technologies, such as multidimensional mass cytometry, single-cell gene expression, and single-cell full genome sequencing techniques.Fil: Candia, Julian Marcelo. University of Maryland; Estados Unidos. Consejo Nacional de Investigaciones Científicas y Técnicas. Centro Científico Tecnológico Conicet - La Plata. Instituto de Física de Líquidos y Sistemas Biológicos. Universidad Nacional de La Plata. Facultad de Ciencias Exactas. Instituto de Física de Líquidos y Sistemas Biológicos; ArgentinaFil: Maunu, Ryan. University of Maryland; Estados UnidosFil: Driscoll, Meghan. University of Maryland; Estados UnidosFil: Biancotto, Angélique. National Institutes of Health; Estados UnidosFil: Dagur, Pradeep. National Institutes of Health; Estados UnidosFil: McCoy Jr., J Philip. National Institutes of Health; Estados UnidosFil: Nida Sen, H.. National Institutes of Health; Estados UnidosFil: Wei, Lai. National Institutes of Health; Estados UnidosFil: Maritan, Amos. Università di Padova; ItaliaFil: Cao, Kan. University of Maryland; Estados UnidosFil: Nussenblatt, Robert B. National Institutes of Health; Estados UnidosFil: Banavar, Jayanth R.. University of Maryland; Estados UnidosFil: Losert, Wolfgang. University of Maryland; Estados Unido
Improve the performance of transfer learning without fine-tuning using dissimilarity-based multi-view learning for breast cancer histology images
Breast cancer is one of the most common types of cancer and leading
cancer-related death causes for women. In the context of ICIAR 2018 Grand
Challenge on Breast Cancer Histology Images, we compare one handcrafted feature
extractor and five transfer learning feature extractors based on deep learning.
We find out that the deep learning networks pretrained on ImageNet have better
performance than the popular handcrafted features used for breast cancer
histology images. The best feature extractor achieves an average accuracy of
79.30%. To improve the classification performance, a random forest
dissimilarity based integration method is used to combine different feature
groups together. When the five deep learning feature groups are combined, the
average accuracy is improved to 82.90% (best accuracy 85.00%). When handcrafted
features are combined with the five deep learning feature groups, the average
accuracy is improved to 87.10% (best accuracy 93.00%)
Toward high-content/high-throughput imaging and analysis of embryonic morphogenesis
In vivo study of embryonic morphogenesis tremendously benefits from recent advances in live microscopy and computational analyses. Quantitative and automated investigation of morphogenetic processes opens the field to high-content and high-throughput strategies. Following experimental workflow currently developed in cell biology, we identify the key challenges for applying such strategies in developmental biology. We review the recent progress in embryo preparation and manipulation, live imaging, data registration, image segmentation, feature computation, and data mining dedicated to the study of embryonic morphogenesis. We discuss a selection of pioneering studies that tackled the current methodological bottlenecks and illustrated the investigation of morphogenetic processes in vivo using quantitative and automated imaging and analysis of hundreds or thousands of cells simultaneously, paving the way for high-content/high-throughput strategies and systems analysis of embryonic morphogenesis
Recommended from our members
A Rapid Segmentation-Insensitive "Digital Biopsy" Method for Radiomic Feature Extraction: Method and Pilot Study Using CT Images of Non-Small Cell Lung Cancer.
Quantitative imaging approaches compute features within images' regions of interest. Segmentation is rarely completely automatic, requiring time-consuming editing by experts. We propose a new paradigm, called "digital biopsy," that allows for the collection of intensity- and texture-based features from these regions at least 1 order of magnitude faster than the current manual or semiautomated methods. A radiologist reviewed automated segmentations of lung nodules from 100 preoperative volume computed tomography scans of patients with non-small cell lung cancer, and manually adjusted the nodule boundaries in each section, to be used as a reference standard, requiring up to 45 minutes per nodule. We also asked a different expert to generate a digital biopsy for each patient using a paintbrush tool to paint a contiguous region of each tumor over multiple cross-sections, a procedure that required an average of <3 minutes per nodule. We simulated additional digital biopsies using morphological procedures. Finally, we compared the features extracted from these digital biopsies with our reference standard using intraclass correlation coefficient (ICC) to characterize robustness. Comparing the reference standard segmentations to our digital biopsies, we found that 84/94 features had an ICC >0.7; comparing erosions and dilations, using a sphere of 1.5-mm radius, of our digital biopsies to the reference standard segmentations resulted in 41/94 and 53/94 features, respectively, with ICCs >0.7. We conclude that many intensity- and texture-based features remain consistent between the reference standard and our method while substantially reducing the amount of operator time required
Keeping track of worm trackers
C. elegans is used extensively as a model system in the neurosciences due to its well defined nervous system. However, the seeming simplicity of this nervous system in anatomical structure and neuronal connectivity, at least compared to higher animals, underlies a rich diversity of behaviors. The usefulness of the worm in genome-wide mutagenesis or RNAi screens, where thousands of strains are assessed for phenotype, emphasizes the need for computational methods for automated parameterization of generated behaviors. In addition, behaviors can be modulated upon external cues like temperature, O2 and CO2 concentrations, mechanosensory and chemosensory inputs. Different machine vision tools have been developed to aid researchers in their efforts to inventory and characterize defined behavioral “outputs”. Here we aim at providing an overview of different worm-tracking packages or video analysis tools designed to quantify different aspects of locomotion such as the occurrence of directional changes (turns, omega bends), curvature of the sinusoidal shape (amplitude, body bend angles) and velocity (speed, backward or forward movement)
- …