29 research outputs found
Large-Scale Automatic Reconstruction of Neuronal Processes from Electron Microscopy Images
Automated sample preparation and electron microscopy enables acquisition of
very large image data sets. These technical advances are of special importance
to the field of neuroanatomy, as 3D reconstructions of neuronal processes at
the nm scale can provide new insight into the fine grained structure of the
brain. Segmentation of large-scale electron microscopy data is the main
bottleneck in the analysis of these data sets. In this paper we present a
pipeline that provides state-of-the art reconstruction performance while
scaling to data sets in the GB-TB range. First, we train a random forest
classifier on interactive sparse user annotations. The classifier output is
combined with an anisotropic smoothing prior in a Conditional Random Field
framework to generate multiple segmentation hypotheses per image. These
segmentations are then combined into geometrically consistent 3D objects by
segmentation fusion. We provide qualitative and quantitative evaluation of the
automatic segmentation and demonstrate large-scale 3D reconstructions of
neuronal processes from a volume of brain
tissue over a cube of in each dimension corresponding to
1000 consecutive image sections. We also introduce Mojo, a proofreading tool
including semi-automated correction of merge errors based on sparse user
scribbles
A connectomic approach to the lateral geniculate nucleus
AbstractAlthough the core functions and structure of the lateral geniculate nucleus (LGN) are well understood, this core is surrounded by questions about the integration of feedforward and feedback connections, interactions between different channels of information, and how activity dependent development restructures synaptic networks. Our understanding of the organization of the mouse LGN is particularly limited given how important it has become as a model system. Advances in circuit scale electron microscopy (cellular connectomics) have made it possible to reconstruct the synaptic connectivity of hundreds of neurons within in a circuit the size of the mouse LGN. These circuit reconstructions can reveal cell type-to-cell type canonical wiring diagrams as well as the higher order wiring motifs that are only visible in reconstructions of intact networks. Connectomic analysis of the LGN therefore not only can answer longstanding questions about the organization of the visual thalamus but also presents unique opportunities for investigating fundamental properties of mammalian circuit formation.</jats:p
Guided Proofreading of Automatic Segmentations for Connectomics
Automatic cell image segmentation methods in connectomics produce merge and
split errors, which require correction through proofreading. Previous research
has identified the visual search for these errors as the bottleneck in
interactive proofreading. To aid error correction, we develop two classifiers
that automatically recommend candidate merges and splits to the user. These
classifiers use a convolutional neural network (CNN) that has been trained with
errors in automatic segmentations against expert-labeled ground truth. Our
classifiers detect potentially-erroneous regions by considering a large context
region around a segmentation boundary. Corrections can then be performed by a
user with yes/no decisions, which reduces variation of information 7.5x faster
than previous proofreading methods. We also present a fully-automatic mode that
uses a probability threshold to make merge/split decisions. Extensive
experiments using the automatic approach and comparing performance of novice
and expert users demonstrate that our method performs favorably against
state-of-the-art proofreading methods on different connectomics datasets.Comment: Supplemental material available at
http://rhoana.org/guidedproofreading/supplemental.pd
Convolutional nets for reconstructing neural circuits from brain images acquired by serial section electron microscopy
Neural circuits can be reconstructed from brain images acquired by serial
section electron microscopy. Image analysis has been performed by manual labor
for half a century, and efforts at automation date back almost as far.
Convolutional nets were first applied to neuronal boundary detection a dozen
years ago, and have now achieved impressive accuracy on clean images. Robust
handling of image defects is a major outstanding challenge. Convolutional nets
are also being employed for other tasks in neural circuit reconstruction:
finding synapses and identifying synaptic partners, extending or pruning
neuronal reconstructions, and aligning serial section images to create a 3D
image stack. Computational systems are being engineered to handle petavoxel
images of cubic millimeter brain volumes
An Automated Images-to-Graphs Framework for High Resolution Connectomics
Reconstructing a map of neuronal connectivity is a critical challenge in
contemporary neuroscience. Recent advances in high-throughput serial section
electron microscopy (EM) have produced massive 3D image volumes of nanoscale
brain tissue for the first time. The resolution of EM allows for individual
neurons and their synaptic connections to be directly observed. Recovering
neuronal networks by manually tracing each neuronal process at this scale is
unmanageable, and therefore researchers are developing automated image
processing modules. Thus far, state-of-the-art algorithms focus only on the
solution to a particular task (e.g., neuron segmentation or synapse
identification).
In this manuscript we present the first fully automated images-to-graphs
pipeline (i.e., a pipeline that begins with an imaged volume of neural tissue
and produces a brain graph without any human interaction). To evaluate overall
performance and select the best parameters and methods, we also develop a
metric to assess the quality of the output graphs. We evaluate a set of
algorithms and parameters, searching possible operating points to identify the
best available brain graph for our assessment metric. Finally, we deploy a
reference end-to-end version of the pipeline on a large, publicly available
data set. This provides a baseline result and framework for community analysis
and future algorithm development and testing. All code and data derivatives
have been made publicly available toward eventually unlocking new biofidelic
computational primitives and understanding of neuropathologies.Comment: 13 pages, first two authors contributed equally V2: Added additional
experiments and clarifications; added information on infrastructure and
pipeline environmen