510 research outputs found

    linkcomm: an R package for the generation, visualization, and analysis of link communities in networks of arbitrary size and type

    Get PDF
    Summary: An essential element when analysing the structure, function, and dynamics of biological networks is the identification of communities of related nodes. An algorithm proposed recently enhances this process by clustering the links between nodes, rather than the nodes themselves, thereby allowing each node to belong to multiple overlapping or nested communities. The R package ‘linkcomm’ implements this algorithm and extends it in several aspects: (i) the clustering algorithm handles networks that are weighted, directed, or both weighted and directed; (ii) several visualization methods are implemented that facilitate the representation of the link communities and their relationships; (iii) a suite of functions are included for the downstream analysis of the link communities including novel community-based measures of node centrality; (iv) the main algorithm is written in C++ and designed to handle networks of any size; and (v) several clustering methods are available for networks that can be handled in memory, and the number of communities can be adjusted by the user

    An automated workflow for parallel processing of large multiview SPIM recordings

    Get PDF
    Multiview light sheet fluorescence microscopy (LSFM) allows to image developing organisms in 3D at unprecedented temporal resolution over long periods of time. The resulting massive amounts of raw image data requires extensive processing interactively via dedicated graphical user interface (GUI) applications. The consecutive processing steps can be easily automated and the individual time points can be processed independently, which lends itself to trivial parallelization on a high performance cluster (HPC). Here we introduce an automated workflow for processing large multiview, multi-channel, multi-illumination time-lapse LSFM data on a single workstation or in parallel on a HPC. The pipeline relies on snakemake to resolve dependencies among consecutive processing steps and can be easily adapted to any cluster environment for processing LSFM data in a fraction of the time required to collect it.Comment: 13 pages with supplement, LATEX; 1 table, 1 figure, 2 supplementary figures, 2 supplementary lists, 2 supplementary tables; corrected error in results table, results unchange

    An Adaptive Threshold in Mammalian Neocortical Evolution

    Full text link
    Expansion of the neocortex is a hallmark of human evolution. However, it remains an open question what adaptive mechanisms facilitated its expansion. Here we show, using gyrencephaly index (GI) and other physiological and life-history data for 102 mammalian species, that gyrencephaly is an ancestral mammalian trait. We provide evidence that the evolution of a highly folded neocortex, as observed in humans, requires the traversal of a threshold of 10^9 neurons, and that species above and below the threshold exhibit a bimodal distribution of physiological and life-history traits, establishing two phenotypic groups. We identify, using discrete mathematical models, proliferative divisions of progenitors in the basal compartment of the developing neocortex as evolutionarily necessary and sufficient for generating a fourteen-fold increase in daily prenatal neuron production and thus traversal of the neuronal threshold. We demonstrate that length of neurogenic period, rather than any novel progenitor-type, is sufficient to distinguish cortical neuron number between species within the same phenotypic group.Comment: Currently under review; 38 pages, 5 Figures, 13 Supplementary Figures, 2 Table

    Mapping the complexity of transcription control in higher eukaryotes

    Get PDF
    Recent large analyses suggest the importance of combinatorial regulation by broadly expressed transcription factors rather than expression domains characterized by highly specific factors

    Globally optimal stitching of tiled 3D microscopic image acquisitions

    Get PDF
    Motivation: Modern anatomical and developmental studies often require high-resolution imaging of large specimens in three dimensions (3D). Confocal microscopy produces high-resolution 3D images, but is limited by a relatively small field of view compared with the size of large biological specimens. Therefore, motorized stages that move the sample are used to create a tiled scan of the whole specimen. The physical coordinates provided by the microscope stage are not precise enough to allow direct reconstruction (Stitching) of the whole image from individual image stacks

    CATMAID: collaborative annotation toolkit for massive amounts of image data

    Get PDF
    Summary: High-resolution, three-dimensional (3D) imaging of large biological specimens generates massive image datasets that are difficult to navigate, annotate and share effectively. Inspired by online mapping applications like GoogleMaps™, we developed a decentralized web interface that allows seamless navigation of arbitrarily large image stacks. Our interface provides means for online, collaborative annotation of the biological image data and seamless sharing of regions of interest by bookmarking. The CATMAID interface enables synchronized navigation through multiple registered datasets even at vastly different scales such as in comparisons between optical and electron microscopy

    ImgLib2-generic image processing in Java

    Get PDF
    Summary: ImgLib2 is an open-source Java library for n-dimensional data representation and manipulation with focus on image processing. It aims at minimizing code duplication by cleanly separating pixel-algebra, data access and data representation in memory. Algorithms can be implemented for classes of pixel types and generic access patterns by which they become independent of the specific dimensionality, pixel type and data representation. ImgLib2 illustrates that an elegant high-level programming interface can be achieved without sacrificing performance. It provides efficient implementations of common data types, storage layouts and algorithms. It is the data model underlying ImageJ2, the KNIME Image Processing toolbox and an increasing number of Fiji-Plugins

    Bioimage informatics in the context of drosophila research

    Get PDF
    Modern biological research relies heavily on microscopic imaging. The advanced genetic toolkit of drosophila makes it possible to label molecular and cellular components with unprecedented level of specificity necessitating the application of the most sophisticated imaging technologies. Imaging in drosophila spans all scales from single molecules to the entire populations of adult organisms, from electron microscopy to live imaging of developmental processes. As the imaging approaches become more complex and ambitious, there is an increasing need for quantitative, computer-mediated image processing and analysis to make sense of the imagery. Bioimage informatics is an emerging research field that covers all aspects of biological image analysis from data handling, through processing, to quantitative measurements, analysis and data presentation. Some of the most advanced, large scale projects, combining cutting edge imaging with complex bioimage informatics pipelines, are realized in the drosophila research community. In this review, we discuss the current research in biological image analysis specifically relevant to the type of systems level image datasets that are uniquely available for the drosophila model system. We focus on how state-of-the-art computer vision algorithms are impacting the ability of drosophila researchers to analyze biological systems in space and time. We pay particular attention to how these algorithmic advances from computer science are made usable to practicing biologists through open source platforms and how biologists can themselves participate in their further development
    corecore