11,460 research outputs found
Benchmarking SciDB Data Import on HPC Systems
SciDB is a scalable, computational database management system that uses an
array model for data storage. The array data model of SciDB makes it ideally
suited for storing and managing large amounts of imaging data. SciDB is
designed to support advanced analytics in database, thus reducing the need for
extracting data for analysis. It is designed to be massively parallel and can
run on commodity hardware in a high performance computing (HPC) environment. In
this paper, we present the performance of SciDB using simulated image data. The
Dynamic Distributed Dimensional Data Model (D4M) software is used to implement
the benchmark on a cluster running the MIT SuperCloud software stack. A peak
performance of 2.2M database inserts per second was achieved on a single node
of this system. We also show that SciDB and the D4M toolbox provide more
efficient ways to access random sub-volumes of massive datasets compared to the
traditional approaches of reading volumetric data from individual files. This
work describes the D4M and SciDB tools we developed and presents the initial
performance results. This performance was achieved by using parallel inserts, a
in-database merging of arrays as well as supercomputing techniques, such as
distributed arrays and single-program-multiple-data programming.Comment: 5 pages, 4 figures, IEEE High Performance Extreme Computing (HPEC)
2016, best paper finalis
Veni Vidi Vici, A Three-Phase Scenario For Parameter Space Analysis in Image Analysis and Visualization
Automatic analysis of the enormous sets of images is a critical task in life
sciences. This faces many challenges such as: algorithms are highly
parameterized, significant human input is intertwined, and lacking a standard
meta-visualization approach. This paper proposes an alternative iterative
approach for optimizing input parameters, saving time by minimizing the user
involvement, and allowing for understanding the workflow of algorithms and
discovering new ones. The main focus is on developing an interactive
visualization technique that enables users to analyze the relationships between
sampled input parameters and corresponding output. This technique is
implemented as a prototype called Veni Vidi Vici, or "I came, I saw, I
conquered." This strategy is inspired by the mathematical formulas of numbering
computable functions and is developed atop ImageJ, a scientific image
processing program. A case study is presented to investigate the proposed
framework. Finally, the paper explores some potential future issues in the
application of the proposed approach in parameter space analysis in
visualization
Recommended from our members
Collaborative analysis of multi-gigapixel imaging data using Cytomine
Motivation: Collaborative analysis of massive imaging datasets is essential to enable scientific discoveries.
Results: We developed Cytomine to foster active and distributed collaboration of multidisciplinary teams for large-scale image-based studies. It uses web development methodologies and machine learning in order to readily organize, explore, share and analyze (semantically and quantitatively) multi-gigapixel imaging data over the internet. We illustrate how it has been used in several biomedical applications
An end-to-end software solution for the analysis of high-throughput single-cell migration data
The systematic study of single-cell migration requires the availability of software for assisting data inspection, quality control and analysis. This is especially important for high-throughput experiments, where multiple biological conditions are tested in parallel. Although the field of cell migration can count on different computational tools for cell segmentation and tracking, downstream data visualization, parameter extraction and statistical analysis are still left to the user and are currently not possible within a single tool. This article presents a completely new module for the open-source, cross-platform CellMissy software for cell migration data management. This module is the first tool to focus specifically on single-cell migration data downstream of image processing. It allows fast comparison across all tested conditions, providing automated data visualization, assisted data filtering and quality control, extraction of various commonly used cell migration parameters, and non-parametric statistical analysis. Importantly, the module enables parameters computation both at the trajectory-and at the step-level. Moreover, this single-cell analysis module is complemented by a new data import module that accommodates multiwell plate data obtained from high-throughput experiments, and is easily extensible through a plugin architecture. In conclusion, the end-to-end software solution presented here tackles a key bioinformatics challenge in the cell migration field, assisting researchers in their highthroughput data processing
Exploring the relationship between the Engineering and Physical Sciences and the Health and Life Sciences by advanced bibliometric methods
We investigate the extent to which advances in the health and life sciences
(HLS) are dependent on research in the engineering and physical sciences (EPS),
particularly physics, chemistry, mathematics, and engineering. The analysis
combines two different bibliometric approaches. The first approach to analyze
the 'EPS-HLS interface' is based on term map visualizations of HLS research
fields. We consider 16 clinical fields and five life science fields. On the
basis of expert judgment, EPS research in these fields is studied by
identifying EPS-related terms in the term maps. In the second approach, a
large-scale citation-based network analysis is applied to publications from all
fields of science. We work with about 22,000 clusters of publications, each
representing a topic in the scientific literature. Citation relations are used
to identify topics at the EPS-HLS interface. The two approaches complement each
other. The advantages of working with textual data compensate for the
limitations of working with citation relations and the other way around. An
important advantage of working with textual data is in the in-depth qualitative
insights it provides. Working with citation relations, on the other hand,
yields many relevant quantitative statistics. We find that EPS research
contributes to HLS developments mainly in the following five ways: new
materials and their properties; chemical methods for analysis and molecular
synthesis; imaging of parts of the body as well as of biomaterial surfaces;
medical engineering mainly related to imaging, radiation therapy, signal
processing technology, and other medical instrumentation; mathematical and
statistical methods for data analysis. In our analysis, about 10% of all EPS
and HLS publications are classified as being at the EPS-HLS interface. This
percentage has remained more or less constant during the past decade
- …