324,775 research outputs found
Journal Staff
This book constitutes the refereed proceedings of the 18th Scandinavian Conference on Image Analysis, SCIA 2013, held in Espoo, Finland, in June 2013. The 67 revised full papers presented were carefully reviewed and selected from 132 submissions. The papers are organized in topical sections on feature extraction and segmentation, pattern recognition and machine learning, medical and biomedical image analysis, faces and gestures, object and scene recognition, matching, registration, and alignment, 3D vision, color and multispectral image analysis, motion analysis, systems and applications, human-centered computing, and video and multimedia analysis
Suggestive Annotation: A Deep Active Learning Framework for Biomedical Image Segmentation
Image segmentation is a fundamental problem in biomedical image analysis.
Recent advances in deep learning have achieved promising results on many
biomedical image segmentation benchmarks. However, due to large variations in
biomedical images (different modalities, image settings, objects, noise, etc),
to utilize deep learning on a new application, it usually needs a new set of
training data. This can incur a great deal of annotation effort and cost,
because only biomedical experts can annotate effectively, and often there are
too many instances in images (e.g., cells) to annotate. In this paper, we aim
to address the following question: With limited effort (e.g., time) for
annotation, what instances should be annotated in order to attain the best
performance? We present a deep active learning framework that combines fully
convolutional network (FCN) and active learning to significantly reduce
annotation effort by making judicious suggestions on the most effective
annotation areas. We utilize uncertainty and similarity information provided by
FCN and formulate a generalized version of the maximum set cover problem to
determine the most representative and uncertain areas for annotation. Extensive
experiments using the 2015 MICCAI Gland Challenge dataset and a lymph node
ultrasound image segmentation dataset show that, using annotation suggestions
by our method, state-of-the-art segmentation performance can be achieved by
using only 50% of training data.Comment: Accepted at MICCAI 201
Medical Image Segmentation Based on Multi-Modal Convolutional Neural Network: Study on Image Fusion Schemes
Image analysis using more than one modality (i.e. multi-modal) has been
increasingly applied in the field of biomedical imaging. One of the challenges
in performing the multimodal analysis is that there exist multiple schemes for
fusing the information from different modalities, where such schemes are
application-dependent and lack a unified framework to guide their designs. In
this work we firstly propose a conceptual architecture for the image fusion
schemes in supervised biomedical image analysis: fusing at the feature level,
fusing at the classifier level, and fusing at the decision-making level.
Further, motivated by the recent success in applying deep learning for natural
image analysis, we implement the three image fusion schemes above based on the
Convolutional Neural Network (CNN) with varied structures, and combined into a
single framework. The proposed image segmentation framework is capable of
analyzing the multi-modality images using different fusing schemes
simultaneously. The framework is applied to detect the presence of soft tissue
sarcoma from the combination of Magnetic Resonance Imaging (MRI), Computed
Tomography (CT) and Positron Emission Tomography (PET) images. It is found from
the results that while all the fusion schemes outperform the single-modality
schemes, fusing at the feature level can generally achieve the best performance
in terms of both accuracy and computational cost, but also suffers from the
decreased robustness in the presence of large errors in any image modalities.Comment: Zhe Guo and Xiang Li contribute equally to this wor
Visual parameter optimisation for biomedical image processing
Background: Biomedical image processing methods require users to optimise input parameters to ensure high quality
output. This presents two challenges. First, it is difficult to optimise multiple input parameters for multiple
input images. Second, it is difficult to achieve an understanding of underlying algorithms, in particular, relationships
between input and output.
Results: We present a visualisation method that transforms users’ ability to understand algorithm behaviour by
integrating input and output, and by supporting exploration of their relationships. We discuss its application to a
colour deconvolution technique for stained histology images and show how it enabled a domain expert to
identify suitable parameter values for the deconvolution of two types of images, and metrics to quantify
deconvolution performance. It also enabled a breakthrough in understanding by invalidating an underlying
assumption about the algorithm.
Conclusions: The visualisation method presented here provides analysis capability for multiple inputs and outputs
in biomedical image processing that is not supported by previous analysis software. The analysis supported by our
method is not feasible with conventional trial-and-error approaches
- …