8,103 research outputs found
Active Mean Fields for Probabilistic Image Segmentation: Connections with Chan-Vese and Rudin-Osher-Fatemi Models
Segmentation is a fundamental task for extracting semantically meaningful
regions from an image. The goal of segmentation algorithms is to accurately
assign object labels to each image location. However, image-noise, shortcomings
of algorithms, and image ambiguities cause uncertainty in label assignment.
Estimating the uncertainty in label assignment is important in multiple
application domains, such as segmenting tumors from medical images for
radiation treatment planning. One way to estimate these uncertainties is
through the computation of posteriors of Bayesian models, which is
computationally prohibitive for many practical applications. On the other hand,
most computationally efficient methods fail to estimate label uncertainty. We
therefore propose in this paper the Active Mean Fields (AMF) approach, a
technique based on Bayesian modeling that uses a mean-field approximation to
efficiently compute a segmentation and its corresponding uncertainty. Based on
a variational formulation, the resulting convex model combines any
label-likelihood measure with a prior on the length of the segmentation
boundary. A specific implementation of that model is the Chan-Vese segmentation
model (CV), in which the binary segmentation task is defined by a Gaussian
likelihood and a prior regularizing the length of the segmentation boundary.
Furthermore, the Euler-Lagrange equations derived from the AMF model are
equivalent to those of the popular Rudin-Osher-Fatemi (ROF) model for image
denoising. Solutions to the AMF model can thus be implemented by directly
utilizing highly-efficient ROF solvers on log-likelihood ratio fields. We
qualitatively assess the approach on synthetic data as well as on real natural
and medical images. For a quantitative evaluation, we apply our approach to the
icgbench dataset
Spatial-temporal data mining procedure: LASR
This paper is concerned with the statistical development of our
spatial-temporal data mining procedure, LASR (pronounced ``laser''). LASR is
the abbreviation for Longitudinal Analysis with Self-Registration of
large--small- data. It was motivated by a study of ``Neuromuscular
Electrical Stimulation'' experiments, where the data are noisy and
heterogeneous, might not align from one session to another, and involve a large
number of multiple comparisons. The three main components of LASR are: (1) data
segmentation for separating heterogeneous data and for distinguishing outliers,
(2) automatic approaches for spatial and temporal data registration, and (3)
statistical smoothing mapping for identifying ``activated'' regions based on
false-discovery-rate controlled -maps and movies. Each of the components is
of interest in its own right. As a statistical ensemble, the idea of LASR is
applicable to other types of spatial-temporal data sets beyond those from the
NMES experiments.Comment: Published at http://dx.doi.org/10.1214/074921706000000707 in the IMS
Lecture Notes--Monograph Series
(http://www.imstat.org/publications/lecnotes.htm) by the Institute of
Mathematical Statistics (http://www.imstat.org
3D Shape Segmentation with Projective Convolutional Networks
This paper introduces a deep architecture for segmenting 3D objects into
their labeled semantic parts. Our architecture combines image-based Fully
Convolutional Networks (FCNs) and surface-based Conditional Random Fields
(CRFs) to yield coherent segmentations of 3D shapes. The image-based FCNs are
used for efficient view-based reasoning about 3D object parts. Through a
special projection layer, FCN outputs are effectively aggregated across
multiple views and scales, then are projected onto the 3D object surfaces.
Finally, a surface-based CRF combines the projected outputs with geometric
consistency cues to yield coherent segmentations. The whole architecture
(multi-view FCNs and CRF) is trained end-to-end. Our approach significantly
outperforms the existing state-of-the-art methods in the currently largest
segmentation benchmark (ShapeNet). Finally, we demonstrate promising
segmentation results on noisy 3D shapes acquired from consumer-grade depth
cameras.Comment: This is an updated version of our CVPR 2017 paper. We incorporated
new experiments that demonstrate ShapePFCN performance under the case of
consistent *upright* orientation and an additional input channel in our
rendered images for encoding height from the ground plane (upright axis
coordinate values). Performance is improved in this settin
Eigenspectra optoacoustic tomography achieves quantitative blood oxygenation imaging deep in tissues
Light propagating in tissue attains a spectrum that varies with location due
to wavelength-dependent fluence attenuation by tissue optical properties, an
effect that causes spectral corruption. Predictions of the spectral variations
of light fluence in tissue are challenging since the spatial distribution of
optical properties in tissue cannot be resolved in high resolution or with high
accuracy by current methods. Spectral corruption has fundamentally limited the
quantification accuracy of optical and optoacoustic methods and impeded the
long sought-after goal of imaging blood oxygen saturation (sO2) deep in
tissues; a critical but still unattainable target for the assessment of
oxygenation in physiological processes and disease. We discover a new principle
underlying light fluence in tissues, which describes the wavelength dependence
of light fluence as an affine function of a few reference base spectra,
independently of the specific distribution of tissue optical properties. This
finding enables the introduction of a previously undocumented concept termed
eigenspectra Multispectral Optoacoustic Tomography (eMSOT) that can effectively
account for wavelength dependent light attenuation without explicit knowledge
of the tissue optical properties. We validate eMSOT in more than 2000
simulations and with phantom and animal measurements. We find that eMSOT can
quantitatively image tissue sO2 reaching in many occasions a better than
10-fold improved accuracy over conventional spectral optoacoustic methods.
Then, we show that eMSOT can spatially resolve sO2 in muscle and tumor;
revealing so far unattainable tissue physiology patterns. Last, we related
eMSOT readings to cancer hypoxia and found congruence between eMSOT tumor sO2
images and tissue perfusion and hypoxia maps obtained by correlative
histological analysis
Interactive semantic mapping: Experimental evaluation
Robots that are launched in the consumer market need to provide more effective human robot interaction, and, in particular, spoken language interfaces. However, in order to support the execution of high level commands as they are specified in natural language, a semantic map is required. Such a map is a representation that enables the robot to ground the commands into the actual places and objects located in the environment. In this paper, we present the experimental evaluation of a system specifically designed to build semantically rich maps, through the interaction with the user. The results of the experiments not only provide the basis for a discussion of the features of the proposed approach, but also highlight the manifold issues that arise in the evaluation of semantic mapping
Towards the development of a smart flying sensor: illustration in the field of precision agriculture
Sensing is an important element to quantify productivity, product quality and to make decisions. Applications, such as mapping, surveillance, exploration and precision agriculture, require a reliable platform for remote sensing. This paper presents the first steps towards the development of a smart flying sensor based on an unmanned aerial vehicle (UAV). The concept of smart remote sensing is illustrated and its performance tested for the task of mapping the volume of grain inside a trailer during forage harvesting. Novelty lies in: (1) the development of a position-estimation method with time delay compensation based on inertial measurement unit (IMU) sensors and image processing; (2) a method to build a 3D map using information obtained from a regular camera; and (3) the design and implementation of a path-following control algorithm using model predictive control (MPC). Experimental results on a lab-scale system validate the effectiveness of the proposed methodology
- …