57,039 research outputs found
Scene Parsing with Multiscale Feature Learning, Purity Trees, and Optimal Covers
Scene parsing, or semantic segmentation, consists in labeling each pixel in
an image with the category of the object it belongs to. It is a challenging
task that involves the simultaneous detection, segmentation and recognition of
all the objects in the image.
The scene parsing method proposed here starts by computing a tree of segments
from a graph of pixel dissimilarities. Simultaneously, a set of dense feature
vectors is computed which encodes regions of multiple sizes centered on each
pixel. The feature extractor is a multiscale convolutional network trained from
raw pixels. The feature vectors associated with the segments covered by each
node in the tree are aggregated and fed to a classifier which produces an
estimate of the distribution of object categories contained in the segment. A
subset of tree nodes that cover the image are then selected so as to maximize
the average "purity" of the class distributions, hence maximizing the overall
likelihood that each segment will contain a single object. The convolutional
network feature extractor is trained end-to-end from raw pixels, alleviating
the need for engineered features. After training, the system is parameter free.
The system yields record accuracies on the Stanford Background Dataset (8
classes), the Sift Flow Dataset (33 classes) and the Barcelona Dataset (170
classes) while being an order of magnitude faster than competing approaches,
producing a 320 \times 240 image labeling in less than 1 second.Comment: 9 pages, 4 figures - Published in 29th International Conference on
Machine Learning (ICML 2012), Jun 2012, Edinburgh, United Kingdo
Does segmentation always improve model performance in credit scoring?
Credit scoring allows for the credit risk assessment of bank customers. A single scoring model (scorecard) can be developed for the entire customer population, e.g. using logistic regression. However, it is often expected that segmentation, i.e. dividing the population into several groups and building separate scorecards for them, will improve the model performance. The most common statistical methods for segmentation are the two-step approaches, where logistic regression follows Classification and Regression Trees (CART) or Chi-squared Automatic Interaction Detection (CHAID) trees etc. In this research, the two-step approaches are applied as well as a new, simultaneous method, in which both segmentation and scorecards are optimised at the same time: Logistic Trees with Unbiased Selection (LOTUS). For reference purposes, a single-scorecard model is used. The above-mentioned methods are applied to the data provided by two of the major UK banks and one of the European credit bureaus. The model performance measures are then compared to examine whether there is improvement due to the segmentation methods used. It is found that segmentation does not always improve model performance in credit scoring: for none of the analysed real-world datasets, the multi-scorecard models perform considerably better than the single-scorecard ones. Moreover, in this application, there is no difference in performance between the two-step and simultaneous approache
Extracting Tree-structures in CT data by Tracking Multiple Statistically Ranked Hypotheses
In this work, we adapt a method based on multiple hypothesis tracking (MHT)
that has been shown to give state-of-the-art vessel segmentation results in
interactive settings, for the purpose of extracting trees. Regularly spaced
tubular templates are fit to image data forming local hypotheses. These local
hypotheses are used to construct the MHT tree, which is then traversed to make
segmentation decisions. However, some critical parameters in this method are
scale-dependent and have an adverse effect when tracking structures of varying
dimensions. We propose to use statistical ranking of local hypotheses in
constructing the MHT tree, which yields a probabilistic interpretation of
scores across scales and helps alleviate the scale-dependence of MHT
parameters. This enables our method to track trees starting from a single seed
point. Our method is evaluated on chest CT data to extract airway trees and
coronary arteries. In both cases, we show that our method performs
significantly better than the original MHT method.Comment: Accepted for publication at the International Journal of Medical
Physics and Practic
Graph Refinement based Airway Extraction using Mean-Field Networks and Graph Neural Networks
Graph refinement, or the task of obtaining subgraphs of interest from
over-complete graphs, can have many varied applications. In this work, we
extract trees or collection of sub-trees from image data by, first deriving a
graph-based representation of the volumetric data and then, posing the tree
extraction as a graph refinement task. We present two methods to perform graph
refinement. First, we use mean-field approximation (MFA) to approximate the
posterior density over the subgraphs from which the optimal subgraph of
interest can be estimated. Mean field networks (MFNs) are used for inference
based on the interpretation that iterations of MFA can be seen as feed-forward
operations in a neural network. This allows us to learn the model parameters
using gradient descent. Second, we present a supervised learning approach using
graph neural networks (GNNs) which can be seen as generalisations of MFNs.
Subgraphs are obtained by training a GNN-based graph refinement model to
directly predict edge probabilities. We discuss connections between the two
classes of methods and compare them for the task of extracting airways from 3D,
low-dose, chest CT data. We show that both the MFN and GNN models show
significant improvement when compared to one baseline method, that is similar
to a top performing method in the EXACT'09 Challenge, and a 3D U-Net based
airway segmentation model, in detecting more branches with fewer false
positives.Comment: Accepted for publication at Medical Image Analysis. 14 page
A graph-based mathematical morphology reader
This survey paper aims at providing a "literary" anthology of mathematical
morphology on graphs. It describes in the English language many ideas stemming
from a large number of different papers, hence providing a unified view of an
active and diverse field of research
Integrating Prosodic and Lexical Cues for Automatic Topic Segmentation
We present a probabilistic model that uses both prosodic and lexical cues for
the automatic segmentation of speech into topically coherent units. We propose
two methods for combining lexical and prosodic information using hidden Markov
models and decision trees. Lexical information is obtained from a speech
recognizer, and prosodic features are extracted automatically from speech
waveforms. We evaluate our approach on the Broadcast News corpus, using the
DARPA-TDT evaluation metrics. Results show that the prosodic model alone is
competitive with word-based segmentation methods. Furthermore, we achieve a
significant reduction in error by combining the prosodic and word-based
knowledge sources.Comment: 27 pages, 8 figure
- …