24,245 research outputs found
Veni Vidi Vici, A Three-Phase Scenario For Parameter Space Analysis in Image Analysis and Visualization
Automatic analysis of the enormous sets of images is a critical task in life
sciences. This faces many challenges such as: algorithms are highly
parameterized, significant human input is intertwined, and lacking a standard
meta-visualization approach. This paper proposes an alternative iterative
approach for optimizing input parameters, saving time by minimizing the user
involvement, and allowing for understanding the workflow of algorithms and
discovering new ones. The main focus is on developing an interactive
visualization technique that enables users to analyze the relationships between
sampled input parameters and corresponding output. This technique is
implemented as a prototype called Veni Vidi Vici, or "I came, I saw, I
conquered." This strategy is inspired by the mathematical formulas of numbering
computable functions and is developed atop ImageJ, a scientific image
processing program. A case study is presented to investigate the proposed
framework. Finally, the paper explores some potential future issues in the
application of the proposed approach in parameter space analysis in
visualization
Distance Measures for Reduced Ordering Based Vector Filters
Reduced ordering based vector filters have proved successful in removing
long-tailed noise from color images while preserving edges and fine image
details. These filters commonly utilize variants of the Minkowski distance to
order the color vectors with the aim of distinguishing between noisy and
noise-free vectors. In this paper, we review various alternative distance
measures and evaluate their performance on a large and diverse set of images
using several effectiveness and efficiency criteria. The results demonstrate
that there are in fact strong alternatives to the popular Minkowski metrics
Local Variation as a Statistical Hypothesis Test
The goal of image oversegmentation is to divide an image into several pieces,
each of which should ideally be part of an object. One of the simplest and yet
most effective oversegmentation algorithms is known as local variation (LV)
(Felzenszwalb and Huttenlocher 2004). In this work, we study this algorithm and
show that algorithms similar to LV can be devised by applying different
statistical models and decisions, thus providing further theoretical
justification and a well-founded explanation for the unexpected high
performance of the LV approach. Some of these algorithms are based on
statistics of natural images and on a hypothesis testing decision; we denote
these algorithms probabilistic local variation (pLV). The best pLV algorithm,
which relies on censored estimation, presents state-of-the-art results while
keeping the same computational complexity of the LV algorithm
Reading Scene Text in Deep Convolutional Sequences
We develop a Deep-Text Recurrent Network (DTRN) that regards scene text
reading as a sequence labelling problem. We leverage recent advances of deep
convolutional neural networks to generate an ordered high-level sequence from a
whole word image, avoiding the difficult character segmentation problem. Then a
deep recurrent model, building on long short-term memory (LSTM), is developed
to robustly recognize the generated CNN sequences, departing from most existing
approaches recognising each character independently. Our model has a number of
appealing properties in comparison to existing scene text recognition methods:
(i) It can recognise highly ambiguous words by leveraging meaningful context
information, allowing it to work reliably without either pre- or
post-processing; (ii) the deep CNN feature is robust to various image
distortions; (iii) it retains the explicit order information in word image,
which is essential to discriminate word strings; (iv) the model does not depend
on pre-defined dictionary, and it can process unknown words and arbitrary
strings. Codes for the DTRN will be available.Comment: To appear in the 13th AAAI Conference on Artificial Intelligence
(AAAI-16), 201
Shaped extensions of singular spectrum analysis
Extensions of singular spectrum analysis (SSA) for processing of
non-rectangular images and time series with gaps are considered. A circular
version is suggested, which allows application of the method to the data given
on a circle or on a cylinder, e.g. cylindrical projection of a 3D ellipsoid.
The constructed Shaped SSA method with planar or circular topology is able to
produce low-rank approximations for images of complex shapes. Together with
Shaped SSA, a shaped version of the subspace-based ESPRIT method for frequency
estimation is developed. Examples of 2D circular SSA and 2D Shaped ESPRIT are
presented
Computing Multidimensional Persistence
The theory of multidimensional persistence captures the topology of a
multifiltration -- a multiparameter family of increasing spaces.
Multifiltrations arise naturally in the topological analysis of scientific
data. In this paper, we give a polynomial time algorithm for computing
multidimensional persistence. We recast this computation as a problem within
computational algebraic geometry and utilize algorithms from this area to solve
it. While the resulting problem is Expspace-complete and the standard
algorithms take doubly-exponential time, we exploit the structure inherent
withing multifiltrations to yield practical algorithms. We implement all
algorithms in the paper and provide statistical experiments to demonstrate
their feasibility.Comment: This paper has been withdrawn by the authors. Journal of
Computational Geometry, 1(1) 2010, pages 72-100.
http://jocg.org/index.php/jocg/article/view/1
- …