60,491 research outputs found
Learning scale-variant and scale-invariant features for deep image classification
Convolutional Neural Networks (CNNs) require large image corpora to be
trained on classification tasks. The variation in image resolutions, sizes of
objects and patterns depicted, and image scales, hampers CNN training and
performance, because the task-relevant information varies over spatial scales.
Previous work attempting to deal with such scale variations focused on
encouraging scale-invariant CNN representations. However, scale-invariant
representations are incomplete representations of images, because images
contain scale-variant information as well. This paper addresses the combined
development of scale-invariant and scale-variant representations. We propose a
multi- scale CNN method to encourage the recognition of both types of features
and evaluate it on a challenging image classification task involving
task-relevant characteristics at multiple scales. The results show that our
multi-scale CNN outperforms single-scale CNN. This leads to the conclusion that
encouraging the combined development of a scale-invariant and scale-variant
representation in CNNs is beneficial to image recognition performance
Sparse Bayesian mass-mapping with uncertainties: hypothesis testing of structure
A crucial aspect of mass-mapping, via weak lensing, is quantification of the
uncertainty introduced during the reconstruction process. Properly accounting
for these errors has been largely ignored to date. We present results from a
new method that reconstructs maximum a posteriori (MAP) convergence maps by
formulating an unconstrained Bayesian inference problem with Laplace-type
-norm sparsity-promoting priors, which we solve via convex
optimization. Approaching mass-mapping in this manner allows us to exploit
recent developments in probability concentration theory to infer theoretically
conservative uncertainties for our MAP reconstructions, without relying on
assumptions of Gaussianity. For the first time these methods allow us to
perform hypothesis testing of structure, from which it is possible to
distinguish between physical objects and artifacts of the reconstruction. Here
we present this new formalism, demonstrate the method on illustrative examples,
before applying the developed formalism to two observational datasets of the
Abel-520 cluster. In our Bayesian framework it is found that neither Abel-520
dataset can conclusively determine the physicality of individual local massive
substructure at significant confidence. However, in both cases the recovered
MAP estimators are consistent with both sets of data
A Neural Model for Self Organizing Feature Detectors and Classifiers in a Network Hierarchy
Many models of early cortical processing have shown how local learning rules can produce efficient, sparse-distributed codes in which nodes have responses that are statistically independent and low probability. However, it is not known how to develop a useful hierarchical representation, containing sparse-distributed codes at each level of the hierarchy, that incorporates predictive feedback from the environment. We take a step in that direction by proposing a biologically plausible neural network model that develops receptive fields, and learns to make class predictions, with or without the help of environmental feedback. The model is a new type of predictive adaptive resonance theory network called Receptive Field ARTMAP, or RAM. RAM self organizes internal category nodes that are tuned to activity distributions in topographic input maps. Each receptive field is composed of multiple weight fields that are adapted via local, on-line learning, to form smooth receptive ftelds that reflect; the statistics of the activity distributions in the input maps. When RAM generates incorrect predictions, its vigilance is raised, amplifying subtractive inhibition and sharpening receptive fields until the error is corrected. Evaluation on several classification benchmarks shows that RAM outperforms a related (but neurally implausible) model called Gaussian ARTMAP, as well as several standard neural network and statistical classifters. A topographic version of RAM is proposed, which is capable of self organizing hierarchical representations. Topographic RAM is a model for receptive field development at any level of the cortical hierarchy, and provides explanations for a variety of perceptual learning data.Defense Advanced Research Projects Agency and Office of Naval Research (N00014-95-1-0409
Limitations for shapelet-based weak-lensing measurements
We seek to understand the impact on shape estimators obtained from circular
and elliptical shapelet models under two realistic conditions: (a) only a
limited number of shapelet modes is available for the model, and (b) the
intrinsic galactic shapes are not restricted to shapelet models.
We create a set of simplistic simulations, in which the galactic shapes
follow a Sersic profile. By varying the Sersic index and applied shear, we
quantify the amount of bias on shear estimates which arises from insufficient
modeling. Additional complications due to PSF convolution, pixelation and pixel
noise are also discussed.
Steep and highly elliptical galaxy shapes cannot be accurately modeled within
the circular shapelet basis system and are biased towards shallower and less
elongated shapes. This problem can be cured partially by allowing elliptical
basis functions, but for steep profiles elliptical shapelet models still depend
critically on accurate ellipticity priors. As a result, shear estimates are
typically biased low. Independently of the particular form of the estimator,
the bias depends on the true intrinsic galaxy morphology, but also on the size
and shape of the PSF.
As long as the issues discussed here are not solved, the shapelet method
cannot provide weak-lensing measurements with an accuracy demanded by upcoming
missions and surveys, unless one can provide an accurate and reliable
calibration, specific for the dataset under investigation.Comment: 8 pages, 5 figures, submitted to A&
Information-Theoretic Active Learning for Content-Based Image Retrieval
We propose Information-Theoretic Active Learning (ITAL), a novel batch-mode
active learning method for binary classification, and apply it for acquiring
meaningful user feedback in the context of content-based image retrieval.
Instead of combining different heuristics such as uncertainty, diversity, or
density, our method is based on maximizing the mutual information between the
predicted relevance of the images and the expected user feedback regarding the
selected batch. We propose suitable approximations to this computationally
demanding problem and also integrate an explicit model of user behavior that
accounts for possible incorrect labels and unnameable instances. Furthermore,
our approach does not only take the structure of the data but also the expected
model output change caused by the user feedback into account. In contrast to
other methods, ITAL turns out to be highly flexible and provides
state-of-the-art performance across various datasets, such as MIRFLICKR and
ImageNet.Comment: GCPR 2018 paper (14 pages text + 2 pages references + 6 pages
appendix
- …