12,940 research outputs found
Integrating a Non-Uniformly Sampled Software Retina with a Deep CNN Model
We present a biologically inspired method for pre-processing images applied to CNNs
that reduces their memory requirements while increasing their invariance to scale and rotation
changes. Our method is based on the mammalian retino-cortical transform: a
mapping between a pseudo-randomly tessellated retina model (used to sample an input
image) and a CNN. The aim of this first pilot study is to demonstrate a functional retinaintegrated
CNN implementation and this produced the following results: a network using
the full retino-cortical transform yielded an F1 score of 0.80 on a test set during a 4-way
classification task, while an identical network not using the proposed method yielded an
F1 score of 0.86 on the same task. The method reduced the visual data by e×7, the input
data to the CNN by 40% and the number of CNN training epochs by 64%. These results
demonstrate the viability of our method and hint at the potential of exploiting functional
traits of natural vision systems in CNNs
Manitest: Are classifiers really invariant?
Invariance to geometric transformations is a highly desirable property of
automatic classifiers in many image recognition tasks. Nevertheless, it is
unclear to which extent state-of-the-art classifiers are invariant to basic
transformations such as rotations and translations. This is mainly due to the
lack of general methods that properly measure such an invariance. In this
paper, we propose a rigorous and systematic approach for quantifying the
invariance to geometric transformations of any classifier. Our key idea is to
cast the problem of assessing a classifier's invariance as the computation of
geodesics along the manifold of transformed images. We propose the Manitest
method, built on the efficient Fast Marching algorithm to compute the
invariance of classifiers. Our new method quantifies in particular the
importance of data augmentation for learning invariance from data, and the
increased invariance of convolutional neural networks with depth. We foresee
that the proposed generic tool for measuring invariance to a large class of
geometric transformations and arbitrary classifiers will have many applications
for evaluating and comparing classifiers based on their invariance, and help
improving the invariance of existing classifiers.Comment: BMVC 201
- …