268 research outputs found
Compressing the Input for CNNs with the First-Order Scattering Transform
International audienceWe study the first-order scattering transform as a candidate for reducing the signal processed by a convolutional neural network (CNN). We study this transformation and show theoretical and empirical evidence that in the case of natural images and sufficiently small translation invariance, this transform preserves most of the signal information needed for classification while substantially reducing the spatial resolution and total signal size. We show that cascading a CNN with this representation performs on par with ImageNet classification models commonly used in downstream tasks such as the ResNet-50. We subsequently apply our trained hybrid ImageNet model as a base model on a detection system, which has typically larger image inputs. On Pascal VOC and COCO detection tasks we deliver substantial improvements in the inference speed and training memory consumption compared to models trained directly on the input image
Boosting Deep Neural Networks with Geometrical Prior Knowledge: A Survey
While Deep Neural Networks (DNNs) achieve state-of-the-art results in many
different problem settings, they are affected by some crucial weaknesses. On
the one hand, DNNs depend on exploiting a vast amount of training data, whose
labeling process is time-consuming and expensive. On the other hand, DNNs are
often treated as black box systems, which complicates their evaluation and
validation. Both problems can be mitigated by incorporating prior knowledge
into the DNN.
One promising field, inspired by the success of convolutional neural networks
(CNNs) in computer vision tasks, is to incorporate knowledge about symmetric
geometrical transformations of the problem to solve. This promises an increased
data-efficiency and filter responses that are interpretable more easily. In
this survey, we try to give a concise overview about different approaches to
incorporate geometrical prior knowledge into DNNs. Additionally, we try to
connect those methods to the field of 3D object detection for autonomous
driving, where we expect promising results applying those methods.Comment: Survey Pape
Harmonic Convolutional Networks based on Discrete Cosine Transform
Convolutional neural networks (CNNs) learn filters in order to capture local
correlation patterns in feature space. We propose to learn these filters as
combinations of preset spectral filters defined by the Discrete Cosine
Transform (DCT). Our proposed DCT-based harmonic blocks replace conventional
convolutional layers to produce partially or fully harmonic versions of new or
existing CNN architectures. Using DCT energy compaction properties, we
demonstrate how the harmonic networks can be efficiently compressed by
truncating high-frequency information in harmonic blocks thanks to the
redundancies in the spectral domain. We report extensive experimental
validation demonstrating benefits of the introduction of harmonic blocks into
state-of-the-art CNN models in image classification, object detection and
semantic segmentation applications.Comment: arXiv admin note: substantial text overlap with arXiv:1812.0320
Harmonic Networks for Image Classification
Convolutional neural networks (CNNs) learn filters in order to capture local correlation patterns in feature space. In contrast, in this paper we propose harmonic blocks that
produce features by learning optimal combinations of responses to preset spectral filters.
We rely on the use of the Discrete Cosine Transform filters which have excellent energy
compaction properties and are widely used for image compression. The proposed harmonic blocks are intended to replace conventional convolutional layers to produce partially or fully harmonic versions of new or existing CNN architectures. We demonstrate
how the harmonic networks can be efficiently compressed by exploiting redundancy in
spectral domain and truncating high-frequency information. We extensively validate our
approach and show that the introduction of harmonic blocks into state-of-the-art CNN
models results in improved classification performance on CIFAR and ImageNet datasets
- …