16 research outputs found
Catastrophic forgetting: still a problem for DNNs
We investigate the performance of DNNs when trained on class-incremental
visual problems consisting of initial training, followed by retraining with
added visual classes. Catastrophic forgetting (CF) behavior is measured using a
new evaluation procedure that aims at an application-oriented view of
incremental learning. In particular, it imposes that model selection must be
performed on the initial dataset alone, as well as demanding that retraining
control be performed only using the retraining dataset, as initial dataset is
usually too large to be kept. Experiments are conducted on class-incremental
problems derived from MNIST, using a variety of different DNN models, some of
them recently proposed to avoid catastrophic forgetting. When comparing our new
evaluation procedure to previous approaches for assessing CF, we find their
findings are completely negated, and that none of the tested methods can avoid
CF in all experiments. This stresses the importance of a realistic empirical
measurement procedure for catastrophic forgetting, and the need for further
research in incremental learning for DNNs.Comment: 10 pages, 11 figures, Artificial Neural Networks and Machine Learning
- ICANN 201
Biased Competition in Visual Processing Hierarchies: A Learning Approach Using Multiple Cues
In this contribution, we present a large-scale hierarchical system for object detection fusing bottom-up (signal-driven) processing results with top-down (model or task-driven) attentional modulation. Specifically, we focus on the question of how the autonomous learning of invariant models can be embedded into a performing system and how such models can be used to define object-specific attentional modulation signals. Our system implements bi-directional data flow in a processing hierarchy. The bottom-up data flow proceeds from a preprocessing level to the hypothesis level where object hypotheses created by exhaustive object detection algorithms are represented in a roughly retinotopic way. A competitive selection mechanism is used to determine the most confident hypotheses, which are used on the system level to train multimodal models that link object identity to invariant hypothesis properties. The top-down data flow originates at the system level, where the trained multimodal models are used to obtain space- and feature-based attentional modulation signals, providing biases for the competitive selection process at the hypothesis level. This results in object-specific hypothesis facilitation/suppression in certain image regions which we show to be applicable to different object detection mechanisms. In order to demonstrate the benefits of this approach, we apply the system to the detection of cars in a variety of challenging traffic videos. Evaluating our approach on a publicly available dataset containing approximately 3,500 annotated video images from more than 1 h of driving, we can show strong increases in performance and generalization when compared to object detection in isolation. Furthermore, we compare our results to a late hypothesis rejection approach, showing that early coupling of top-down and bottom-up information is a favorable approach especially when processing resources are constrained