Reviewing plays an important role when learning knowledge. The knowledge
acquisition at a certain time point may be strongly inspired with the help of
previous experience. Thus the knowledge growing procedure should show strong
relationship along the temporal dimension. In our research, we find that during
the network training, the evolution of feature map follows temporal sequence
property. A proper temporal supervision may further improve the network
training performance. Inspired by this observation, we propose Temporal
Supervised Knowledge Distillation (TSKD). Specifically, we extract the
spatiotemporal features in the different training phases of student by
convolutional Long Short-term memory network (Conv-LSTM). Then, we train the
student net through a dynamic target, rather than static teacher network
features. This process realizes the refinement of old knowledge in student
network, and utilizes it to assist current learning. Extensive experiments
verify the effectiveness and advantages of our method over existing knowledge
distillation methods, including various network architectures and different
tasks (image classification and object detection) .Comment: Under review in AAAI 202