Learning from Students: Online Contrastive Distillation Network for General Continual Learning
Learning from Students: Online Contrastive Distillation Network for General Continual Learning
Jin Li, Zhong Ji, Gang Wang, Qiang Wang, Feng Gao
Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence
Main Track. Pages 3215-3221.
https://doi.org/10.24963/ijcai.2022/446
The goal of General Continual Learning (GCL) is to preserve learned knowledge and learn new knowledge with constant memory from an infinite data stream where task boundaries are blurry. Distilling the model's response of reserved samples between the old and the new models is an effective way to achieve promise performance on GCL. However, it accumulates the inherent old model's response bias and is not robust to model changes. To this end, we propose an Online Contrastive Distillation Network (OCD-Net) to tackle these problems, which explores the merit of the student model in each time step to guide the training process of the student model. Concretely, the teacher model is devised to help the student model to consolidate the learned knowledge, which is trained online via integrating the model weights of the student model to accumulate the new knowledge. Moreover, our OCD-Net incorporates both relation and adaptive response to help the student model alleviate the catastrophic forgetting, which is also beneficial for the teacher model preserves the learned knowledge. Extensive experiments on six benchmark datasets demonstrate that our proposed OCD-Net significantly outperforms state-of-the-art approaches in 3.26%~8.71% with various buffer sizes. Our code is available at https://github.com/lijincm/OCD-Net.
Keywords:
Machine Learning: Incremental Learning
Computer Vision: Recognition (object detection, categorization)