Skip to Main content Skip to Navigation
New interface
Conference papers

Attenuating Catastrophic Forgetting by Joint Contrastive and Incremental Learning

Abstract : In class incremental learning, discriminative models are trained to classify images while adapting to new instances and classes incrementally. Training a model to adapt to new classes without total access to previous class data, however, leads to the known problem of catastrophic forgetting of the previously learnt classes. To alleviate this problem, we show how we can build upon recent progress on contrastive learning methods. In particular, we develop an incremental learning approach for deep neural networks operating both at classification and representation level which alleviates forgetting and learns more general features for data classification. Experiments performed on several datasets demonstrate the superiority of the proposed method with respect to well known state-of-the-art methods.
Document type :
Conference papers
Complete list of metadata

https://hal-ensta-bretagne.archives-ouvertes.fr/hal-03784379
Contributor : Benoit Clement Connect in order to contact the contributor
Submitted on : Friday, September 23, 2022 - 8:45:44 AM
Last modification on : Thursday, September 29, 2022 - 4:56:02 AM

File

Ferdinand_Attenuating_Catastro...
Files produced by the author(s)

Identifiers

Citation

Quentin Ferdinand, Benoit Clement, Quentin Oliveau, Gilles Le Chenadec, Panagiotis Papadakis. Attenuating Catastrophic Forgetting by Joint Contrastive and Incremental Learning. IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Jun 2022, New-Orleans, United States. ⟨10.1109/CVPRW56347.2022.00423⟩. ⟨hal-03784379⟩

Share

Metrics

Record views

20

Files downloads

11