Despite their effectiveness in a wide range of tasks, deep architectures suffer from some important limitations. In particular, they are vulnerable to catastrophic forgetting, i.e. they perform poorly when they are required to update their model as new classes are available but the original training set is not retained. This paper addresses this problem in the context of semantic segmentation. Current strategies fail on this task because they do not consider a peculiar aspect of semantic segmentation: since each training step provides annotation only for a subset of all possible classes, pixels of the background class (i.e. pixels that do not belong to any other classes) exhibit a semantic distribution shift. In this work we revisit classical incremental learning methods, proposing a new distillation-based framework which explicitly accounts for this shift. Furthermore, we introduce a novel strategy to initialize classifier's parameters, thus preventing biased predictions toward the ba...

Despite their effectiveness in a wide range of tasks, deep architectures suffer from some important limitations. In particular, they are vulnerable to catastrophic forgetting, i.e. they perform poorly when they are required to update their model as new classes are available but the original training set is not retained. This paper addresses this problem in the context of semantic segmentation. Current strategies fail on this task because they do not consider a peculiar aspect of semantic segmentation: since each training step provides annotation only for a subset of all possible classes, pixels of the background class (i.e. pixels that do not belong to any other classes) exhibit a semantic distribution shift. In this work we revisit classical incremental learning methods, proposing a new distillation-based framework which explicitly accounts for this shift. Furthermore, we introduce a novel strategy to initialize classifier's parameters, thus preventing biased predictions toward the background class. We demonstrate the effectiveness of our approach with an extensive evaluation on the Pascal-VOC 2012 and ADE20K datasets, significantly outperforming state of the art incremental learning methods.

Modeling the Background for Incremental Learning in Semantic Segmentation / Cermelli, Fabio; Mancini, Massimiliano; Bulo, Samuel Rota; Ricci, Elisa; Caputo, Barbara. - (2020), pp. 9230-9239. ( 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020 usa 14 June- 19 June 2020) [10.1109/CVPR42600.2020.00925].

Modeling the Background for Incremental Learning in Semantic Segmentation

Mancini, Massimiliano;Bulo, Samuel Rota;Ricci, Elisa;
2020-01-01

Abstract

Despite their effectiveness in a wide range of tasks, deep architectures suffer from some important limitations. In particular, they are vulnerable to catastrophic forgetting, i.e. they perform poorly when they are required to update their model as new classes are available but the original training set is not retained. This paper addresses this problem in the context of semantic segmentation. Current strategies fail on this task because they do not consider a peculiar aspect of semantic segmentation: since each training step provides annotation only for a subset of all possible classes, pixels of the background class (i.e. pixels that do not belong to any other classes) exhibit a semantic distribution shift. In this work we revisit classical incremental learning methods, proposing a new distillation-based framework which explicitly accounts for this shift. Furthermore, we introduce a novel strategy to initialize classifier's parameters, thus preventing biased predictions toward the ba...
2020
Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
10662 LOS VAQUEROS CIRCLE, PO BOX 3014, LOS ALAMITOS, CA 90720-1264 USA
IEEE Computer Society
978-1-7281-7168-5
Cermelli, Fabio; Mancini, Massimiliano; Bulo, Samuel Rota; Ricci, Elisa; Caputo, Barbara
Modeling the Background for Incremental Learning in Semantic Segmentation / Cermelli, Fabio; Mancini, Massimiliano; Bulo, Samuel Rota; Ricci, Elisa; Caputo, Barbara. - (2020), pp. 9230-9239. ( 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020 usa 14 June- 19 June 2020) [10.1109/CVPR42600.2020.00925].
File in questo prodotto:
File Dimensione Formato  
Cermelli_Modeling_the_Background_for_Incremental_Learning_in_Semantic_Segmentation_CVPR_2020_paper.pdf

accesso aperto

Tipologia: Post-print referato (Refereed author’s manuscript)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 1 MB
Formato Adobe PDF
1 MB Adobe PDF Visualizza/Apri
Modeling_the_Background_for_Incremental_Learning_in_Semantic_Segmentation.pdf

Solo gestori archivio

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 1.41 MB
Formato Adobe PDF
1.41 MB Adobe PDF   Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/440171
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 281
  • ???jsp.display-item.citation.isi??? 232
  • OpenAlex 280
social impact