Over the past years, semantic segmentation, as many other tasks in computer vision, benefited from the progress in deep neural networks, resulting in significantly improved performance. However, deep architectures trained with gradient-based techniques suffer from catastrophic forgetting, which is the tendency to forget previously learned knowledge while learning new tasks. Aiming at devising strategies to counteract this effect, incremental learning approaches have gained popularity over the past years. However, the first incremental learning methods for semantic segmentation appeared only recently. While effective, these approaches do not account for a crucial aspect in pixel-level dense prediction problems, i.e. the role of attention mechanisms. To fill this gap, in this paper we introduce a novel attentive feature distillation approach to mitigate catastrophic forgetting while accounting for semantic spatial- and channel-level dependencies. Furthermore, we propose a {cross-modal attention} structure, which takes advantage of the attention learned from the new and the old tasks while learning features for the new task. Finally, we also introduce a novel strategy to account for the background class in the distillation loss, thus preventing biased predictions. We demonstrate the effectiveness of our approach with an extensive evaluation on Pascal-VOC 2012 and ADE20K, setting a new state of the art.

Continual Attentive Fusion for Incremental Learning in Semantic Segmentation / Yang, G.; Fini, E.; Xu, D.; Rota, P.; Ding, M.; Hao, T.; Alameda-Pineda, X.; Ricci, E.. - In: IEEE TRANSACTIONS ON MULTIMEDIA. - ISSN 1520-9210. - 2022:(2022), pp. 1-14. [10.1109/TMM.2022.3167555]

Continual Attentive Fusion for Incremental Learning in Semantic Segmentation

Fini E.;Xu D.;Rota P.;Ding M.;Ricci E.
2022-01-01

Abstract

Over the past years, semantic segmentation, as many other tasks in computer vision, benefited from the progress in deep neural networks, resulting in significantly improved performance. However, deep architectures trained with gradient-based techniques suffer from catastrophic forgetting, which is the tendency to forget previously learned knowledge while learning new tasks. Aiming at devising strategies to counteract this effect, incremental learning approaches have gained popularity over the past years. However, the first incremental learning methods for semantic segmentation appeared only recently. While effective, these approaches do not account for a crucial aspect in pixel-level dense prediction problems, i.e. the role of attention mechanisms. To fill this gap, in this paper we introduce a novel attentive feature distillation approach to mitigate catastrophic forgetting while accounting for semantic spatial- and channel-level dependencies. Furthermore, we propose a {cross-modal attention} structure, which takes advantage of the attention learned from the new and the old tasks while learning features for the new task. Finally, we also introduce a novel strategy to account for the background class in the distillation loss, thus preventing biased predictions. We demonstrate the effectiveness of our approach with an extensive evaluation on Pascal-VOC 2012 and ADE20K, setting a new state of the art.
2022
Yang, G.; Fini, E.; Xu, D.; Rota, P.; Ding, M.; Hao, T.; Alameda-Pineda, X.; Ricci, E.
Continual Attentive Fusion for Incremental Learning in Semantic Segmentation / Yang, G.; Fini, E.; Xu, D.; Rota, P.; Ding, M.; Hao, T.; Alameda-Pineda, X.; Ricci, E.. - In: IEEE TRANSACTIONS ON MULTIMEDIA. - ISSN 1520-9210. - 2022:(2022), pp. 1-14. [10.1109/TMM.2022.3167555]
File in questo prodotto:
File Dimensione Formato  
2202.00432 (1) (1).pdf

accesso aperto

Tipologia: Post-print referato (Refereed author’s manuscript)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 1.41 MB
Formato Adobe PDF
1.41 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/341664
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 12
  • ???jsp.display-item.citation.isi??? 11
social impact