Semantic change detection (SCD) involves temporal changes and spatial semantics. Its working principle and processing flow usually include land semantic segmentation (LSS) and binary change detection (BCD). Due to its significant impact and practical value, SCD has received consistently wide attention in Earth observation. Nowadays, remote sensing (RS) data in various modalities are proliferating, calling for an urgent need to develop intelligent algorithms for multimodal RS data. However, no efficient multimodal SCD methods exist currently. To address this limitation, this work proposes the first deep learning-based multimodal SCD method: MSCD-Net. MSCD-Net extracts multiscale semantic and difference features after fusing multimodal features, and then aggregates and refines these features to output high-quality semantic segmentation and change maps. Additionally, a semantic difference decoder (SDD) module is designed to model semantic and difference features jointly. It can be integrated with existing methods to increase accuracy. Experimental results demonstrate that MSCD-Net achieves state-of-the-art performance on both multimodal and unimodal SCD datasets, and SDD has strong feature learning ability and compatibility. These findings imply that MSCD-Net is expected to promote the development and application of multimodal SCD.

Semantic change detection (SCD) involves temporal changes and spatial semantics. Its working principle and processing flow usually include land semantic segmentation (LSS) and binary change detection (BCD). Due to its significant impact and practical value, SCD has received consistently wide attention in Earth observation. Nowadays, remote sensing (RS) data in various modalities are proliferating, calling for an urgent need to develop intelligent algorithms for multimodal RS data. However, no efficient multimodal SCD methods exist currently. To address this limitation, this work proposes the first deep learning-based multimodal SCD method: MSCD-Net. MSCD-Net extracts multiscale semantic and difference features after fusing multimodal features, and then aggregates and refines these features to output high-quality semantic segmentation and change maps. Additionally, a semantic difference decoder (SDD) module is designed to model semantic and difference features jointly. It can be integrated with existing methods to increase accuracy. Experimental results demonstrate that MSCD-Net achieves state-of-the-art performance on both multimodal and unimodal SCD datasets, and SDD has strong feature learning ability and compatibility. These findings imply that MSCD-Net is expected to promote the development and application of multimodal SCD.

MSCD-Net: From Unimodal to Multimodal Semantic Change Detection / Wang, Jian; Xie, Hong; Yan, Li; Zhou, Tingyuan; Wang, Yanheng; Zhang, Jing; Bruzzone, Lorenzo; Atkinson, Peter M.. - In: IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING. - ISSN 1558-0644. - 63:4508017(2025), pp. 1-17. [10.1109/TGRS.2025.3591814]

MSCD-Net: From Unimodal to Multimodal Semantic Change Detection

Jing Zhang;Lorenzo Bruzzone;
2025-01-01

Abstract

Semantic change detection (SCD) involves temporal changes and spatial semantics. Its working principle and processing flow usually include land semantic segmentation (LSS) and binary change detection (BCD). Due to its significant impact and practical value, SCD has received consistently wide attention in Earth observation. Nowadays, remote sensing (RS) data in various modalities are proliferating, calling for an urgent need to develop intelligent algorithms for multimodal RS data. However, no efficient multimodal SCD methods exist currently. To address this limitation, this work proposes the first deep learning-based multimodal SCD method: MSCD-Net. MSCD-Net extracts multiscale semantic and difference features after fusing multimodal features, and then aggregates and refines these features to output high-quality semantic segmentation and change maps. Additionally, a semantic difference decoder (SDD) module is designed to model semantic and difference features jointly. It can be integrated with existing methods to increase accuracy. Experimental results demonstrate that MSCD-Net achieves state-of-the-art performance on both multimodal and unimodal SCD datasets, and SDD has strong feature learning ability and compatibility. These findings imply that MSCD-Net is expected to promote the development and application of multimodal SCD.
2025
4508017
Wang, Jian; Xie, Hong; Yan, Li; Zhou, Tingyuan; Wang, Yanheng; Zhang, Jing; Bruzzone, Lorenzo; Atkinson, Peter M.
MSCD-Net: From Unimodal to Multimodal Semantic Change Detection / Wang, Jian; Xie, Hong; Yan, Li; Zhou, Tingyuan; Wang, Yanheng; Zhang, Jing; Bruzzone, Lorenzo; Atkinson, Peter M.. - In: IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING. - ISSN 1558-0644. - 63:4508017(2025), pp. 1-17. [10.1109/TGRS.2025.3591814]
File in questo prodotto:
File Dimensione Formato  
TGRS3591814.pdf

accesso aperto

Descrizione: This article has been accepted for publication in IEEE Transactions on Geoscience and Remote Sensing. This is the author's version which has not been fully edited and content may change prior to final publication. Citation information: DOI 10.1109/TGRS.2025.3591814
Tipologia: Post-print referato (Refereed author’s manuscript)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 26.16 MB
Formato Adobe PDF
26.16 MB Adobe PDF Visualizza/Apri
MSCD-Net_From_Unimodal_to_Multimodal_Semantic_Change_Detection.pdf

Solo gestori archivio

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 14.21 MB
Formato Adobe PDF
14.21 MB Adobe PDF   Visualizza/Apri
MSCD-Net_From_Unimodal_to_Multimodal_Semantic_Change_Detection_compressed (1).pdf

Solo gestori archivio

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 3.97 MB
Formato Adobe PDF
3.97 MB Adobe PDF   Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/475691
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 2
  • ???jsp.display-item.citation.isi??? 1
  • OpenAlex 0
social impact