Cross-view image translation is challenging because it involves images with drastically different views and severe deformation. In this paper, we propose a novel approach named Multi-Channel Attention SelectionGAN (Selection-GAN) that makes it possible to generate images of natural scenes in arbitrary viewpoints, based on an image of the scene and a novel semantic map. The proposed SelectionGAN explicitly utilizes the semantic information and consists of two stages. In the first stage, the condition image and the target semantic map are fed into a cycled semantic-guided generation network to produce initial coarse results. In the second stage, we refine the initial results by using a multi-channel attention selection mechanism. Moreover, uncertainty maps automatically learned from attentions are used to guide the pixel loss for better network optimization. Extensive experiments on Dayton [41], CVUSA [43] and Ego2Top [1] datasets show that our model is able to generate significantly better results than the state-of-the-art methods. The source code, data and trained models are available at https://github. com/Ha0Tang/SelectionGAN.

Multi-Channel Attention Selection GAN With Cascaded Semantic Guidance for Cross-View Image Translation / Tang, Hao; Xu, Dan; Sebe, Nicu; Wang, Yanzhi; Corso, Jason J.; Yan, Yan. - (2019), pp. 2412-2421. (Intervento presentato al convegno IEEE Comference on Computer Vision and Pattern Recognition (CVPR'19) tenutosi a Long Beach nel 16-20 June, 2019) [10.1109/CVPR.2019.00252].

Multi-Channel Attention Selection GAN With Cascaded Semantic Guidance for Cross-View Image Translation

Tang, Hao;Xu, Dan;Sebe, Nicu;Yan, Yan
2019-01-01

Abstract

Cross-view image translation is challenging because it involves images with drastically different views and severe deformation. In this paper, we propose a novel approach named Multi-Channel Attention SelectionGAN (Selection-GAN) that makes it possible to generate images of natural scenes in arbitrary viewpoints, based on an image of the scene and a novel semantic map. The proposed SelectionGAN explicitly utilizes the semantic information and consists of two stages. In the first stage, the condition image and the target semantic map are fed into a cycled semantic-guided generation network to produce initial coarse results. In the second stage, we refine the initial results by using a multi-channel attention selection mechanism. Moreover, uncertainty maps automatically learned from attentions are used to guide the pixel loss for better network optimization. Extensive experiments on Dayton [41], CVUSA [43] and Ego2Top [1] datasets show that our model is able to generate significantly better results than the state-of-the-art methods. The source code, data and trained models are available at https://github. com/Ha0Tang/SelectionGAN.
2019
IEEE Comference on Computer Vision and Pattern Recognition (CVPR'19)
New York
IEEE
978-1-7281-3293-8
Tang, Hao; Xu, Dan; Sebe, Nicu; Wang, Yanzhi; Corso, Jason J.; Yan, Yan
Multi-Channel Attention Selection GAN With Cascaded Semantic Guidance for Cross-View Image Translation / Tang, Hao; Xu, Dan; Sebe, Nicu; Wang, Yanzhi; Corso, Jason J.; Yan, Yan. - (2019), pp. 2412-2421. (Intervento presentato al convegno IEEE Comference on Computer Vision and Pattern Recognition (CVPR'19) tenutosi a Long Beach nel 16-20 June, 2019) [10.1109/CVPR.2019.00252].
File in questo prodotto:
File Dimensione Formato  
Tang_Multi-Channel_Attention_Selection_GAN_With_Cascaded_Semantic_Guidance_for_Cross-View_CVPR_2019_paper.pdf

accesso aperto

Tipologia: Post-print referato (Refereed author’s manuscript)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 4.63 MB
Formato Adobe PDF
4.63 MB Adobe PDF Visualizza/Apri
08954196.pdf

Solo gestori archivio

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 4.17 MB
Formato Adobe PDF
4.17 MB Adobe PDF   Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/250763
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 181
  • ???jsp.display-item.citation.isi??? 141
  • OpenAlex ND
social impact