Existing self-supervised learning methods learn representation by means of pretext tasks which are either (1) dis-criminating that explicitly specify which features should be separated or (2) aligning that precisely indicate which features should be closed together, but ignore the fact how to jointly and principally define which features to be repelled and which ones to be attracted. In this work, we combine the positive aspects of the discriminating and aligning methods, and design a hybrid method that addresses the above issue. Our method explicitly specifies the repulsion and attraction mechanism respectively by discriminative predictive task and concurrently maximizing mutual information between paired views sharing redundant information. We qualitatively and quantitatively show that our proposed model learns better features that are more effective for the diverse downstream tasks ranging from classification to semantic segmentation. Our experiments on nine established benchmarks s...

Existing self-supervised learning methods learn representation by means of pretext tasks which are either (1) dis-criminating that explicitly specify which features should be separated or (2) aligning that precisely indicate which features should be closed together, but ignore the fact how to jointly and principally define which features to be repelled and which ones to be attracted. In this work, we combine the positive aspects of the discriminating and aligning methods, and design a hybrid method that addresses the above issue. Our method explicitly specifies the repulsion and attraction mechanism respectively by discriminative predictive task and concurrently maximizing mutual information between paired views sharing redundant information. We qualitatively and quantitatively show that our proposed model learns better features that are more effective for the diverse downstream tasks ranging from classification to semantic segmentation. Our experiments on nine established benchmarks show that the proposed model consistently outperforms the existing state-of-the-art results of self-supervised and transfer learning protocol. Code can be found at https://github.com/AnjanDutta/codial.

Concurrent Discrimination and Alignment for Self-Supervised Feature Learning / Dutta, Anjan; Mancini, Massimiliano; Akata, Zeynep. - 2021-:(2021), pp. 2189-2198. ( 18th IEEE/CVF International Conference on Computer Vision Workshops, ICCVW 2021 can 2021) [10.1109/ICCVW54120.2021.00248].

Concurrent Discrimination and Alignment for Self-Supervised Feature Learning

Massimiliano Mancini;
2021-01-01

Abstract

Existing self-supervised learning methods learn representation by means of pretext tasks which are either (1) dis-criminating that explicitly specify which features should be separated or (2) aligning that precisely indicate which features should be closed together, but ignore the fact how to jointly and principally define which features to be repelled and which ones to be attracted. In this work, we combine the positive aspects of the discriminating and aligning methods, and design a hybrid method that addresses the above issue. Our method explicitly specifies the repulsion and attraction mechanism respectively by discriminative predictive task and concurrently maximizing mutual information between paired views sharing redundant information. We qualitatively and quantitatively show that our proposed model learns better features that are more effective for the diverse downstream tasks ranging from classification to semantic segmentation. Our experiments on nine established benchmarks s...
2021
2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW)
10662 LOS VAQUEROS CIRCLE, PO BOX 3014, LOS ALAMITOS, CA 90720-1264 USA
Institute of Electrical and Electronics Engineers Inc.
978-1-6654-0191-3
Dutta, Anjan; Mancini, Massimiliano; Akata, Zeynep
Concurrent Discrimination and Alignment for Self-Supervised Feature Learning / Dutta, Anjan; Mancini, Massimiliano; Akata, Zeynep. - 2021-:(2021), pp. 2189-2198. ( 18th IEEE/CVF International Conference on Computer Vision Workshops, ICCVW 2021 can 2021) [10.1109/ICCVW54120.2021.00248].
File in questo prodotto:
File Dimensione Formato  
Concurrent_Discrimination_and_Alignment_for_Self-Supervised_Feature_Learning.pdf

Solo gestori archivio

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 9 MB
Formato Adobe PDF
9 MB Adobe PDF   Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/437734
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? 0
  • OpenAlex 0
social impact