A prominent goal of representation learning research is to achieve representations which are factorized in a useful manner with respect to the ground truth factors of variation. The fields of disentangled and equivariant representation learning have approached this ideal from a range of complimentary perspectives; however, to date, most approaches have proven to either be ill-specified or insufficiently flexible to effectively separate all realistic factors of interest in a learned latent space. In this work, we propose an alternative viewpoint on such structured representation learning which we call Flow Factorized Representation Learning, and demonstrate it to learn both more efficient and more usefully structured representations than existing frameworks. Specifically, we introduce a generative model which specifies a distinct set of latent probability paths that define different input transformations. Each latent flow is generated by the gradient field of a learned potential following dynamic optimal transport. Our novel setup brings new understandings to both disentanglement and equivariance. We show that our model achieves higher likelihoods on standard representation learning benchmarks while simultaneously being closer to approximately equivariant models. Furthermore, we demonstrate that the transformations learned by our model are flexibly composable and can also extrapolate to new data, implying a degree of robustness and generalizability approaching the ultimate goal of usefully factorized representation learning. © 2023 Neural information processing systems foundation. All rights reserved.

Flow Factorized Representation Learning / Song, Yue; Anderson Keller, T.; Sebe, Nicu; Welling, Max. - 36:(2023), pp. 49761-49782. ( 37th Conference on Neural Information Processing Systems, NeurIPS 2023 New Orleans December 2023).

Flow Factorized Representation Learning

Yue Song
;
Nicu Sebe;
2023-01-01

Abstract

A prominent goal of representation learning research is to achieve representations which are factorized in a useful manner with respect to the ground truth factors of variation. The fields of disentangled and equivariant representation learning have approached this ideal from a range of complimentary perspectives; however, to date, most approaches have proven to either be ill-specified or insufficiently flexible to effectively separate all realistic factors of interest in a learned latent space. In this work, we propose an alternative viewpoint on such structured representation learning which we call Flow Factorized Representation Learning, and demonstrate it to learn both more efficient and more usefully structured representations than existing frameworks. Specifically, we introduce a generative model which specifies a distinct set of latent probability paths that define different input transformations. Each latent flow is generated by the gradient field of a learned potential following dynamic optimal transport. Our novel setup brings new understandings to both disentanglement and equivariance. We show that our model achieves higher likelihoods on standard representation learning benchmarks while simultaneously being closer to approximately equivariant models. Furthermore, we demonstrate that the transformations learned by our model are flexibly composable and can also extrapolate to new data, implying a degree of robustness and generalizability approaching the ultimate goal of usefully factorized representation learning. © 2023 Neural information processing systems foundation. All rights reserved.
2023
Thirty-seventh Conference on Neural Information Processing Systems (NeurIPS 2023)
New York
Openreview
9781713899921
Song, Yue; Anderson Keller, T.; Sebe, Nicu; Welling, Max
Flow Factorized Representation Learning / Song, Yue; Anderson Keller, T.; Sebe, Nicu; Welling, Max. - 36:(2023), pp. 49761-49782. ( 37th Conference on Neural Information Processing Systems, NeurIPS 2023 New Orleans December 2023).
File in questo prodotto:
File Dimensione Formato  
2717_flow_factorized_representation (1)-min.pdf

accesso aperto

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 3.62 MB
Formato Adobe PDF
3.62 MB Adobe PDF Visualizza/Apri
NeurIPS-2023-flow-factorized-representation-learning-Supplemental-Conference.pdf

accesso aperto

Descrizione: Supplementary Material
Tipologia: Altro materiale allegato (Other attachments)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 9.05 MB
Formato Adobe PDF
9.05 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/401003
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 3
  • ???jsp.display-item.citation.isi??? 0
  • OpenAlex 0
social impact