There is growing interest in concept-based models (CBMs) that combine highperformance and interpretability by acquiring and reasoning with a vocabulary of high-level concepts. A key requirement is that the concepts be interpretable. Existing CBMs tackle this desideratum using a variety of heuristics based on unclear notions of interpretability, and fail to acquire concepts with the intended semantics. We address this by providing a clear definition of interpretability in terms of alignment between the model’s representation and an underlying data generation process, and introduce GlanceNets, a new CBM that exploits techniques from disentangled representation learning and open-set recognition to achieve alignment, thus improving the interpretability of the learned concepts. We show that GlanceNets, paired with concept-level supervision, achieve better alignment than state-of-the-art approaches while preventing spurious concepts from unintentionally affecting its predictions. The code is available at https://github.com/ema-marconato/glancenet.

GlanceNets: Interpretabile, Leak-proof Concept-based Models / Marconato, Emanuele; Passerini, Andrea; Teso, Stefano. - ELETTRONICO. - 35:(2022), pp. 1-23. (Intervento presentato al convegno 36th Conference on Neural Information Processing Systems, NeurIPS 2022 tenutosi a New Orleans, United States nel 28/11/2022 - 9/12/2022).

GlanceNets: Interpretabile, Leak-proof Concept-based Models

Marconato, Emanuele;Passerini, Andrea;Teso, Stefano
2022-01-01

Abstract

There is growing interest in concept-based models (CBMs) that combine highperformance and interpretability by acquiring and reasoning with a vocabulary of high-level concepts. A key requirement is that the concepts be interpretable. Existing CBMs tackle this desideratum using a variety of heuristics based on unclear notions of interpretability, and fail to acquire concepts with the intended semantics. We address this by providing a clear definition of interpretability in terms of alignment between the model’s representation and an underlying data generation process, and introduce GlanceNets, a new CBM that exploits techniques from disentangled representation learning and open-set recognition to achieve alignment, thus improving the interpretability of the learned concepts. We show that GlanceNets, paired with concept-level supervision, achieve better alignment than state-of-the-art approaches while preventing spurious concepts from unintentionally affecting its predictions. The code is available at https://github.com/ema-marconato/glancenet.
2022
Advances in Neural Information Processing Systems 35 (NeurIPS 2022)
San Diego, CA
Neural information processing systems foundation
9781713871088
Marconato, Emanuele; Passerini, Andrea; Teso, Stefano
GlanceNets: Interpretabile, Leak-proof Concept-based Models / Marconato, Emanuele; Passerini, Andrea; Teso, Stefano. - ELETTRONICO. - 35:(2022), pp. 1-23. (Intervento presentato al convegno 36th Conference on Neural Information Processing Systems, NeurIPS 2022 tenutosi a New Orleans, United States nel 28/11/2022 - 9/12/2022).
File in questo prodotto:
File Dimensione Formato  
glancenets___neurips22.pdf

accesso aperto

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 3.87 MB
Formato Adobe PDF
3.87 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/364701
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 14
  • ???jsp.display-item.citation.isi??? ND
  • OpenAlex ND
social impact