Concept Bottleneck Models (CBMs) are neural networks designed to conjoin high perfor- mance with ante-hoc interpretability. CBMs work by first mapping inputs (e.g., images) to high-level concepts (e.g., visible objects and their properties) and then use these to solve a downstream task (e.g., tagging or scoring an image) in an interpretable manner. Their performance and interpretability, however, hinge on the quality of the concepts they learn. The go-to strategy for ensuring good quality concepts is to leverage expert annota- tions, which are expensive to collect and seldom available in applications. Researchers have recently addressed this issue by introducing “VLM-CBM” architectures that replace manual annotations with weak supervision from foundation models. It is however unclear what the impact of doing so is on the quality of the learned concepts. To answer this question, we put state-of-the-art VLM-CBMs to the test, analyzing their learned concepts empirically using a selection of significant metrics. Our results show that, depending on the task, VLM supervision can noticeably differ from expert annotations, and that concept accuracy and quality are not strongly correlated.

If Concept Bottlenecks are the Question, are Foundation Models the Answer? / Debole, Nicola; Barbiero, Pietro; Giannini, Francesco; Passerini, Andrea; Teso, Stefano; Marconato, Emanuele. - In: MACHINE LEARNING. - ISSN 0885-6125. - ELETTRONICO. - 115:5(2026), pp. 1-37. [10.1007/s10994-026-06999-y]

If Concept Bottlenecks are the Question, are Foundation Models the Answer?

Debole, Nicola
Primo
;
Passerini, Andrea;Teso, Stefano
Co-ultimo
;
Marconato, Emanuele
Co-ultimo
2026-01-01

Abstract

Concept Bottleneck Models (CBMs) are neural networks designed to conjoin high perfor- mance with ante-hoc interpretability. CBMs work by first mapping inputs (e.g., images) to high-level concepts (e.g., visible objects and their properties) and then use these to solve a downstream task (e.g., tagging or scoring an image) in an interpretable manner. Their performance and interpretability, however, hinge on the quality of the concepts they learn. The go-to strategy for ensuring good quality concepts is to leverage expert annota- tions, which are expensive to collect and seldom available in applications. Researchers have recently addressed this issue by introducing “VLM-CBM” architectures that replace manual annotations with weak supervision from foundation models. It is however unclear what the impact of doing so is on the quality of the learned concepts. To answer this question, we put state-of-the-art VLM-CBMs to the test, analyzing their learned concepts empirically using a selection of significant metrics. Our results show that, depending on the task, VLM supervision can noticeably differ from expert annotations, and that concept accuracy and quality are not strongly correlated.
2026
5
Settore IINF-05/A - Sistemi di elaborazione delle informazioni
Debole, Nicola; Barbiero, Pietro; Giannini, Francesco; Passerini, Andrea; Teso, Stefano; Marconato, Emanuele
If Concept Bottlenecks are the Question, are Foundation Models the Answer? / Debole, Nicola; Barbiero, Pietro; Giannini, Francesco; Passerini, Andrea; Teso, Stefano; Marconato, Emanuele. - In: MACHINE LEARNING. - ISSN 0885-6125. - ELETTRONICO. - 115:5(2026), pp. 1-37. [10.1007/s10994-026-06999-y]
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/487298
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
  • OpenAlex ND
social impact