Text-to-image generative models are becoming increasingly popular and accessible to the general public. As these models see large-scale deployments, it is necessary to deeply investigate their safety and fairness to not disseminate and perpetuate any kind of biases. However, existing works focus on detecting closed sets of biases defined a priori, limiting the studies to well-known concepts. In this paper, we tackle the challenge of open-set bias detection in text-to-image generative models presenting OpenBias, a new pipeline that identifies and quantifies the severity of biases agnostically, without access to any precompiled set. OpenBias has three stages. In the first phase, we leverage a Large Language Model (LLM) to propose biases given a set of captions. Secondly, the target generative model produces images using the same set of captions. Lastly, a Vision Question Answering model recognizes the presence and extent of the previously proposed biases. We study the behavior of Stable Diffusion 1.5, 2, and XL emphasizing new biases, never investigated before. Via quantitative experiments, we demonstrate that OpenBias agrees with current closed-set bias detection methods and human judgement.

OpenBias: Open-Set Bias Detection in Text-to-Image Generative Models / D'Incà, Moreno; Peruzzo, Elia; Mancini, Massimiliano; Xu, Dejia; Goe, Vidit; Xu, Xingqian; Wang, Zhangyang; Shi, Humphrey; Sebe, Nicu. - (2024), pp. 12225-12235. (Intervento presentato al convegno IEEE/CVF Conference on Computer Vision and Pattern Recognition tenutosi a Seattle, WA, USA nel 16-22 June 2024) [10.1109/cvpr52733.2024.01162].

OpenBias: Open-Set Bias Detection in Text-to-Image Generative Models

D'Incà, Moreno;Peruzzo, Elia;Mancini, Massimiliano;Sebe, Nicu
2024-01-01

Abstract

Text-to-image generative models are becoming increasingly popular and accessible to the general public. As these models see large-scale deployments, it is necessary to deeply investigate their safety and fairness to not disseminate and perpetuate any kind of biases. However, existing works focus on detecting closed sets of biases defined a priori, limiting the studies to well-known concepts. In this paper, we tackle the challenge of open-set bias detection in text-to-image generative models presenting OpenBias, a new pipeline that identifies and quantifies the severity of biases agnostically, without access to any precompiled set. OpenBias has three stages. In the first phase, we leverage a Large Language Model (LLM) to propose biases given a set of captions. Secondly, the target generative model produces images using the same set of captions. Lastly, a Vision Question Answering model recognizes the presence and extent of the previously proposed biases. We study the behavior of Stable Diffusion 1.5, 2, and XL emphasizing new biases, never investigated before. Via quantitative experiments, we demonstrate that OpenBias agrees with current closed-set bias detection methods and human judgement.
2024
IEEE/CVF Conference on Computer Vision and Pattern Recognition
Piscataway, NJ USA
IEEE
979-8-3503-5300-6
979-8-3503-5301-3
D'Incà, Moreno; Peruzzo, Elia; Mancini, Massimiliano; Xu, Dejia; Goe, Vidit; Xu, Xingqian; Wang, Zhangyang; Shi, Humphrey; Sebe, Nicu
OpenBias: Open-Set Bias Detection in Text-to-Image Generative Models / D'Incà, Moreno; Peruzzo, Elia; Mancini, Massimiliano; Xu, Dejia; Goe, Vidit; Xu, Xingqian; Wang, Zhangyang; Shi, Humphrey; Sebe, Nicu. - (2024), pp. 12225-12235. (Intervento presentato al convegno IEEE/CVF Conference on Computer Vision and Pattern Recognition tenutosi a Seattle, WA, USA nel 16-22 June 2024) [10.1109/cvpr52733.2024.01162].
File in questo prodotto:
File Dimensione Formato  
DInca_OpenBias_Open-set_Bias_Detection_in_Text-to-Image_Generative_Models_CVPR_2024_paper (1).pdf

accesso aperto

Tipologia: Post-print referato (Refereed author’s manuscript)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 6.13 MB
Formato Adobe PDF
6.13 MB Adobe PDF Visualizza/Apri
OpenBias_Open-Set_Bias_Detection_in_Text-to-Image_Generative_Models.pdf

Solo gestori archivio

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 6.05 MB
Formato Adobe PDF
6.05 MB Adobe PDF   Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/430714
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 2
  • ???jsp.display-item.citation.isi??? ND
  • OpenAlex ND
social impact