Text encoders have recently been convicted of en-coding unjustified social stereotypes, which lead models to make biased and prejudiced predictions when trained on downstream tasks such as sentiment analysis or question answering. The presence of bias in NLP models is dangerous since it promotes the divide between different social groups. Thus, attempts at mitigating biases from NLP models constitute an active line of research. However, these methods assume that models replicate exactly the same stereotypes ingrained in society's impression, leading to potential inaccuracies in the normative framing of bias. In this work we confirm that text encoders are indeed biased. Nonetheless, we show that encoded biases are slightly different from survey-based biases proper to human prejudice. We ground our findings on the Stereotype Content Model, an acclaimed framework to interpret stereotypes, prejudice and inter-group relations in social psychology.

Societal Versus Encoded Stereotypes in Text Encoders / Gaci, Yacine; Benatallah, Boualem; Casati, Fabio; Benabdeslem, Khalid. - (2023), pp. 46-53. ( 35th IEEE International Conference on Tools with Artificial Intelligence, ICTAI 2023 Atlanta, GA, USA nov 2023) [10.1109/ICTAI59109.2023.00015].

Societal Versus Encoded Stereotypes in Text Encoders

Boualem Benatallah;Fabio Casati;
2023-01-01

Abstract

Text encoders have recently been convicted of en-coding unjustified social stereotypes, which lead models to make biased and prejudiced predictions when trained on downstream tasks such as sentiment analysis or question answering. The presence of bias in NLP models is dangerous since it promotes the divide between different social groups. Thus, attempts at mitigating biases from NLP models constitute an active line of research. However, these methods assume that models replicate exactly the same stereotypes ingrained in society's impression, leading to potential inaccuracies in the normative framing of bias. In this work we confirm that text encoders are indeed biased. Nonetheless, we show that encoded biases are slightly different from survey-based biases proper to human prejudice. We ground our findings on the Stereotype Content Model, an acclaimed framework to interpret stereotypes, prejudice and inter-group relations in social psychology.
2023
2023 IEEE 35th International Conference on Tools with Artificial Intelligence (ICTAI)
10662 LOS VAQUEROS CIRCLE, PO BOX 3014, LOS ALAMITOS, CA 90720-1264 USA
IEEE
9798350342734
Gaci, Yacine; Benatallah, Boualem; Casati, Fabio; Benabdeslem, Khalid
Societal Versus Encoded Stereotypes in Text Encoders / Gaci, Yacine; Benatallah, Boualem; Casati, Fabio; Benabdeslem, Khalid. - (2023), pp. 46-53. ( 35th IEEE International Conference on Tools with Artificial Intelligence, ICTAI 2023 Atlanta, GA, USA nov 2023) [10.1109/ICTAI59109.2023.00015].
File in questo prodotto:
File Dimensione Formato  
Societal_Versus_Encoded_Stereotypes_in_Text_Encoders.pdf

Solo gestori archivio

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 620.68 kB
Formato Adobe PDF
620.68 kB Adobe PDF   Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/441175
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 1
  • ???jsp.display-item.citation.isi??? 2
  • OpenAlex 1
social impact