Text encoders have recently been convicted of en-coding unjustified social stereotypes, which lead models to make biased and prejudiced predictions when trained on downstream tasks such as sentiment analysis or question answering. The presence of bias in NLP models is dangerous since it promotes the divide between different social groups. Thus, attempts at mitigating biases from NLP models constitute an active line of research. However, these methods assume that models replicate exactly the same stereotypes ingrained in society's impression, leading to potential inaccuracies in the normative framing of bias. In this work we confirm that text encoders are indeed biased. Nonetheless, we show that encoded biases are slightly different from survey-based biases proper to human prejudice. We ground our findings on the Stereotype Content Model, an acclaimed framework to interpret stereotypes, prejudice and inter-group relations in social psychology.

Societal Versus Encoded Stereotypes in Text Encoders / Gaci, Y.; Benatallah, B.; Casati, F.; Benabdeslem, K.. - (2023), pp. 46-53. ( 35th IEEE International Conference on Tools with Artificial Intelligence, ICTAI 2023 Atlanta, GA, USA nov 2023) [10.1109/ICTAI59109.2023.00015].

Societal Versus Encoded Stereotypes in Text Encoders

Benatallah B.;Casati F.;
2023-01-01

Abstract

Text encoders have recently been convicted of en-coding unjustified social stereotypes, which lead models to make biased and prejudiced predictions when trained on downstream tasks such as sentiment analysis or question answering. The presence of bias in NLP models is dangerous since it promotes the divide between different social groups. Thus, attempts at mitigating biases from NLP models constitute an active line of research. However, these methods assume that models replicate exactly the same stereotypes ingrained in society's impression, leading to potential inaccuracies in the normative framing of bias. In this work we confirm that text encoders are indeed biased. Nonetheless, we show that encoded biases are slightly different from survey-based biases proper to human prejudice. We ground our findings on the Stereotype Content Model, an acclaimed framework to interpret stereotypes, prejudice and inter-group relations in social psychology.
2023
IEEE 35th International Conference on Tools with Artificial Intelligence (ICTAI)
10662 LOS VAQUEROS CIRCLE, PO BOX 3014, LOS ALAMITOS, CA 90720-1264 USA
IEEE
9798350342734
Gaci, Y.; Benatallah, B.; Casati, F.; Benabdeslem, K.
Societal Versus Encoded Stereotypes in Text Encoders / Gaci, Y.; Benatallah, B.; Casati, F.; Benabdeslem, K.. - (2023), pp. 46-53. ( 35th IEEE International Conference on Tools with Artificial Intelligence, ICTAI 2023 Atlanta, GA, USA nov 2023) [10.1109/ICTAI59109.2023.00015].
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/441175
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 1
  • ???jsp.display-item.citation.isi??? 1
  • OpenAlex 1
social impact