Recent advances in Representation Learning have discovered a strong inclination for pre-trained word embeddings to demonstrate unfair and discriminatory gender stereotypes. These usually come in the shape of unjustified associations between representations of group words (e.g., male or female) and attribute words (e.g. driving, cooking, doctor, nurse, etc.) In this paper, we propose an iterative and adversarial procedure to reduce gender bias in word vectors. We aim to remove gender influence from word representations that should otherwise be free of it, while retaining meaningful gender information in words that are inherently charged with gender polarity (male or female). We confine these gender signals in a sub-vector of word embeddings to make them more interpretable. Quantitative and qualitative experiments confirm that our method successfully reduces gender bias in pre-trained word embeddings with minimal semantic offset.

Iterative Adversarial Removal of Gender Bias in Pretrained Word Embeddings / Gaci, Y; Benatallah, B; Casati, F; Benabdeslem, K. - (2022), pp. 829-836. (Intervento presentato al convegno SAC 22 tenutosi a Online nel 25 - 29 April 2022) [10.1145/3477314.3507274].

Iterative Adversarial Removal of Gender Bias in Pretrained Word Embeddings

Benatallah, B;Casati, F;
2022-01-01

Abstract

Recent advances in Representation Learning have discovered a strong inclination for pre-trained word embeddings to demonstrate unfair and discriminatory gender stereotypes. These usually come in the shape of unjustified associations between representations of group words (e.g., male or female) and attribute words (e.g. driving, cooking, doctor, nurse, etc.) In this paper, we propose an iterative and adversarial procedure to reduce gender bias in word vectors. We aim to remove gender influence from word representations that should otherwise be free of it, while retaining meaningful gender information in words that are inherently charged with gender polarity (male or female). We confine these gender signals in a sub-vector of word embeddings to make them more interpretable. Quantitative and qualitative experiments confirm that our method successfully reduces gender bias in pre-trained word embeddings with minimal semantic offset.
2022
SAC '22: Proceedings of the 37th ACM/SIGAPP Symposium on Applied Computing
1601 Broadway, 10th Floor, NEW YORK, NY, UNITED STATES
ASSOC COMPUTING MACHINERY
9781450387132
Gaci, Y; Benatallah, B; Casati, F; Benabdeslem, K
Iterative Adversarial Removal of Gender Bias in Pretrained Word Embeddings / Gaci, Y; Benatallah, B; Casati, F; Benabdeslem, K. - (2022), pp. 829-836. (Intervento presentato al convegno SAC 22 tenutosi a Online nel 25 - 29 April 2022) [10.1145/3477314.3507274].
File in questo prodotto:
File Dimensione Formato  
SAC___Iterative_Adversarial_removal_of_gender_bias (1).pdf

accesso aperto

Tipologia: Post-print referato (Refereed author’s manuscript)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 943.83 kB
Formato Adobe PDF
943.83 kB Adobe PDF Visualizza/Apri
3477314.3507274.pdf

Solo gestori archivio

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 1.32 MB
Formato Adobe PDF
1.32 MB Adobe PDF   Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/397742
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 6
  • ???jsp.display-item.citation.isi??? 3
social impact