Disentangling how gender and occupations are encoded by LLMs is crucial to identify possible biases and prevent harms, especially given the widespread use of LLMs in sensitive domains such as human resources.In this work, we carry out an in-depth investigation of gender and occupational biases in English and Italian as expressed by 9 different LLMs (both base and instruction-tuned). Specifically, we focus on the analysis of sentence completions when LLMs are prompted with job-related sentences including different gender representations. We carry out a manual analysis of 4,500 generated texts over 4 dimensions that can reflect bias, we propose a novel embedding-based method to investigate biases in generated texts and, finally, we carry out a lexical analysis of the model completions. In our qualitative and quantitative evaluation we show that many facets of social bias remain unaccounted for even in aligned models, and LLMs in general still reflect existing gender biases in both languages. Finally, we find that models still struggle with gender-neutral expressions, especially beyond English.

Job Unfair: An Investigation of Gender and Occupational Bias in Free-Form Text Completions by LLMs / Casula, Camilla; Vecellio Salto, Sebastiano; Leonardelli, Elisa; Tonelli, Sara. - (2025), pp. 22770-22788. ( 2025 Conference on Empirical Methods in Natural Language Processing, EMNLP 2025 Suzhou, China 2025) [10.18653/v1/2025.emnlp-main.1159].

Job Unfair: An Investigation of Gender and Occupational Bias in Free-Form Text Completions by LLMs

Casula, Camilla;Vecellio Salto, Sebastiano;Leonardelli, Elisa;Tonelli, Sara
2025-01-01

Abstract

Disentangling how gender and occupations are encoded by LLMs is crucial to identify possible biases and prevent harms, especially given the widespread use of LLMs in sensitive domains such as human resources.In this work, we carry out an in-depth investigation of gender and occupational biases in English and Italian as expressed by 9 different LLMs (both base and instruction-tuned). Specifically, we focus on the analysis of sentence completions when LLMs are prompted with job-related sentences including different gender representations. We carry out a manual analysis of 4,500 generated texts over 4 dimensions that can reflect bias, we propose a novel embedding-based method to investigate biases in generated texts and, finally, we carry out a lexical analysis of the model completions. In our qualitative and quantitative evaluation we show that many facets of social bias remain unaccounted for even in aligned models, and LLMs in general still reflect existing gender biases in both languages. Finally, we find that models still struggle with gender-neutral expressions, especially beyond English.
2025
EMNLP 2025 - 2025 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference
Suzhou, China
Association for Computational Linguistics (ACL)
Casula, Camilla; Vecellio Salto, Sebastiano; Leonardelli, Elisa; Tonelli, Sara
Job Unfair: An Investigation of Gender and Occupational Bias in Free-Form Text Completions by LLMs / Casula, Camilla; Vecellio Salto, Sebastiano; Leonardelli, Elisa; Tonelli, Sara. - (2025), pp. 22770-22788. ( 2025 Conference on Empirical Methods in Natural Language Processing, EMNLP 2025 Suzhou, China 2025) [10.18653/v1/2025.emnlp-main.1159].
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/469610
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
  • OpenAlex ND
social impact