Equitable access to reliable health information is vital for public health, but the quality of online health resources varies by language, raising concerns about inconsistencies in Large Language Models (LLMs) for healthcare. In this study, we examine the consistency of responses provided by LLMs to health-related questions across English, German, Turkish, and Chinese. We largely expand the HealthFC dataset by categorizing health-related questions by disease type and broadening its multilingual scope with Turkish and Chinese translations. We reveal significant inconsistencies in responses that could spread healthcare misinformation. Our main contributions are 1) a multilingual health-related inquiry dataset with meta-information on disease categories, and 2) a novel prompt-based evaluation workflow that enables sub-dimensional comparisons between two languages through parsing. Our findings highlight key challenges in deploying LLM-based tools in multilingual contexts and emphasize the need for improved cross-lingual alignment to ensure accurate and equitable healthcare information.

Do LLMs Provide Consistent Answers to Health-Related Questions Across Languages? / Schlicht, Ipek Baris; Zhao, Zhixue; Sayin, Burcu; Flek, Lucie; Rosso, Paolo. - 15574 LNCS:(2025), pp. 314-322. ( 47th European Conference on Information Retrieval, ECIR 2025 Lucca, Italy 06-10 April 2025) [10.1007/978-3-031-88714-7_30].

Do LLMs Provide Consistent Answers to Health-Related Questions Across Languages?

Sayin, Burcu;
2025-01-01

Abstract

Equitable access to reliable health information is vital for public health, but the quality of online health resources varies by language, raising concerns about inconsistencies in Large Language Models (LLMs) for healthcare. In this study, we examine the consistency of responses provided by LLMs to health-related questions across English, German, Turkish, and Chinese. We largely expand the HealthFC dataset by categorizing health-related questions by disease type and broadening its multilingual scope with Turkish and Chinese translations. We reveal significant inconsistencies in responses that could spread healthcare misinformation. Our main contributions are 1) a multilingual health-related inquiry dataset with meta-information on disease categories, and 2) a novel prompt-based evaluation workflow that enables sub-dimensional comparisons between two languages through parsing. Our findings highlight key challenges in deploying LLM-based tools in multilingual contexts and emphasize the need for improved cross-lingual alignment to ensure accurate and equitable healthcare information.
2025
Advances in Information Retrieval. ECIR 2025. Lecture Notes in Computer Science, vol 15574.
Cham, Switzerland
Springer Science and Business Media Deutschland GmbH
9783031887130
9783031887147
Schlicht, Ipek Baris; Zhao, Zhixue; Sayin, Burcu; Flek, Lucie; Rosso, Paolo
Do LLMs Provide Consistent Answers to Health-Related Questions Across Languages? / Schlicht, Ipek Baris; Zhao, Zhixue; Sayin, Burcu; Flek, Lucie; Rosso, Paolo. - 15574 LNCS:(2025), pp. 314-322. ( 47th European Conference on Information Retrieval, ECIR 2025 Lucca, Italy 06-10 April 2025) [10.1007/978-3-031-88714-7_30].
File in questo prodotto:
File Dimensione Formato  
978-3-031-88714-7_30 (1).pdf

Solo gestori archivio

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 1.98 MB
Formato Adobe PDF
1.98 MB Adobe PDF   Visualizza/Apri
Do LLMs Provide Consistent Answers to Health-Related Questions across Languages.pdf

embargo fino al 04/04/2026

Tipologia: Post-print referato (Refereed author’s manuscript)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 726.43 kB
Formato Adobe PDF
726.43 kB Adobe PDF   Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/465810
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 1
  • ???jsp.display-item.citation.isi??? ND
  • OpenAlex ND
social impact