Gender bias in English-to-Greek machine translation

Publication type
C1
Publication status
Published
Authors
Gkovedarou, EG, Daems, J, & Bruyne, L.
Editor
Janica Hackenbuchner, Luisa Bentivogli, Joke Daems, Chiara Manna, Beatrice Savoldi and Eva Vanmassenhove
Series
Proceedings of the 3rd Workshop on Gender-Inclusive Translation Technologies (GITT 2025)
Pagination
17-45
Publisher
European Association for Machine Translation (EAMT)
Conference
3rd Workshop on Gender-Inclusive Translation Technologies (GITT 2025) (Geneva, Switzerland)
Download
(.pdf)
View in Biblio
(externe link)

Abstract

As the demand for inclusive language increases, concern has grown over the susceptibility of machine translation (MT) systems to reinforce gender stereotypes. This study investigates gender bias in two commercial MT systems, Google Translate and DeepL, focusing on the understudied English-to-Greek language pair. We address three aspects of gender bias: i) male bias, ii) occupational stereotyping, and iii) errors in anti-stereotypical translations. Additionally, we explore the potential of prompted GPT-4o as a bias mitigation tool that provides both gender-explicit and gender-neutral alternatives when necessary. To achieve this, we introduce GendEL, a manually crafted bilingual dataset of 240 gender-ambiguous and unambiguous sentences that feature stereotypical occupational nouns and adjectives. We find persistent gender bias in translations by both MT systems; while they perform well in cases where gender is explicitly defined, with DeepL outperforming both Google Translate and GPT-4o in feminine gender-unambiguous sentences, they are far from producing gender-inclusive or neutral translations when the gender is unspecified. GPT-4o shows promise, generating appropriate gendered and neutral alternatives for most ambiguous cases, though residual biases remain evident. As one of the first comprehensive studies on gender bias in English-to-Greek MT, we provide both our data and code at [github link].