论文标题

多种语言的蒙版语言模型中的性别偏见

Gender Bias in Masked Language Models for Multiple Languages

论文作者

Kaneko, Masahiro, Imankulova, Aizhan, Bollegala, Danushka, Okazaki, Naoaki

论文摘要

通过预测大型语料库上的蒙版令牌预测的蒙版语言模型(MLMS)已成功地用于各种语言的自然语言处理任务中。不幸的是,据报道,MLMS还学习有关性别和种族等属性的歧视性偏见。由于大多数研究都集中在英语中的MLMS上,因此很少研究MLMS在其他语言中的偏见。由于招募注释者的成本和困难,对英语以外的其他语言的评估数据的手动注释一直具有挑战性。此外,现有的偏见评估方法需要由属性单词组成的陈规定型句子对(例如,他/她是护士)。我们提出多语言偏见评估(MBE)得分,以使用英语属性单词列表和目标语言和英语之间的平行语言来评估各种语言的偏见,而无需手动注释数据。我们使用MBE以八种语言评估了MLM,并确认与性别相关的偏见是在所有这些语言中用MLMS编码的。我们手动创建了日语和俄语的性别偏见的数据集,以评估MBE的有效性。结果表明,MBE报告的偏差分数与从上面手动创建的数据集和现有英语数据集计算出的性别偏见的偏差分数显着相关。

Masked Language Models (MLMs) pre-trained by predicting masked tokens on large corpora have been used successfully in natural language processing tasks for a variety of languages. Unfortunately, it was reported that MLMs also learn discriminative biases regarding attributes such as gender and race. Because most studies have focused on MLMs in English, the bias of MLMs in other languages has rarely been investigated. Manual annotation of evaluation data for languages other than English has been challenging due to the cost and difficulty in recruiting annotators. Moreover, the existing bias evaluation methods require the stereotypical sentence pairs consisting of the same context with attribute words (e.g. He/She is a nurse). We propose Multilingual Bias Evaluation (MBE) score, to evaluate bias in various languages using only English attribute word lists and parallel corpora between the target language and English without requiring manually annotated data. We evaluated MLMs in eight languages using the MBE and confirmed that gender-related biases are encoded in MLMs for all those languages. We manually created datasets for gender bias in Japanese and Russian to evaluate the validity of the MBE. The results show that the bias scores reported by the MBE significantly correlates with that computed from the above manually created datasets and the existing English datasets for gender bias.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源