【24h】

Detoxifying Language Models Risks Marginalizing Minority Voices

机译:解毒语言模型风险限制边缘化少数民族声音

获取原文

摘要

Language models (LMs) must be both safe and equitable to be responsibly deployed in practice. With safety in mind, numerous detoxification techniques (e.g., Dathathri et al. 2020; Krause et al. 2020) have been proposed to mitigate toxic LM generations. In this work, we show that these detoxification techniques hurt equity: they decrease the utility of LMs on language used by marginalized groups (e.g., African-American English and minority identity mentions). In particular, we perform automatic and human evaluations of text generation quality when LMs are conditioned on inputs with different dialects and group identifiers. We find that detoxification makes LMs more brittle to distribution shift, especially on language used by marginalized groups. We identify that these failures stem from detoxification methods exploiting spurious correlations in toxicity datasets. Overall, our results highlight the tension between the controllability and distributional robustness of LMs.
机译:语言模型(LMS)必须安全且公平地在实践中负责任地部署。 为了安全记,许多解毒技术(例如,Dathathri等,2020; Krause等,2020)已提出减轻毒性LM代。 在这项工作中,我们表明这些排毒技术受到公平的伤害:它们减少了边缘化群体使用的语言的LMS的效用(例如,非洲裔美国英语和少数民族身份提出)。 特别是,当LMS在具有不同方言和组标识符的输入上调节LMS时,我们执行文本生成质量的自动和人为评估。 我们发现解毒使LMS更加脆弱,尤其是边缘化群体使用的语言。 我们确定这些失败源于解毒方法,利用毒性数据集的杂散相关性。 总体而言,我们的结果突出了LMS的可控性和分布稳健性之间的张力。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号