【24h】

Privacy-Aware Text Rewriting

机译:隐私感知文本重写

获取原文

摘要

Biased decisions made by automatic systems have led to growing concerns in research communities. Recent work from the NLP community focuses on building systems that make fair decisions based on text. Instead of relying on unknown decision systems or human decisionmakers, we argue that a better way to protect data providers is to remove the trails of sensitive information before publishing the data. In light of this, we propose a new privacy-aware text rewriting task and explore two privacy-aware back-translation methods for the task, based on adversarial training and approximate fairness risk. Our extensive experiments on three real-world datasets with varying demo-graphical attributes show that our methods are effective in obfuscating sensitive attributes. We have also observed that the fairness risk method retains better semantics and fluency, while the adversaria] training method tends to leak less sensitive information.
机译:自动系统做出的有偏见的决策引起了研究界越来越大的关注。 NLP社区最近的工作集中在构建基于文本做出公平决策的系统。我们认为,保护数据提供者的更好方法不是依赖未知的决策系统或人为的决策者,而是在发布数据之前删除敏感信息的痕迹。有鉴于此,我们提出了一个新的具有隐私意识的文本重写任务,并基于对抗性训练和近似公平风险,探索了该任务的两种具有隐私意识的反向翻译方法。我们对具有不同人口统计属性的三个现实世界数据集进行的广泛实验表明,我们的方法可以有效地混淆敏感属性。我们还观察到,公平风险方法保留了更好的语义和流畅性,而对抗训练方法则倾向于泄漏不太敏感的信息。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号