首页> 外文会议>Annual meeting of the Association for Computational Linguistics >Semantically Equivalent Adversarial Rules for Debugging NLP Models
【24h】

Semantically Equivalent Adversarial Rules for Debugging NLP Models

机译:用于调试NLP模型的语义等效的对抗条件

获取原文

摘要

Complex machine learning models for NLP are often brittle, making different predictions for input instances that are extremely similar semantically. To automatically detect this behavior for individual instances, we present semantically equivalent adversaries (SEAs) - semantic-preserving perturbations that induce changes in the model's predictions. We generalize these adversaries into semantically equivalent adversarial rules (SEARs) - simple, universal replacement rules that induce adversaries on many instances. We demonstrate the usefulness and flexibility of SEAs and SEARs by detecting bugs in black-box state-of-the-art models for three domains: machine comprehension, visual question-answering, and sentiment analysis. Via user studies, we demonstrate that we generate high-quality local adversaries for more instances than humans, and that SEARs induce four times as many mistakes as the bugs discovered by human experts. SEARs are also actionable: retraining models using data augmentation significantly reduces bugs, while maintaining accuracy.
机译:复杂的机器学习模型NLP往往易碎,因此对于那些极其相似语义上输入情况下不同的预测。自动检测此行为个别情况下,我们提出了语义上等价的对手有限公司(SEAS) - 语义保留摄动诱导模型的预测变化。我们概括这些对手成语义等价对抗规则(西尔斯) - 简单,通用的替换规则诱发许多情况下对手。机器理解,视觉问题回答和情感分析:我们通过检测在国家的最先进的黑盒模型错误的三个领域展示海洋和西尔斯的实用性和灵活性。通过用户研究,我们证明了我们产生比人类更实例当地优质的敌人,西尔斯诱导四倍多的错误作为错误发现人类专家。西尔斯也是可行的:使用再培训模型的数据增强显著减少错误,同时保持准确度。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号