首页> 外文会议>International Joint Conference on Natural Language Processing;Annual Meeting of the Association for Computational Linguistics >Regression Bugs Are In Your Model! Measuring, Reducing and Analyzing Regressions In NLP Model Updates
【24h】

Regression Bugs Are In Your Model! Measuring, Reducing and Analyzing Regressions In NLP Model Updates

机译:回归错误在您的模型中! NLP模型更新中的回归测量,减少和分析回归

获取原文

摘要

Behavior of deep neural networks can be inconsistent between different versions. Regressions 1 during model update are a common cause of concern that often over-weigh the benefits in accuracy or efficiency gain. This work focuses on quantifying, reducing and analyzing regression errors in the NLP model updates. Using negative flip rate as regression measure, we show that regression has a prevalent presence across tasks in the GLUE benchmark. We formulate the regression-free model updates into a constrained optimization problem, and further reduce it into a relaxed form which can be approximately optimized through knowledge distillation training method. We empirically analyze how model ensemble reduces regression. Finally, we conduct CHECKLIST behavioral testing to understand the distribution of regressions across linguistic phenomena, and the efficacy of ensemble and distillation methods.
机译:深度神经网络的行为可能在不同版本之间不一致。 在模型更新期间的回归1是常见的令人担忧的原因,这通常会在准确性或效率增益中超越益处。 这项工作侧重于量化,减少和分析NLP模型更新中的回归误差。 使用负面翻转率作为回归措施,我们显示回归在胶水基准中的任务中具有普遍存在的存在。 我们将回归模型更新分为约束的优化问题,并进一步将其降低到宽松的形式中,通过知识蒸馏训练方法可以大致优化。 我们经验分析了模型集合如何降低回归。 最后,我们进行清单行为测试以了解跨语言现象的回归分布,以及集合和蒸馏方法的疗效。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号