【24h】

Exploiting Attention to Reveal Shortcomings in Memory Models

机译:利用注意力来揭示记忆模型中的缺点

获取原文
获取原文并翻译 | 示例

摘要

The decision making processes of deep networks are difficult to understand and while their accuracy often improves with increased architectural complexity, so too does their opacity. Practical use of machine learning models, especially for question and answering applications, demands a system that is inter-pretable. We analyze the attention of a memory network model to reconcile contradictory performance on a challenging question-answering dataset that is inspired by theory-of-mind experiments. We equate success on questions to task classification, which explains not only test-time failures but also how well the model generalizes to new training conditions.
机译:深度网络的决策过程很难理解,尽管其准确性通常会随着体系结构复杂性的提高而提高,但其不透明性也是如此。机器学习模型的实际使用,尤其是对于问答应用程序,要求系统是可预知的。我们分析了记忆网络模型的注意,以调和受理论理论实验启发的具有挑战性的问答数据集上的矛盾性能。我们将问题上的成功等同于任务分类,这不仅解释了测试时失败,还解释了该模型对新的训练条件的推广程度。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号