【24h】

State Gradients for RNN Memory Analysis

机译:RNN内存分析的状态梯度

获取原文

摘要

We present a framework for analyzing what the state in RNNs remembers from its input embeddings. We compute the gradients of the states with respect to the input embeddings and decompose the gradient matrix with Singular Value Decomposition to analyze which directions in the embedding space are best transferred to the hidden state space, characterized by the largest singular values. We apply our approach to LSTM language models and investigate to what extent and for how long certain classes of words are remembered on average for a certain corpus. Additionally, the extent to which a specific property or relationship is remembered by the RNN can be tracked by comparing a vector characterizing that property with the direction(s) in embedding space that are best preserved in hidden state space.
机译:我们介绍了一个框架,用于分析RNN中的状态从其输入嵌入中记住的状态。我们对输入嵌入的梯度进行计算,并将梯度矩阵与奇异值分解分解,以分析嵌入空间中的哪些方向最佳地传送到隐藏状态空间,其特征是由最大奇异值的特征。我们将我们的方法应用于LSTM语言模型,并调查某种程度,以及某些语料库的平均记住某些单词的程度。另外,通过将特征在于在隐藏状态空间中最好保留的嵌入空间中的方向,可以通过比较标志的载体来跟踪由RNN记住特定财产或关系的程度。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号