首页> 外文会议>1st EMNLP workshop blackboxNLP: analyzing and interpreting neural networks for NLP 2018 >Trying to Understand Recurrent Neural Networks for Language Processing
【24h】

Trying to Understand Recurrent Neural Networks for Language Processing

机译:试图了解递归神经网络进行语言处理

获取原文
获取原文并翻译 | 示例

摘要

Recurrent neural networks (RNNs), and in particular LSTM networks, emerge as very capable learners for sequential data. Thus, my group started using them everywhere, achieving strong results on many language understanding and modeling tasks. However, little is known about how RNNs represent sequences, what they actually encode, and what they are capable representing. In this talk, Ⅰ will describe some attempts at trying to shed light on the inner-working of RNNs. Particularly, Ⅰ plan to describe at least two of the following: a method for comparing what is captured in vector representations of sentences based on different encoders (Adi et al, ICLR 2017, and more generally the notion of diagnostic classification), a framework for extracting a finite-state automata from trained RNNs (Weiss et al, ICML 2018), and a formal difference between the representation capacity of different RNN variants (Weiss et al, ACL 2018).
机译:递归神经网络(RNN),尤其是LSTM网络,是非常有能力的顺序数据学习者。因此,我的小组开始在各处使用它们,在许多语言理解和建模任务上取得了丰硕的成果。但是,对于RNN如何表示序列,它们实际编码的内容以及它们能够表示的内容知之甚少。在本次演讲中,我将描述一些试图阐明RNN的内部工作的尝试。特别是,我计划至少描述以下两项:一种用于比较基于不同编码器的句子的矢量表示中捕获的内容的方法(Adi等人,ICLR 2017,更笼统的诊断分类概念),从经过训练的RNN中提取有限状态自动机(Weiss等,ICML 2018),以及不同RNN变体的表示能力之间的形式差异(Weiss等,ACL 2018)。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号