【24h】

Colorless green recurrent networks dream hierarchically

机译:无色绿色循环网络分层实现梦想

获取原文

摘要

Recurrent neural networks (RNNs) have achieved impressive results in a variety of linguistic processing tasks, suggesting that they can induce non-trivial properties of language. We investigate here to what extent RNNs leam to track abstract hierarchical syntactic structure. We test whether RNNs trained with a generic language modeling objective in four languages (Italian, English, Hebrew, Russian) can predict long-distance number agreement in various constructions. We include in our evaluation nonsensical sentences where RNNs cannot rely on semantic or lexical cues ("The colorless green ideas I ate with the chair sleep furiously"), and, for Italian, we compare model performance to human intuitions. Our language-model-trained RNNs make reliable predictions about long-distance agreement, and do not lag much behind human performance. We thus bring support to the hypothesis that RNNs are not just shallow-pattern extractors, but they also acquire deeper grammatical competence.
机译:递归神经网络(RNN)在各种语言处理任务中均取得了令人印象深刻的结果,表明它们可以诱导语言的非平凡特性。我们在这里研究RNN在多大程度上可以追踪抽象的层次句法结构。我们测试了以四种语言(意大利语,英语,希伯来语,俄语)进行了通用语言建模目标训练的RNN是否可以预测各种构造中的长途号码协议。我们在评估中包括了无意义的句子,其中RNN不能依赖语义或词汇提示(“我在椅子上吃的无色绿色想法疯狂地睡着了”),并且,对于意大利语,我们将模型性能与人类直觉进行了比较。我们经过语言模型训练的RNN可以对长途协议做出可靠的预测,并且不会对人类的表现造成很大的影响。因此,我们为RNN不仅是浅模式提取器的假设提供了支持,而且它们还获得了更深的语法能力。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号