首页> 外文会议>IEEE World Symposium on Applied Machine Intelligence and Informatics >You Are What You Read: The Effect of Corpus and Training Task on Semantic Absorption in Recurrent Neural Architectures
【24h】

You Are What You Read: The Effect of Corpus and Training Task on Semantic Absorption in Recurrent Neural Architectures

机译:您就是所读的内容:语料库和训练任务对递归神经体系结构中语义吸收的影响

获取原文

摘要

Recurrent neural networks are able to capture semantic meaning within the geometry of the resultant embedding space, but the impact of training corpus and training curriculum on the structure of the learned representations remains poorly understood. This paper sheds some light on the mystery by comparing the performance of a simple recurrent model on three training tasks and two strongly divergent training corpora. The learned representations are compared on tasks including the Semantic Textual Similarity Benchmark, the Stanford Natural Language Inference Corpus, and Google’s Analogical Reasoning Test Set. Results show that context-based training produces the strongest semantic alignment within the embedding space, with reconstruction loss as an interesting close second. Pairwise comparisons of models trained on different corpora show that the choice of corpus also has powerful effects on the learned representations. Most importantly, we observe that the choice of input corpus and training task are not unilaterally independent, but instead interact with each other in interesting ways. This motivates a cautionary position against training neural models by simply throwing as many different training tasks as possible into the mix. Instead, it may be wiser to carefully select only tasks that are compatible with the chosen input corpus, and vice versa.
机译:递归神经网络能够捕获合成的嵌入空间的几何形状内的语义,但是培训语料库和培训课程对学习表示的结构的影响仍然知之甚少。通过比较简单的递归模型在三个训练任务和两个强烈分歧的训练语料库上的表现,本文揭示了这个奥秘。将学习到的表示形式在包括语义文本相似性基准测试,斯坦福自然语言推理语料库和Google的类比推理测试集在内的任务上进行比较。结果表明,基于上下文的训练在嵌入空间内产生最强的语义对齐,而重构损失则是紧随其后的有趣事件。在不同语料库上训练的模型的成对比较显示,语料库的选择对学习的表示形式也有强大的影响。最重要的是,我们观察到输入语料库的选择和培训任务不是单方面独立的,而是以有趣的方式相互影响的。通过简单地将尽可能多的不同训练任务投入混合,这会激发人们对训练神经模型的警惕。相反,明智的选择是仅谨慎选择与所选输入语料库兼容的任务,反之亦然。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号