首页> 外文会议>International conference on computational linguistics >Subword-augmented Embedding for Cloze Reading Comprehension
【24h】

Subword-augmented Embedding for Cloze Reading Comprehension

机译:填充封闭阅读理解的子字增强嵌入

获取原文

摘要

Representation learning is the foundation of machine reading comprehension. In state-of-the-art models, deep learning methods hroadly use word and character level representations. However, character is not naturally the minimal linguistic unit. In addition, with a simple concatenation of character and word embedding, previous models actually give suboptimal solution. In this paper, we propose to use subword rather than character for word embedding enhancement. We also empirically explore different augmentation strategies on subword-augmented embedding to enhance the cloze-style reading comprehension model (reader). In detail, we present a reader that uses subword-level representation to augment word embedding with a short list to handle rare words effectively. A thorough examination is conducted to evaluate the comprehensive performance and generalization ability of the proposed reader. Experimental results show that the proposed approach helps the reader significantly outperform the state-of-the-art baselines on various public datasets.
机译:代表学习是机器阅读理解的基础。在最先进的模型中,深入学习方法的高纳使用Word和字符级别表示。然而,性格并不自然是最小的语言单位。此外,通过简单的字符和单词嵌入级联,之前的模型实际上是提供次优的解决方案。在本文中,我们建议使用子字而不是字符,以便嵌入增强词。我们还经验探索了子字增强嵌入的不同增强策略,以增强隐冻式阅读理解模型(读者)。详细地,我们介绍了一个读者,该读者使用子字级表示来增强单词嵌入短列表以有效地处理稀有单词。进行彻底检查以评估拟议读者的综合性能和泛化能力。实验结果表明,该方法有助于读者显着优于各种公共数据集的最先进的基线。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号