首页> 外文会议>1st EMNLP workshop blackboxNLP: analyzing and interpreting neural networks for NLP 2018 >Predicting and interpreting embeddings for out of vocabulary words in downstream tasks
【24h】

Predicting and interpreting embeddings for out of vocabulary words in downstream tasks

机译:预测和解释下游任务中词汇外的嵌入

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

We propose a novel way to handle out of vocabulary (OOV) words in downstream natural language processing (NLP) tasks. We implement a network that predicts useful em-beddings for OOV words based on their morphology and on the context in which they appear. Our model also incorporates an attention mechanism indicating the focus allocated to the left context words, the right context words or the word's characters, hence making the prediction more interpretable. The model is a "drop-in" module that is jointly trained with the downstream task's neural network, thus producing embeddings specialized for the task at hand. When the task is mostly syntactical, we observe that our model aims most of its attention on surface form characters. On the other hand, for tasks more semantical, the network allocates more attention to the surrounding words. In all our tests, the module helps the network to achieve better performances in comparison to the use of simple random embeddings.
机译:我们提出了一种新颖的方法来处理下游自然语言处理(NLP)任务中的词汇(OOV)单词。我们实现了一个网络,该网络可以根据OOV词的形态和出现的上下文来预测有用的嵌入。我们的模型还包含一种注意机制,该机制指示分配给左上下文单词,右上下文单词或单词字符的焦点,因此使预测更具可解释性。该模型是一个“插入式”模块,与下游任务的神经网络共同训练,从而生成专门用于手头任务的嵌入。当任务主要是语法性的时,我们观察到我们的模型将大部分注意力集中在表面形状特征上。另一方面,对于更具语义的任务,网络将更多的注意力分配给周围的单词。在我们所有的测试中,与使用简单的随机嵌入相比,该模块可帮助网络获得更好的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号