【24h】

Transferable Neural Projection Representations

机译:可转移的神经投影表示

获取原文

摘要

Neural word representations are at the core of many state-of-the-art natural language processing models. A widely used approach is to pre-train, store and look up word or character embedding matrices. While useful, such representations occupy huge memory making it hard to deploy on-device and often do not generalize to unknown words due to vocabulary pruning. In this paper, we propose a skip-gram based architecture coupled with Locality-Sensitive Hashing (LSH) projections to learn efficient dynamically computable representations. Our model does not need to store lookup tables as representations are computed on-the-fly and require low memory footprint. The representations can be trained in an unsupervised fashion and can be easily transferred to other NLP tasks. For qualitative evaluation, we analyze the nearest neighbors of the word representations and discover semantically similar words even with misspellings. For quantitative evaluation, we plug our transferable projections into a simple LSTM and run it on multiple NLP tasks and show how our transferable projections achieve better performance compared to prior work.
机译:神经词表示法是许多最先进的自然语言处理模型的核心。一种广泛使用的方法是预训练,存储和查找单词或字符嵌入矩阵。尽管有用,但这种表示形式占用大量内存,因此很难在设备上进行部署,并且由于词汇修剪而常常无法推广到未知单词。在本文中,我们提出了一种基于跳跃语法的架构,并结合了局部敏感哈希(LSH)投影,以学习有效的动态可计算表示形式。我们的模型不需要存储查找表,因为可以动态计算表示形式,并且所需的内存占用少。可以以无监督的方式对表示进行训练,并且可以轻松地将其转移到其他NLP任务。为了进行定性评估,我们分析了单词表示形式的最近邻居,并发现了语义相似的单词,即使拼写错误也是如此。为了进行定量评估,我们将可转移的预测插入到简单的LSTM中,并在多个NLP任务上运行,并显示与以前的工作相比,我们的可转移的预测如何实现更好的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号