【24h】

Obtaining Better Static Word Embeddings Using Contextual Embedding Models

机译:使用上下文嵌入模型获取更好的静态单词嵌入

获取原文

摘要

The advent of contextual word embeddings- representations of words which incorporate semantic and syntactic information from their context-has led to tremendous improvements on a wide variety of NLP tasks. However, recent contextual models have prohibitively high computational cost in many use-cases and are often hard to interpret. In this work, we demonstrate that our proposed distillation method, which is a simple extension of CBOW-based training, allows to significantly improve computational efficiency of NLP applications, while outperforming the quality of existing static embeddings trained from scratch as well as those distilled from previously proposed methods. As a side-effect, our approach also allows a fair comparison of both contextual and static embeddings via standard lexical evaluation tasks.
机译:中文单词嵌入的出现 - 从他们的上下文中包含语义和句法信息的词语 - 导致各种NLP任务的巨大改进。 然而,最近的上下文模型在许多用例中具有过高的计算成本,并且通常很难解释。 在这项工作中,我们证明了我们提出的蒸馏方法,这是基于CBOW培训的简单延伸,可以显着提高NLP应用的计算效率,同时优于从头划痕培训的现有静态嵌入品的质量以及蒸馏出来的现有静态嵌入品的质量。 以前提出的方法。 作为副作用,我们的方法还允许通过标准的词法评估任务进行体内和静态嵌入的公平比较。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号