首页> 外文会议>IEEE International Conference on Data Mining Workshops >Intrinsic or Extrinsic Evaluation: An Overview of Word Embedding Evaluation
【24h】

Intrinsic or Extrinsic Evaluation: An Overview of Word Embedding Evaluation

机译:内部或外部评估:词嵌入评估概述

获取原文

摘要

Compared with traditional methods, word em-bedding is an efficient language representation that can learn syntax and semantics by using neural networks. As the result, more and more promising experiments in natural language processing (NLP) get the state-of-the-art results by introducing word embedding. In principle, embedding representation learning embeds words to a low-dimensional vector space, there-fore vectors support initialization of NLP tasks such as text classification, sentiment analysis, language understanding, etc. However, polysemy is very common in many languages, which causes word ambiguation, further influences the accuracy of the system. Additionally, language models based on distributed hypotheses mostly focused on word properties rather than morphology were our primary focus. This leads to unreasonable performance in different evaluations. At the same time, word embedding learning and measuring are two vital components of word representation. In this paper, we overviewed many language models including single sense and multiple sense word embedding, and many evaluated approaches including intrinsic and extrinsic evaluation. We found that there are obvious gaps between vectors and manual annotations in word similarity evaluation, and language models that achieved good performance in intrinsic evaluations could not produce similar results in extrinsic evaluations. To the best of our knowledge, there is no universal language model and embedding learning method for most NLP task, and each evaluations also hidden natural defects compared to human knowledge. More evaluated datasets are also investigated such as datasets used in intrinsic and extrinsic evaluations. We believe that an improved evaluation dataset and a more rational evaluation method would benefit from this overview.
机译:与传统方法相比,词嵌入是一种有效的语言表示形式,可以使用神经网络来学习语法和语义。结果,越来越多有前途的自然语言处理(NLP)实验通过引入词嵌入技术获得了最先进的结果。原则上,嵌入表示学习将单词嵌入到低维向量空间中,因此向量支持NLP任务的初始化,例如文本分类,情感分析,语言理解等。但是,多义性在许多语言中非常普遍,这导致单词歧义,进一步影响系统的准确性。此外,我们主要关注基于分布式假设的语言模型,这些语言模型主要关注词的属性而不是词法。这会导致在不同评估中出现不合理的性能。同时,词嵌入学习和度量是词表示的两个重要组成部分。在本文中,我们概述了许多语言模型,包括单义和多义词嵌入,以及许多评估方法,包括内在和外在评估。我们发现在单词相似性评估中向量和手动注释之间存在明显的差距,在固有评估中表现良好的语言模型在外部评估中无法产生相似的结果。据我们所知,大多数NLP任务都没有通用的语言模型和嵌入学习方法,并且与人类知识相比,每次评估也都隐藏了自然缺陷。还研究了更多评估的数据集,例如用于内部和外部评估的数据集。我们认为,改进的评估数据集和更合理的评估方法将从此概述中受益。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号