首页> 外文会议>International workshop on semantic evaluation;Annual meeting of the Association for Computational Linguistics >SEW-EMBED at SemEval-2017 Task 2: Language-Independent Concept Representations from a Semantically Enriched Wikipedia
【24h】

SEW-EMBED at SemEval-2017 Task 2: Language-Independent Concept Representations from a Semantically Enriched Wikipedia

机译:SEW-EMBED在SemEval-2017任务2:来自语义丰富的维基百科的与语言无关的概念表示

获取原文

摘要

This paper describes SEW-EMBED, our language-independent approach to multilingual and cross-lingual semantic word similarity as part of the SemEval-2017 Task 2. We leverage the Wikipedia-based concept representations developed by Raganato et al. (2016), and propose an embedded augmentation of their explicit high-dimensional vectors, which we obtain by plugging in an arbitrary word (or sense) embedding representation, and computing a weighted average in the continuous vector space. We evaluate Sew-EMBED with two different off-the-shelf embedding representations, and report their performances across all monolingual and cross-lingual benchmarks available for the task. Despite its simplicity, especially compared with supervised or overly tuned approaches, SEW-EMBED achieves competitive results in the cross-lingual setting (3rd best result in the global ranking of subtask 2, score 0.56).
机译:本文描述了SEW-EMBED,这是我们与语言无关的多语言和跨语言语义词相似性方法,是SemEval-2017 Task 2的一部分。我们利用Raganato等人开发的基于Wikipedia的概念表示。 (2016年),并提出了其显式高维向量的嵌入扩充,这是通过插入任意词(或有义)嵌入表示形式并在连续向量空间中计算加权平均值而获得的。我们使用两种不同的现成嵌入表示来评估Sew-EMBED,并在可用于该任务的所有单语言和跨语言基准测试中报告其性能。尽管SEW-EMBED非常简单,尤其是与有监督或过度调整的方法相比,它在跨语言环境中仍取得了竞争性的结果(在子任务2的全球排名中,第三名的最佳成绩是0.56)。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号