首页> 外文会议>IEEE International Conference on Semantic Computing >HWE: Word Embedding with Heterogeneous Features
【24h】

HWE: Word Embedding with Heterogeneous Features

机译:HWE:具有异构特征的词嵌入

获取原文

摘要

Distributed word representation is widely used in Natural Language Processing. However, traditional approaches, which learn word representations from the co-occurrence information in large corpora, might not capture fine-grained syntactic and semantic information. In this paper, we propose a general and flexible framework to incorporate heterogeneous features (e.g., word-sense, part-of-speech, topic) for learning feature-specific word embeddings in an explicit fashion, namely Heterogeneous Word Embedding (HWE). Experimental results on both intrinsic and extrinsic tasks show that HWE outperforms the baseline and various state-of-the-art models. Moreover, through the concatenation over HWE and the corresponding feature embeddings, each word would have different contextual representation under different contexts, which achieves even more significant improvement. Finally, we illustrate the insight of our model via visualization of the learned word embeddings.
机译:分布式单词表示法在自然语言处理中被广泛使用。但是,从大语料库中的共现信息中学习单词表示的传统方法可能无法捕获细粒度的句法和语义信息。在本文中,我们提出了一个通用且灵活的框架,以结合异构特征(例如,词义,词性,主题)以显式方式学习特定于特征的词嵌入,即异构词嵌入(HWE)。内在和外在任务的实验结果表明,HWE的性能优于基线模型和各种最新模型。此外,通过在HWE上进行级联和相应的特征嵌入,每个单词在不同上下文中将具有不同的上下文表示,这将实现更大的改进。最后,我们通过可视化学习到的单词嵌入来说明我们的模型的洞察力。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号