【24h】

Cross-Topic Distributional Semantic Representations Via Unsupervised Mappings

机译:通过无监督映射的跨主题分布语义表示

获取原文

摘要

In traditional Distributional Semantic Models (DSMs) the multiple senses of a polyse-mous word are conflated into a single vector space representation. In this work, we propose a DSM that learns multiple distributional representations of a word based on different topics. First, a separate DSM is trained for each topic and then each of the topic-based DSMs is aligned to a common vector space. Our unsupervised mapping approach is motivated by the hypothesis that words preserving their relative distances in different topic semantic sub-spaces constitute robust semantic anchors that define the mappings between them. Aligned cross-topic representations achieve state-of-the-art results for the task of contextual word similarity. Furthermore, evaluation on NLP downstream tasks shows that multiple topic-based embeddings outperform single-prototype models.
机译:在传统的分布语义模型(DSM)中,多义词的多种含义被合并为单个向量空间表示形式。在这项工作中,我们提出了一种DSM,可以基于不同的主题学习单词的多个分布表示形式。首先,为每个主题训练一个单独的DSM,然后将每个基于主题的DSM与一个公共向量空间对齐。我们的无监督映射方法受到以下假设的启发:在不同主题语义子空间中保留其相对距离的单词构成了定义它们之间映射的强大语义锚。对齐的跨主题表示法可实现上下文词相似性任务的最新结果。此外,对NLP下游任务的评估表明,多个基于主题的嵌入优于单一原型模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号