首页> 外文会议>International joint conference on natural language processing >A Multilingual Topic Model for Learning Weighted Topic Links Across Corpora with Low Comparability
【24h】

A Multilingual Topic Model for Learning Weighted Topic Links Across Corpora with Low Comparability

机译:具有低可比性跨学习加权主题链接的多语言主题模型

获取原文

摘要

Multilingual topic models (MTMs) learn topics on documents in multiple languages. Past models align topics across languages by implicitly assuming the documents in different languages are highly comparable, often a false assumption. We introduce a new model that does not rely on this assumption, particularly useful in important low-resource language scenarios. Our MTM learns weighted topic links and connects cross-lingual topics only when the dominant words defining them are similar, outperforming LDA and previous MTMs in classification tasks using documents' topic posteriors as features. It also learns coherent topics on documents with low comparability.
机译:多语言主题模型(MTMS)在多种语言中学习文档的主题。过去的模型通过隐式假设不同语言的文档来对齐跨语言的主题是非常可比的,通常是错误的假设。我们介绍了一个不依赖此假设的新模型,特别适用于重要的低资源语言情景。我们的MTM学习加权主题链接,并仅在定义它们的主导单词在分类任务中使用文档的主题后页中的定义单词和之前的MTMS作为特征时,只有当定义它们的主导单词和之前的MTMS时,才能连接到交叉语言。它还在具有低可比性的文档上学习相干主题。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号