...
【24h】

Supervised labeled latent Dirichlet allocation for document categorization

机译:监督标记的潜在Dirichlet分配用于文档分类

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Recently, supervised topic modeling approaches have received considerable attention. However, the representative labeled latent Dirichlet allocation (L-LDA) method has a tendency to over-focus on the pre-assigned labels, and does not give potentially lost labels and common semantics sufficient consideration. To overcome these problems, we propose an extension of L-LDA, namely supervised labeled latent Dirichlet allocation (SL-LDA), for document categorization. Our model makes two fundamental assumptions, i.e., Prior 1 and Prior 2, that relax the restriction of label sampling and extend the concept of topics. In this paper, we develop a Gibbs expectation-maximization algorithm to learn the SL-LDA model. Quantitative experimental results demonstrate that SL-LDA is competitive with state-of-the-art approaches on both single-label and multi-label corpora.
机译:最近,监督主题建模方法受到了相当大的关注。 然而,代表性标记的潜在Dirichlet分配(L-LDA)方法具有过度关注预先分配的标签的趋势,并且不会给出可能丢失的标签和常用语义充分考虑。 为了克服这些问题,我们提出了L-LDA的延伸,即监督标记的潜在Dirichlet分配(SL-LDA),用于文档分类。 我们的模型是两个基本的假设,即之前的1和事先2,放松了标签采样的限制并扩展了主题的概念。 在本文中,我们开发了GIBBS期望最大化算法来学习SL-LDA模型。 定量实验结果表明,SL-LDA对单一标签和多标签语料库的最先进方法具有竞争力。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号