首页> 外文会议>Annual conference on Neural Information Processing Systems >Sparse Additive Text Models with Low Rank Background
【24h】

Sparse Additive Text Models with Low Rank Background

机译:稀疏添加剂文本模型与低排名背景

获取原文

摘要

The sparse additive model for text modeling involves the sum-of-exp computing, whose cost is consuming for large scales. Moreover, the assumption of equal background across all classes/topics may be too strong. This paper extends to propose sparse additive model with low rank background (SAM-LRB) and obtains simple yet efficient estimation. Particularly, employing a double majorization bound, we approximate log-likelihood into a quadratic lower-bound without the log-sum-exp terms. The constraints of low rank and sparsity are then simply embodied by nuclear norm and ?_1-norm regularizers. Interestingly, we find that the optimization task of SAM-LRB can be transformed into the same form as in Robust PCA. Consequently, parameters of supervised SAM-LRB can be efficiently learned using an existing algorithm for Robust PCA based on accelerated proximal gradient. Besides the supervised case, we extend SAM-LRB to favor unsupervised and mul-tifaceted scenarios. Experiments on three real data demonstrate the effectiveness and efficiency of SAM-LRB, compared with a few state-of-the-art models.
机译:文本建模的稀疏添加剂模型涉及到exp计算,其成本为大尺度消耗。此外,在所有类/主题上的相同背景的假设可能太强烈。本文延伸到提出具有低等级背景(SAM-LRB)的稀疏添加剂模型,并获得简单但有效的估计。特别是,采用双大大化绑定,我们将对数似然近似于在没有日志和exp术语的情况下在二次下限中。然后,低等级和稀疏性的约束简单地体现了核标准和?_1常规规则。有趣的是,我们发现SAM-LRB的优化任务可以转换为与鲁棒PCA中相同的形式。因此,可以使用基于加速的近端梯度,使用现有的鲁棒PCA算法有效地学习监督SAM-LRB的参数。除了受监督的情况外,我们还扩展了SAM-LRB,以支持无监督和多种程度的情景。与少数最先进的模型相比,三个真实数据的实验证明了SAM-LRB的有效性和效率。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号