首页> 外文会议>Pacific Rim International Conference on Artificial Intelligence >Mini-Batch Variational Inference for Time-Aware Topic Modeling
【24h】

Mini-Batch Variational Inference for Time-Aware Topic Modeling

机译:迷你批量变化推论时间感知主题建模

获取原文

摘要

This paper proposes a time-aware topic model and its mini-batch variational inference for exploring chronological trends in document contents. Our contribution is twofold. First, to extract topics in a time-aware manner, our method uses two vector embeddings: the embedding of latent topics and that of document timestamps. By combining these two embeddings and applying the softmax function, we have as many word probability distributions as document timestamps for each topic. This modeling enables us to extract remarkable topical trends. Second, to achieve memory efficiency, the variational inference is implemented as mini-batch gradient ascent maximizing the evidence lower bound. This enables us to perform parameter estimation in the way similar to neural networks. Our method was actually implemented with deep learning framework. The evaluation results show that we could improve test set perplexity by using document timestamps and also that our test perplexity was comparable with that of collapsed Gibbs sampling, which is less efficient in memory usage than the proposed inference.
机译:本文提出了一个时光的主题模型及其迷你批量变异推断,用于探索文档内容的时间顺序趋势。我们的贡献是双重的。首先,要以时空地提取主题,我们的方法使用两个向量嵌入:嵌入潜在主题和文档时间戳的嵌入。通过组合这两个嵌入并应用软MAX函数,我们具有多个单词概率分布作为每个主题的文档时间戳。此型号使我们能够提取显着的题目趋势。其次,为了实现记忆效率,变分推理实现为迷你批量梯度上升,最大化证据下限。这使我们能够以类似于神经网络的方式执行参数估计。我们的方法实际上以深入学习框架实施。评估结果表明,我们可以通过使用文档时间戳来改善测试集的困惑,以及我们的测试困惑与折叠的GIBBS采样相当,这在内存使用中的效率低于所提出的推断。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号