首页> 外文会议>Pacific Rim international conference on artificial intelligence >Mini-Batch Variational Inference for Time-Aware Topic Modeling
【24h】

Mini-Batch Variational Inference for Time-Aware Topic Modeling

机译:用于时间感知主题建模的小批量变分推理

获取原文

摘要

This paper proposes a time-aware topic model and its mini-batch variational inference for exploring chronological trends in document contents. Our contribution is twofold. First, to extract topics in a time-aware manner, our method uses two vector embeddings: the embedding of latent topics and that of document timestamps. By combining these two embeddings and applying the softmax function, we have as many word probability distributions as document timestamps for each topic. This modeling enables us to extract remarkable topical trends. Second, to achieve memory efficiency, the variational inference is implemented as mini-batch gradient ascent maximizing the evidence lower bound. This enables us to perform parameter estimation in the way similar to neural networks. Our method was actually implemented with deep learning framework. The evaluation results show that we could improve test set perplexity by using document timestamps and also that our test perplexity was comparable with that of collapsed Gibbs sampling, which is less efficient in memory usage than the proposed inference.
机译:本文提出了一个时间感知主题模型及其小批量变分推理,以探索文档内容中的时间趋势。我们的贡献是双重的。首先,为了以时间感知的方式提取主题,我们的方法使用了两个向量嵌入:潜在主题的嵌入和文档时间戳的嵌入。通过结合这两个嵌入并应用softmax函数,我们每个主题的单词概率分布与文档时间戳一样多。这种建模使我们能够提取出显着的主题趋势。其次,为了实现存储效率,将变分推理实现为最小批量梯度上升以最大化证据下界。这使我们能够以类似于神经网络的方式执行参数估计。我们的方法实际上是通过深度学习框架实现的。评估结果表明,我们可以通过使用文档时间戳来提高测试集的困惑度,并且我们的测试困惑度与塌陷的Gibbs采样相当,后者在内存使用方面的效率低于所提出的推断。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号