【24h】

TopicCheck: Interactive Alignment for Assessing Topic Model Stability

机译:TopicCheck:用于评估主题模型稳定性的交互式对齐

获取原文

摘要

Content analysis, a widely-applied social science research method, is increasingly being supplemented by topic modeling. However, while the discourse on content analysis centers heavily on reproducibility, computer scientists often focus more on scalability and less on coding reliability, leading to growing skepticism on the usefulness of topic models for automated content analysis. In response, we introduce TopicCheck, an interactive tool for assessing topic model stability. Our contributions are threefold. First, from established guidelines on reproducible content analysis, we distill a set of design requirements on how to computationally assess the stability of an automated coding process. Second, we devise an interactive alignment algorithm for matching latent topics from multiple models, and enable sensitivity evaluation across a large number of models. Finally, we demonstrate that our tool enables social scientists to gain novel insights into three active research questions.
机译:内容分析是一种广泛应用的社会科学研究方法,正越来越多地由主题建模来补充。但是,尽管有关内容分析的论述主要集中在可重复性上,但计算机科学家经常将重点更多地放在可伸缩性上,而不是在编码可靠性上,这导致人们越来越怀疑主题模型对自动内容分析的实用性。作为回应,我们介绍了TopicCheck,这是一种用于评估主题模型稳定性的交互式工具。我们的贡献是三倍。首先,从已建立的有关可重复内容分析的指南中,我们提炼出了一组关于如何计算评估自动编码过程的稳定性的设计要求。其次,我们设计了一种交互式比对算法,用于匹配来自多个模型的潜在主题,并能够对大量模型进行灵敏度评估。最后,我们证明了我们的工具使社会科学家能够获得对三个活跃研究问题的新颖见解。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号