首页> 外文会议>International Workshop on Machine Learning for Multimodal Interaction >Meeting State Recognition from Visual and Aural Labels
【24h】

Meeting State Recognition from Visual and Aural Labels

机译:从视觉和听觉标签达到国家识别

获取原文

摘要

In this paper we present a meeting state recognizer based on a combination of multi-modal sensor data in a smart room. Our approach is based on the training of a statistical model to use semantical cues generated by perceptual components. These perceptual components generate these cues in processing the output of one or multiple sensors. The presented recognizer is designed to work with an arbitrary combination of multi-modal input sensors. We have defined a set of states representing both meeting and non-meeting situations, and a set of features we base our classification on. Thus, we can model situations like presentation or break which are important information for many applications. We have hand-annotated a set of meeting recordings to verify our statistical classification, as appropriate multi-modal corpora are currently very sparse. We have also used several statistical classification methods for the best classification, which we validated on the hand-annotated corpus of real meeting data.
机译:在本文中,我们基于智能房间中的多模态传感器数据的组合呈现会议状态识别器。我们的方法是基于培训统计模型来使用感知组件产生的语义线索。这些感知组件在处理一个或多个传感器的输出时生成这些提示。所提出的识别器旨在使用多模态输入传感器的任意组合。我们已经确定了一组代表会议和非会议情况的状态,以及我们基于分类的一组功能。因此,我们可以模拟演示文稿或中断的情况,这是许多应用的重要信息。我们手工注释了一系列会议录音,以验证我们的统计分类,因为适当的多模态语料目前是非常稀疏的。我们还使用了几种统计分类方法,以获得最佳分类,我们在真实会议数据的手动注释语料库上验证。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号