...
首页> 外文期刊>International Journal of Computer Vision >Exploiting Privileged Information from Web Data for Action and Event Recognition
【24h】

Exploiting Privileged Information from Web Data for Action and Event Recognition

机译:利用Web数据中的特权信息进行操作和事件识别

获取原文
获取原文并翻译 | 示例

摘要

In the conventional approaches for action and event recognition, sufficient labelled training videos are generally required to learn robust classifiers with good generalization capability on new testing videos. However, collecting labelled training videos is often time consuming and expensive. In this work, we propose new learning frameworks to train robust classifiers for action and event recognition by using freely available web videos as training data. We aim to address three challenging issues: (1) the training web videos are generally associated with rich textual descriptions, which are not available in test videos; (2) the labels of training web videos are noisy and may be inaccurate; (3) the data distributions between training and test videos are often considerably different. To address the first two issues, we propose a new framework called multi-instance learning with privileged information (MIL-PI) together with three new MIL methods, in which we not only take advantage of the additional textual descriptions of training web videos as privileged information, but also explicitly cope with noise in the loose labels of training web videos. When the training and test videos come from different data distributions, we further extend our MIL-PI as a new framework called domain adaptive MIL-PI. We also propose another three new domain adaptation methods, which can additionally reduce the data distribution mismatch between training and test videos. Comprehensive experiments for action and event recognition demonstrate the effectiveness of our proposed approaches.
机译:在用于动作和事件识别的常规方法中,通常需要足够的带标签的训练视频来学习在新的测试视频上具有良好泛化能力的鲁棒分类器。但是,收集带有标签的培训视频通常很耗时且昂贵。在这项工作中,我们提出了新的学习框架,通过使用免费提供的网络视频作为训练数据来训练用于动作和事件识别的强大分类器。我们的目标是解决三个具有挑战性的问题:(1)培训网络视频通常与丰富的文字说明相关联,而测试视频中没有这些内容; (2)培训网络视频的标签嘈杂,可能不准确; (3)培训视频和测试视频之间的数据分配通常存在很大差异。为了解决前两个问题,我们提出了一个新的框架,即具有特权信息的多实例学习(MIL-PI)和三种新的MIL方法,其中我们不仅利用了特权培训网络视频的其他文字描述,信息,但也可以明确应对培训网络视频的松散标签中的噪音。当培训和测试视频来自不同的数据分布时,我们进一步将MIL-PI扩展为称为域自适应MIL-PI的新框架。我们还提出了另外三种新的领域自适应方法,它们可以另外减少训练视频和测试视频之间的数据分配不匹配。动作和事件识别的综合实验证明了我们提出的方法的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号