首页> 外文期刊>Multimedia Tools and Applications >Tag refinement of micro-videos by learning from multiple data sources
【24h】

Tag refinement of micro-videos by learning from multiple data sources

机译:通过从多个数据源中学习来优化微视频的标签

获取原文
获取原文并翻译 | 示例
           

摘要

Micro-video is an increasingly prevalent social media form, which attracts much attention for its convenient acquisition and expressive ability. However, for the user-generated hashtags of micro-videos have seriously unbalanced distribution and low quality, the management of micro-videos becomes challenging. In this paper, we propose a novel tag refinement approach for micro-videos by learning from multiple public data sources with manually labelled tags, which can overcome the difficulty of directly refining the imprecise hashtags and address the problem of lacking manually labelled micro-video datasets for training. We define a set of target tags by referring to the widely used datasets for object, activity and scene detection. In tag refinement, we firstly transfer the tags from the images in NUS-WIDE to the micro-video keyframes by similarity measurement. Meanwhile, we complete the tags by detecting the objects, activities and scenes in micro-videos based on appearance features and motion features with the assistance of the datasets, namely, ImageNet, PASCAL VOC, HMDB51, UCF50 and SUN. We also denoise the hashtags by constructing the mapping relationships among hashtags and target tags based on the statistics on NUS-WIDE. The results of tag transfer, complement and denoising are finally linearly combined to generate the tag refinement results of micro-videos. To validate the performance, we construct a dataset with 600 micro-videos from Vine, and manually labelled the micro-videos with target tags. The experimental results show that our approach can obtain good performance in tag refinement of micro-videos by learning from multiple data sources.
机译:微型视频是一种越来越流行的社交媒体形式,其便捷的获取和表达能力引起了人们的广泛关注。但是,由于用户生成的微视频主题标签的分布严重不均衡且质量低下,因此微视频的管理变得充满挑战。在本文中,我们通过学习多个带有手动标记标签的公共数据源,提出了一种新颖的微视频标签细化方法,该方法可以克服直接精炼不精确的标签的困难,并解决了缺少手动标记微视频数据集的问题为了训练。我们通过引用广泛用于对象,活动和场景检测的数据集来定义一组目标标签。在标签优化中,我们首先通过相似性测量将标签从NUS-WIDE中的图像传输到微视频关键帧。同时,我们借助ImageNet,PASCAL VOC,HMDB51,UCF50和SUN等数据集,根据外观特征和运动特征检测微视频中的对象,活动和场景,从而完成标签的制作。我们还基于NUS-WIDE上的统计信息,通过在主题标签和目标标签之间构建映射关系来对主题标签进行去噪。标签传递,补码和去噪的结果最终线性组合,以生成微视频的标签细化结果。为了验证性能,我们使用来自Vine的600个微型视频构建了一个数据集,并使用目标标签手动标记了这些微型视频。实验结果表明,通过从多个数据源中学习,我们的方法可以在微视频的标签细化中获得良好的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号