首页> 外文会议>European conference on computer vision >End-to-End Joint Semantic Segmentation of Actors and Actions in Video
【24h】

End-to-End Joint Semantic Segmentation of Actors and Actions in Video

机译:视频中演员和动作的端到端联合语义分割

获取原文

摘要

Traditional video understanding tasks include human action recognition and actor/object semantic segmentation. However, the combined task of providing semantic segmentation for different actor classes simultaneously with their action class remains a challenging but necessary task for many applications. In this work, we propose a new end-to-end architecture for tackling this task in videos. Our model effectively leverages multiple input modalities, contextual information, and multitask learning in the video to directly output semantic segmentations in a single unified framework. We train and benchmark our model on the Actor-Action Dataset (A2D) for joint actor-action semantic segmentation, and demonstrate state-of-the-art performance for both segmentation and detection. We also perform experiments verifying our approach improves performance for zero-shot recognition, indicating generalizabil-ity of our jointly learned feature space.
机译:传统的视频理解任务包括人类动作识别和演员/对象语义分割。但是,同时为不同的actor类及其动作类提供语义分割的组合任务对于许多应用程序来说仍然是一项具有挑战性但必不可少的任务。在这项工作中,我们提出了一种新的端到端架构来解决视频中的这一任务。我们的模型有效地利用了视频中的多种输入方式,上下文信息和多任务学习,以在单个统一框架中直接输出语义细分。我们在Actor-Action数据集(A2D)上对模型进行了训练和基准测试,以实现联合actor-action语义分割,并展示了分割和检测方面的最新性能。我们还进行了实验,验证了我们的方法提高了零镜头识别的性能,表明了我们共同学习的特征空间的普遍性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号