首页> 外文会议>European conference on computer vision >Temporal Convolutional Networks: A Unified Approach to Action Segmentation
【24h】

Temporal Convolutional Networks: A Unified Approach to Action Segmentation

机译:时间卷积网络:统一的动作分割方法

获取原文

摘要

The dominant paradigm for video-based action segmentation is composed of two steps: first, compute low-level features for each frame using Dense Trajectories or a Convolutional Neural Network to encode local spatiotemporal information, and second, input these features into a classifier such as a Recurrent Neural Network (RNN) that captures high-level temporal relationships. While often effective, this decoupling requires specifying two separate models, each with their own complexities, and prevents capturing more nuanced long-range spatiotemporal relationships. We propose a unified approach, as demonstrated by our Temporal Convolutional Network (TCN), that hierarchically captures relationships at low-, intermediate-, and high-level time-scales. Our model achieves superior or competitive performance using video or sensor data on three public action segmentation datasets and can be trained in a fraction of the time it takes to train an RNN.
机译:基于视频的动作分割的主要范例包括两个步骤:首先,使用密集轨迹或卷积神经网络为每个帧计算低级特征,以编码局部时空信息;其次,将这些特征输入到分类器中,例如捕获高级时间关系的递归神经网络(RNN)。这种解耦虽然通常很有效,但需要指定两个单独的模型,每个模型都有各自的复杂性,并防止捕获更细微的远程时空关系。正如我们的时间卷积网络(TCN)所展示的,我们提出了一种统一的方法,该方法可以分层地捕获低,中和高时间尺度上的关系。我们的模型使用三个公共行动细分数据集上的视频或传感器数据来获得卓越或竞争的性能,并且可以在训练RNN的时间的一小部分内进行训练。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号