首页> 外文会议>European conference on computer vision >Temporal Convolutional Networks: A Unified Approach to Action Segmentation
【24h】

Temporal Convolutional Networks: A Unified Approach to Action Segmentation

机译:时间卷积网络:行动分割的统一方法

获取原文

摘要

The dominant paradigm for video-based action segmentation is composed of two steps: first, compute low-level features for each frame using Dense Trajectories or a Convolutional Neural Network to encode local spatiotemporal information, and second, input these features into a classifier such as a Recurrent Neural Network (RNN) that captures high-level temporal relationships. While often effective, this decoupling requires specifying two separate models, each with their own complexities, and prevents capturing more nuanced long-range spatiotemporal relationships. We propose a unified approach, as demonstrated by our Temporal Convolutional Network (TCN), that hierarchically captures relationships at low-, intermediate-, and high-level time-scales. Our model achieves superior or competitive performance using video or sensor data on three public action segmentation datasets and can be trained in a fraction of the time it takes to train an RNN.
机译:基于视频动作分割的主导范例由两个步骤组成:首先,使用致密轨迹或卷积神经网络计算每个帧的低级功能,以编码本地时空信息,然后将这些功能输入到诸如此类的分类器中捕获高级时间关系的经常性神经网络(RNN)。虽然经常有效,但这种去耦需要指定两个单独的模型,每个模型都具有自己的复杂性,并防止捕获更细微的远程时空关系。我们提出了一种统一的方法,如我们的时间卷积网络(TCN)所证明的,该方法在分层上捕获低,中间和高级时间尺度的关系。我们的模型在三个公共动作分割数据集上使用视频或传感器数据实现了卓越或竞争性的性能,并且可以在培训RNN的一小部分中进行培训。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号