首页> 外文会议>IEEE Conference on Computer Vision and Pattern Recognition >Temporal Convolutional Networks for Action Segmentation and Detection
【24h】

Temporal Convolutional Networks for Action Segmentation and Detection

机译:动作分割和检测的时间卷积网络

获取原文

摘要

The ability to identify and temporally segment finegrained human actions throughout a video is crucial for robotics, surveillance, education, and beyond. Typical approaches decouple this problem by first extracting local spatiotemporal features from video frames and then feeding them into a temporal classifier that captures high-level temporal patterns. We describe a class of temporal models, which we call Temporal Convolutional Networks (TCNs), that use a hierarchy of temporal convolutions to perform fine-grained action segmentation or detection. Our Encoder-Decoder TCN uses pooling and upsampling to efficiently capture long-range temporal patterns whereas our Dilated TCN uses dilated convolutions. We show that TCNs are capable of capturing action compositions, segment durations, and long-range dependencies, and are over a magnitude faster to train than competing LSTM-based Recurrent Neural Networks. We apply these models to three challenging fine-grained datasets and show large improvements over the state of the art.
机译:在整个视频中识别和暂时分段的能力将在视频中,对机器人,监测,教育和超越来说至关重要。典型的方法通过首先从视频帧提取本地时空特征然后将它们馈送到捕获高级时间模式的时间分类器中来解耦这个问题。我们描述了一类时间模型,我们呼叫时间卷积网络(TCN),该网络使用时间卷积的层次结构来执行细粒度的动作分段或检测。我们的编码器解码器TCN使用池和上采样来有效地捕获远程时间模式,而我们扩张的TCN使用扩张的卷曲。我们表明TCN能够捕获动作组合物,段持续时间和远程依赖性,并且比竞争基于LSTM的经常性神经网络更快地捕获速度。我们将这些模型应用于三个具有挑战性的细粒度的数据集,并对本领域的巨大改进显示出大量的改进。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号