首页> 外文会议>IEEE Conference on Computer Vision and Pattern Recognition >Temporal Convolutional Networks for Action Segmentation and Detection
【24h】

Temporal Convolutional Networks for Action Segmentation and Detection

机译:时间卷积网络用于动作分割和检测

获取原文

摘要

The ability to identify and temporally segment fine-grained human actions throughout a video is crucial for robotics, surveillance, education, and beyond. Typical approaches decouple this problem by first extracting local spatiotemporal features from video frames and then feeding them into a temporal classifier that captures high-level temporal patterns. We describe a class of temporal models, which we call Temporal Convolutional Networks (TCNs), that use a hierarchy of temporal convolutions to perform fine-grained action segmentation or detection. Our Encoder-Decoder TCN uses pooling and upsampling to efficiently capture long-range temporal patterns whereas our Dilated TCN uses dilated convolutions. We show that TCNs are capable of capturing action compositions, segment durations, and long-range dependencies, and are over a magnitude faster to train than competing LSTM-based Recurrent Neural Networks. We apply these models to three challenging fine-grained datasets and show large improvements over the state of the art.
机译:识别和临时分割整个视频中的细粒度人类动作的能力对于机器人技术,监视,教育以及其他方面至关重要。典型的方法是通过首先从视频帧中提取局部时空特征,然后将它们馈入捕获高级时域模式的时域分类器中来解决此问题。我们描述了一类时间模型,我们称其为时间卷积网络(TCN),它使用时间卷积的层次结构执行细粒度的动作分割或检测。我们的编码器-解码器TCN使用池化和上采样来有效捕获远程时间模式,而我们的Dilated TCN使用膨胀卷积。我们表明,TCN能够捕获动作组成,片段持续时间和远程依存关系,并且比基于LSTM的递归神经网络的竞争速度要快得多。我们将这些模型应用于三个具有挑战性的细粒度数据集,并显示出对现有技术的巨大改进。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号