首页> 外文会议>European conference on computer vision >Massively Parallel Video Networks
【24h】

Massively Parallel Video Networks

机译:大规模并行视频网络

获取原文

摘要

We introduce a class of causal video understanding models that aims to improve efficiency of video processing by maximising throughput, minimising latency, and reducing the number of clock cycles. Leveraging operation pipelining and multi-rate clocks, these models perform a minimal amount of computation (e.g. as few as four convolutional layers) for each frame per timestep to produce an output. The models are still very deep, with dozens of such operations being performed but in a pipelined fashion that enables depth-parallel computation. We illustrate the proposed principles by applying them to existing image architectures and analyse their behaviour on two video tasks: action recognition and human keypoint localisation. The results show that a significant degree of parallelism, and implicitly speedup, can be achieved with little loss in performance.
机译:我们引入了一类因果视频理解模型,旨在通过最大化吞吐量,最小化延迟并减少时钟周期数来提高视频处理效率。利用操作流水线和多速率时钟,这些模型为每个时间步的每一帧执行最少的计算量(例如,少至四个卷积层)以产生输出。这些模型仍然非常深入,虽然执行了数十种此类操作,但是它们以能够进行深度并行计算的流水线方式进行。我们通过将它们应用于现有的图像体系结构来说明所提出的原理,并分析它们在两个视频任务上的行为:动作识别和人类关键点定位。结果表明,可以在不损失性能的情况下实现相当程度的并行性和隐式加速。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号