首页> 外文会议>IEEE/CVF Conference on Computer Vision and Pattern Recognition >Deep Spatio-Temporal Random Fields for Efficient Video Segmentation
【24h】

Deep Spatio-Temporal Random Fields for Efficient Video Segmentation

机译:深度时空随机场,用于有效的视频分割

获取原文

摘要

In this work we introduce a time- and memory-efficient method for structured prediction that couples neuron decisions across both space at time. We show that we are able to perform exact and efficient inference on a densely-connected spatio-temporal graph by capitalizing on recent advances on deep Gaussian Conditional Random Fields (GCRFs). Our method, called VideoGCRF is (a) efficient, (b) has a unique global minimum, and (c) can be trained end-to-end alongside contemporary deep networks for video understanding. We experiment with multiple connectivity patterns in the temporal domain, and present empirical improvements over strong baselines on the tasks of both semantic and instance segmentation of videos. Our implementation is based on the Caffe2 framework and will be available at https://github.com/siddharthachandra/gcrf-v3.0.
机译:在这项工作中,我们为结构化预测引入了一种时间和内存效率高的方法,该方法可同时在两个空间上耦合神经元决策。我们表明,我们能够利用深高斯条件随机场(GCRF)的最新进展,对密集连接的时空图执行精确而有效的推断。我们的方法VideoGCRF(a)高效,(b)具有唯一的全局最小值,并且(c)可以与当代深度网络一起进行端到端培训,以进行视频理解。我们在时域中使用多种连接模式进行了实验,并针对视频的语义和实例分割任务在强基准上提出了经验改进。我们的实现基于Caffe2框架,并将在https://github.com/siddharthachandra/gcrf-v3.0上提供。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号