首页> 外文会议>International Conference on 3D Immersion >VST3D-Net:Video-Based Spatio-Temporal Network for 3D Shape Reconstruction from a Video
【24h】

VST3D-Net:Video-Based Spatio-Temporal Network for 3D Shape Reconstruction from a Video

机译:VST3D-Net:基于视频的时空网络,用于视频3D形状重建

获取原文

摘要

In this paper, we propose the Video-based Spatio-Temporal 3D Network (VST3D-Net), which is a novel learning approach of viewpoint-invariant 3D shape reconstruction from monocular video. In our VST3D-Net, a spatial feature extraction subnetwork is designed to encode the local and global spatial relationships of the object in the image. The extracted latent spatial features have implicitly embedded both shape and pose information. Although a single view can also be used to recover a 3D shape, more rich shape information of the dynamic object can be explored and leveraged from video frames. To generate the viewpoint-free 3D shape, we design a temporal correlation feature extractor. It handles the temporal consistency of the shape and pose of the moving object simultaneously. Therefore, both the canonical 3D shape and the corresponding pose at different frame are recovered by the network. We validate our approach on the ShapeNet-based video dataset and ApolloCar3D dataset. The experimental results show the proposed VST3D-Net can outperform the state-of-the-art approaches both in accuracy and efficiency.
机译:在本文中,我们提出了一种基于视频的时空3D网络(VST3D-Net),这是一种从单目一象视频的观点不变三维形状重建的新学习方法。在VST3D-Net中,空间特征提取子网旨在编码图像中对象的本地和全局空间关系。提取的潜在空间特征隐含地嵌入了形状和姿势信息。虽然单个视图也可用于恢复3D形状,但是可以探索动态对象的更丰富的形状信息和从视频帧中杠杆。要生成无视点3D形状,我们设计了一个时间相关特征提取器。它同时处理移动物体的形状和姿势的时间一致性。因此,通过网络恢复不同帧的规范3D形状和相应的姿势。我们在基于ShapEnet​​的视频数据集和Apollocar3D数据集中验证了我们的方法。实验结果表明,所提出的VST3D-Net可以在准确性和效率方面优于最先进的方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号