首页> 外文会议>IEEE Conference on Virtual Reality and 3D User Interfaces >LiveDeep: Online Viewport Prediction for Live Virtual Reality Streaming Using Lifelong Deep Learning
【24h】

LiveDeep: Online Viewport Prediction for Live Virtual Reality Streaming Using Lifelong Deep Learning

机译:LiveDeep:使用终身深度学习进行实时虚拟现实流传输的在线视口预测

获取原文

摘要

Live virtual reality (VR) streaming has become a popular and trending video application in the consumer market providing users with 360-degree, immersive viewing experiences. To provide premium quality of experience, VR streaming faces unique challenges due to the significantly increased bandwidth consumption. To address the bandwidth challenge, VR video viewport prediction has been proposed as a viable solution, which predicts and streams only the user’s viewport of interest with high quality to the VR device. However, most of the existing viewport prediction approaches target only the video-on-demand (VOD) use cases, requiring offline processing of the historical video and/or user data that are not available in the live streaming scenario. In this work, we develop a novel viewport prediction approach for live VR streaming, which only requires video content and user data in the current viewing session. To address the challenges of insufficient training data and real-time processing, we propose a live VR-specific deep learning mechanism, namely LiveDeep, to create the online viewport prediction model and conduct real-time inference. LiveDeep employs a hybrid approach to address the unique challenges in live VR streaming, involving (1) an alternate online data collection, labeling, training, and inference schedule with controlled feedback loop to accommodate for the sparse training data; and (2) a mixture of hybrid neural network models to accommodate for the inaccuracy caused by a single model. We evaluate LiveDeep using 48 users and 14 VR videos of various types obtained from a public VR user head movement dataset. The results indicate around 90% prediction accuracy, around 40% bandwidth savings, and premium processing time, which meets the bandwidth and real-time requirements of live VR streaming.
机译:实时虚拟现实(VR)流已成为消费市场中流行的趋势视频应用程序,可为用户提供360度沉浸式观看体验。为了提供优质的体验,VR流由于带宽消耗的显着增加而面临着独特的挑战。为了解决带宽挑战,已经提出了VR视频视口预测作为可行的解决方案,该方法可以仅将用户感兴趣的视口高质量地预测并流式传输到VR设备。但是,大多数现有的视口预测方法仅针对视频点播(VOD)用例,要求离线处理在实时流传输场景中不可用的历史视频和/或用户数据。在这项工作中,我们为实时VR流开发了一种新颖的视口预测方法,该方法仅在当前观看会话中需要视频内容和用户数据。为了解决训练数据不足和实时处理的挑战,我们提出了一种针对VR的实时深度学习机制,即LiveDeep,以创建在线视口预测模型并进行实时推理。 LiveDeep采用一种混合方法来解决实时VR流中的独特挑战,其中包括:(1)备用在线数据收集,标记,训练和推理进度表,并具有受控的反馈回路,以适应稀疏的训练数据; (2)混合神经网络模型的混合,以适应由单个模型引起的不准确性。我们使用48个用户和14个从公共VR用户头部运动数据集中获得的各种类型的VR视频评估LiveDeep。结果表明,大约90%的预测精度,大约40%的带宽节省以及超长的处理时间,可以满足实时VR流的带宽和实时性要求。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号