首页> 外文期刊>Multimedia Tools and Applications >Data generation and representation method for 3D video conferencing using programming by demonstration
【24h】

Data generation and representation method for 3D video conferencing using programming by demonstration

机译:通过演示编程进行3D视频会议的数据生成和表示方法

获取原文
获取原文并翻译 | 示例
           

摘要

Video conferencing provides an environment for multiple users linked on a network to have meetings. Since a large quantity of audio and video data are transferred to multiple users in real time, research into reducing the quantity of data to be transferred has been drawing attention. Such methods extract and transfer only the features of a user from video data and then reconstruct a video conference using virtual humans. The disadvantage of such an approach is that only the positions and features of hands and heads are extracted and reconstructed, whilst the other virtual body parts do not follow the user. In order to enable a virtual human to accurately mimic the entire body of the user in a 3D virtual conference, we examined what features should be extracted to express a user more clearly and how they can be reproduced by a virtual human. This 3D video conferencing estimates the user's pose by comparing predefined images with a photographed user's image and generates a virtual human that takes the estimated pose. However, this requires predefining a diverse set of images for pose estimation and, moreover, it is difficult to define behaviors that can express poses correctly. This paper proposes a framework to automatically generate the pose-images used to estimate a user's pose and the behaviors required to present a user using a virtual human in a 3D video conference. The method for applying this framework to a 3D video conference on the basis of the automatically generated data is also described. In the experiment, the framework proposed in this paper was implemented in a mobile device. The generation process of poses and behaviors of virtual human was verified. Finally, by applying programming by demonstration, we developed a system that can automatically collect the various data necessary for a video conference directly without any prior knowledge of the video conference system.
机译:视频会议为链接到网络上的多个用户举行会议提供了一种环境。由于大量的音频和视频数据被实时传输到多个用户,因此减少传输数据量的研究已引起人们的关注。这样的方法仅从视频数据中提取和传递用户的特征,然后使用虚拟人重建视频会议。这种方法的缺点是仅提取和重建了手和头部的位置和特征,而其他虚拟身体部位却不跟随用户。为了使虚拟人能够在3D虚拟会议中准确模拟用户的整个身体,我们研究了应提取哪些特征以更清楚地表达用户以及虚拟人如何再现这些特征。此3D视频会议通过将预定义图像与拍摄的用户图像进行比较来估计用户的姿势,并生成采用估计姿势的虚拟人。但是,这需要预定义一组多样化的图像以进行姿势估计,此外,很难定义可以正确表达姿势的行为。本文提出了一种框架,该框架可自动生成用于估计用户姿势的姿势图像以及在3D视频会议中使用虚拟人演示用户所需的行为。还描述了基于自动生成的数据将该框架应用于3D视频会议的方法。在实验中,本文提出的框架是在移动设备中实现的。验证了虚拟人的姿势和行为的生成过程。最后,通过演示编程,我们开发了一个系统,该系统可以自动收集视频会议所需的各种数据,而无需任何视频会议系统的先验知识。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号