首页> 外文会议>IEEE International Conference on Systems, Man, and Cybernetics >Localization of RGB-D Camera Networks by Skeleton-based Viewpoint Invariance Transformation
【24h】

Localization of RGB-D Camera Networks by Skeleton-based Viewpoint Invariance Transformation

机译:基于骨架的Viewpoint Invarification转换,RGB-D相机网络的本地化

获取原文

摘要

Combining depth information and color image, RGB-D cameras provide ready detection of humans and the associated 3D skeleton joints data, facilitating if not revolutionizing conventional image centric researches in, among others, computer vision, surveillance, and human activity analysis. Applicability of a D-RBG camera, however, is restricted by its limited range of frustum of depth in the range of 0.8 to 4 meters. Although a RGB-D camera network, constructed by deployment of several RGB-D cameras at various locations, could extend the range of coverage, it requires precise localization of the camera network: relative location and orientation of neighboring cameras. By introducing a skeleton-based viewpoint invariant transformation (SVIT) that derives relative location and orientation of a detected human's upper torso to a RGB-D camera, this paper presents a reliable automatic localization technique without the need for additional instrument or human intervention. By respectively applying SVIT to two neighboring RGB-D cameras on a commonly observed skeleton, the respective relative position and orientation of the detected human's skeleton to these two respective cameras can be obtained before being combined to yield the relative position and orientation of these two cameras, thus solving the localization problem. Experiments have been conducted where two Kinects are situated with bearing differences of about 45 degrees and 90 degrees when coverage extended by up to 70% with the installment of an additional Kinect. The same localization technique can be applied repeatedly to a larger number of RGB-D cameras, thus extending the applicability of RGB-D cameras to camera networks in making human behavior analysis and context-aware service in a lager surveillance area.
机译:组合深度信息和彩色图像,RGB-D相机提供了人类和相关的3D骨架关节数据的准备检测,便于如果没有彻底改变传统的图像中以常规图像,其中包括计算机视觉,监测和人类活动分析。然而,D-RBG摄像机的适用性受到其有限的深度范围的限制在0.8至4米的范围内。尽管通过在各个位置部署多个RGB-D相机构建的RGB-D相机网络可以扩展覆盖范围,但它需要相机网络的精确定位:相邻摄像机的相对位置和方向。通过引入基于骨架的ViewPoint不变变换(SVIT),该转换导出检测到的人类上躯干的相对位置和定向到RGB-D相机,本文介绍了可靠的自动定位技术,无需额外的仪器或人为干预。通过分别在常见的骨架上施加SVIT到两个相邻的RGB-D相机,在组合之前可以获得检测到的人骨架对这两个相应的相机的相应位置和取向,以产生这两个相机的相对位置和取向,从而解决本地化问题。已经进行了实验,其中两个Kinects位于约45度和90度的轴承差,当覆盖范围高达70%时,分为额外的kinect延伸到达70%。相同的定位技术可以重复地应用于更大数量的RGB-D相机,从而将RGB-D相机的适用性扩展到在Lager监控区域中的人类行为分析和背景感知服务方面将RGB-D摄像机对相机网络进行了应用。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号