首页> 外文会议>SPIE Photonics Europe Conference >Conception of a Touchless Human Machine Interaction system for Operating Rooms using Deep Learning
【24h】

Conception of a Touchless Human Machine Interaction system for Operating Rooms using Deep Learning

机译:使用深度学习的用于手术室的非接触式人机交互系统的构想

获取原文

摘要

Touchless Human-Computer Interaction (HMI) is important in sterile environments, especially, in operating rooms (OR). Surgeons need to interact with images from scanners, rayon X, ultrasound images, etc. Problems about contamination may happen if surgeons must touch a keyboard or the mouse. To reduce the contamination and to give the possibility to the surgeon to be more autonomous during the operation, different projects have been developed in the Medic@ team from 2011. In order to recognize the hand and the gestures, two main projects: Gesture Tool Box and K2A; based on the use of the Kinect's device (with a depth camera) have been prototyped. The detection of the hand gesture was done by segmentation and hand descriptors on RGB images, but always with a dependency on the depth camera (Kinect) to the detection of the hand. Additionally, this approach does not give the possibility that the system adapts to a new gesture demanded by the end-user, for example, if a new gesture is demanded, a new algorithm must be programed and tested. Thanks to the evolution of NVDIA cards to reduce time processing algorithms for CNN, the last approach explored was the use of the deep learning algorithms. The Gesture tool box project done was to analyze the hand gesture detection using a CNN (pre-trained in VGG 16) and transfer learning. The results were very promising showing 85% of accuracy for the detection of 10 different gestures form LSF ( French Sign Language) and also it was possible to create a user interface to give autonomy to the end user to add his own gesture and to do the transfer learning automatically. However, we still had some problems about the real time delay (0,8s) recognition and the dependency of the Kinect device. In this article, a new architecture is proposed, in which we want to use standard cameras and to reduce the real time delay of the hand and gesture detection. The state of the art shows the use of a YOLOv2 using Darknet framework as a good option with faster time recognition compared to other CNN. We have implemented YOLOv2 for the detection of the hand and signs with good results in gesture detection and with 0.10 seconds on gesture time recognition in laboratory conditions. Future work will include reducing the errors of our model, recognizing intuitive and standard gestures and doing tests in real conditions.
机译:非接触式人机交互(HMI)在无菌环境中尤其是在手术室(OR)中非常重要。外科医生需要与来自扫描仪的图像,人造丝X,超声图像等进行交互。如果外科医生必须触摸键盘或鼠标,则可能会发生污染问题。为了减少污染并让外科医生在手术过程中更加自主,从2011年开始,Medic @团队开发了不同的项目。为了识别手和手势,有两个主要项目:手势工具箱和K2A;基于使用Kinect的设备(带有深度摄像头)已被原型化。手势的检测是通过在RGB图像上进行分割和手描述符完成的,但始终依赖于深度相机(Kinect)来检测手。另外,这种方法不会使系统适应最终用户所要求的新手势的可能性,例如,如果需要新手势,则必须对新算法进行编程和测试。由于NVDIA卡的发展,以减少CNN的时间处理算法,因此探索的最后一种方法是使用深度学习算法。手势工具箱项目的完成是使用CNN(在VGG 16中进行了预训练)分析手势检测并进行转移学习。结果非常有希望,显示从LSF(法语手语)中检测10种不同手势的准确率达到85%,并且还可以创建一个用户界面,使最终用户可以自主添加自己的手势并进行操作。自动转移学习。但是,我们仍然在实时延迟(0,8s)识别以及Kinect设备的依赖性方面存在一些问题。在本文中,提出了一种新的体系结构,其中我们要使用标准相机并减少手和手势检测的实时延迟。最新技术表明,与其他CNN相比,使用Darknet框架的YOLOv2是一个不错的选择,具有更快的时间识别。我们已经实现了YOLOv2来检测手和手势,在手势检测中具有良好的结果,在实验室条件下的手势时间识别中具有0.10秒的时间。未来的工作将包括减少我们模型的错误,识别直观和标准手势以及在实际条件下进行测试。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号