【24h】

AirScript - Creating Documents in Air

机译:Airscript - 在空中创建文件

获取原文
获取外文期刊封面目录资料

摘要

This paper presents a novel approach, called AirScript, for creating, recognizing and visualizing documents in air. We present a novel algorithm, called 2-DifViz, that converts the hand movements in air (captured by a Myo-armband worn by a user) into a sequence of x, y coordinates on a 2D Cartesian plane, and visualizes them on a canvas. Existing sensor-based approaches either do not provide visual feedback or represent the recognized characters using prefixed templates. In contrast, AirScript stands out by giving freedom of movement to the user, as well as by providing a real-time visual feedback of the written characters, making the interaction natural. AirScript provides a recognition module to predict the content of the document created in air. To do so, we present a novel approach based on deep learning, which uses the sensor data and the visualizations created by 2-DifViz. The recognition module consists of a Convolutional Neural Network (CNN) and two Gated Recurrent Unit (GRU) Networks. The output from these three networks is fused to get the final prediction about the characters written in air. AirScript can be used in highly sophisticated environments like a smart classroom, a smart factory or a smart laboratory, where it would enable people to annotate pieces of texts wherever they want without any reference surface. We have evaluated AirScript against various well-known learning models (HMM, KNN, SVM, etc.) on the data of 12 participants. Evaluation results show that the recognition module of AirScript largely outperforms all of these models by achieving an accuracy of 91.7% in a person independent evaluation and a 96.7% accuracy in a person dependent evaluation.
机译:本文提出了一种新颖的方法,称为AirScript,用于在空中创造,识别和可视化文件。我们提出了一种名为2-DIFVIZ的新型算法,它将空气中的手动移动(由用户佩戴的Myo-Armband)转换为2D Cartesian平面上的X,Y坐标,并在画布上可视化它们。现有的基于传感器的方法不提供视觉反馈或使用前缀模板表示识别的字符。相比之下,AirScript通过向用户提供行动自由,以及提供书面字符的实时视觉反馈,使得互动自然。 AirScript提供识别模块,以预测在空中创建的文档的内容。为此,我们提出了一种基于深度学习的新方法,它使用了传感器数据和由2-difviz创建的可视化。识别模块由卷积神经网络(CNN)和两个门控复发单元(GRU)网络组成。这三个网络的输出融合以获得关于在空气中写的字符的最终预测。 Airscript可以用于高度复杂的环境,如智能课堂,智能工厂或智能实验室,在那里它将让人们在没有任何参考表面的情况下在没有任何参考表面的情况下向文本注释作品。我们对12名参与者的数据进行了评估了针对各种着名的学习模型(HMM,KNN,SVM等)的AIRScript。评价结果表明,录音识别模块通过在独立评估中实现91.7%的准确性而大大优于所有这些模型,并且在依赖于人员的评估中的准确性为96.7%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号