...
首页> 外文期刊>Multimedia Tools and Applications >Efficient eye typing with 9-direction gaze estimation
【24h】

Efficient eye typing with 9-direction gaze estimation

机译:通过9方向注视估计实现高效的眼睛打字

获取原文
获取原文并翻译 | 示例
           

摘要

Vision based text entry systems aim to help disabled people achieve text communication using eye movement. Most previous methods have employed an existing eye tracker to predict gaze direction and designed an input method based upon that. However, these methods can result in eye tracking quality becoming easily affected by various factors and lengthy amounts of time for calibration. Our paper presents a novel efficient gaze based text input method, which has the advantage of low cost and robustness. Users can type in words by looking at an on-screen keyboard and blinking. Rather than estimate gaze angles directly to track eyes, we introduce a method that divides the human gaze into nine directions. This method can effectively improve the accuracy of making a selection by gaze and blinks. We built a Convolutional Neural Network (CNN) model for 9-direction gaze estimation. On the basis of the 9-direction gaze, we used a nine-key T9 input method which is widely used in candy bar phones. Bar phones were very popular in the world decades ago and have cultivated strong user habits and language models. To train a robust gaze estimator, we created a large-scale dataset with images of eyes sourced from 25 people. According to the results from our experiments, our CNN model is able to accurately estimate different people's gaze under various lighting conditions. In considering disable people's needs, we removed the complex calibration process. The input methods can run in screen mode and portable off-screen mode. Moreover, The datasets used in our experiments are made available to the community to allow further research.
机译:基于视觉的文本输入系统旨在帮助残疾人使用眼球运动实现文本交流。以前的大多数方法都采用了现有的眼动仪来预测视线方向,并以此为基础设计了一种输入法。但是,这些方法可能导致眼动跟踪质量容易受到各种因素的影响以及校准所需的时间较长。本文提出了一种新颖,高效的基于注视的文本输入方法,该方法具有成本低,鲁棒性强的优点。用户可以通过查看屏幕键盘并闪烁来输入单词。我们没有直接估计注视角度来跟踪眼睛,而是引入了一种将人的注视分为九个方向的方法。该方法可以有效地提高注视和眨眼选择的准确性。我们建立了卷积神经网络(CNN)模型用于9方向注视估计。在9方向注视的基础上,我们使用了九键T9输入法,该方法在直板手机中广泛使用。直板电话在几十年前在世界范围内非常流行,并且已经培养了强大的用户习惯和语言模型。为了训练强大的注视估计器,我们创建了一个大规模数据集,其中包含来自25个人的眼睛图像。根据我们的实验结果,我们的CNN模型能够准确估计各种光照条件下不同人的目光。在考虑禁用人们的需求时,我们删除了复杂的校准过程。输入法可以在屏幕模式和便携式离屏模式下运行。此外,我们在实验中使用的数据集已向社区开放,以允许进行进一步的研究。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号