【24h】

EyeSee: Camera to Caption with Attention Mechanism

机译:眼睛:相机与注意机制的标题

获取原文

摘要

According to the WHO, there are currently around 2.2 billion people who are either visually impaired or blind in the world. Previously, these people had to rely only on classic aids such as the white cane and the guide dog for mobility and magnifiers and screen readers amongst others for reading. The massive use of smartphones has opened many new possibilities for the visually impaired and blind. They can now use their smartphones to help them navigate around cities and other places. In this project it is proposed to have an app for smartphones which automatically tells the blind user the objects around him. However, automatically identifying and describing the content of an image is not such a simple task. It involves tasks from 2 complex fields namely computer vision and natural language processing. The proposed application, EyeSee, takes images from a real-time environment, processes these frame by frame and tells the user what the image represents. The app also annotates the images with text. The app uses Deep Learning, more specifically, Show, Attend and Tell and GRU.
机译:根据谁,目前在左右22亿人在视力下障碍或盲目。此前,这些人不得不依赖于经典助剂,如白色甘蔗和导游,用于移动和放大器以及其他人之间的屏幕阅读器。智能手机的大规模使用已经为视障和盲人开辟了许多新的可能性。他们现在可以使用他们的智能手机来帮助他们绕过城市和其他地方。在这个项目中,建议为智能手机提供一个应用程序,它自动告诉盲目用户周围的对象。但是,自动识别和描述图像的内容并不是如此简单的任务。它涉及来自2个复杂字段的任务,即计算机视觉和自然语言处理。建议的应用程序ISEEPEE,拍摄来自实时环境的图像,通过帧来处理这些帧并告诉用户图像所代表的内容。该应用程序还用文本注释图像。该应用程序使用深度学习,更具体地说,展示,参加和告诉GRU。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号