首页> 外文会议>International Symposium on Micro-NanoMechatronics and Human Science >Fusion system of vision and hearing sensation using Deep Learning Fusion system of vision and hearing sensation using Deep Learning
【24h】

Fusion system of vision and hearing sensation using Deep Learning Fusion system of vision and hearing sensation using Deep Learning

机译:深度学习的视觉和听觉融合系统深度学习的视觉和听觉融合系统

获取原文

摘要

Recently, sensing technology has been dramatically developed. Along with this, a wide variety of sensors have been used in a single system such as automated driving technology and the robot industry. However, as the number of sensors in a system increases, a fusion method for information obtained from the sensors becomes a problem. When humans recognize information from environment, the information obtained from the five senses is once transmitted and processed in sensory areas such as the visual and auditory areas of the brain. After that, the information processed in the sensory area is transmitted to the association area, and information fusion is performed. Also in robot's sensor fusion system, development of such human sensor fusion system is expected. In this paper, we propose a method to extract feature value using deep learning for each sensor and fusion the feature value. In this system, a system constructed by combining lipreading and speech recognition using visual and auditory information. We aim to realize sensor fusion by extracting feature value and recognizing words using Convolutional Neural Network (CNN) respectively for visual and auditory information and inputting the recognition results to the Neural Network that fusion the recognition results. Recently, sensing technology has been dramatically developed. Along with this, a wide variety of sensors have been used in a single system such as automated driving technology and the robot industry. However, as the number of sensors in a system increases, a fusion method for information obtained from the sensors becomes a problem. When humans recognize information from environment, the information obtained from the five senses is once transmitted and processed in sensory areas such as the visual and auditory areas of the brain. After that, the information processed in the sensory area is transmitted to the association area, and information fusion is performed. Also in robot's sensor fusion system, development of such human sensor fusion system is expected. In this paper, we propose a method to extract feature value using deep learning for each sensor and fusion the feature value. In this system, a system constructed by combining lipreading and speech recognition using visual and auditory information. We aim to realize sensor fusion by extracting feature value and recognizing words using Convolutional Neural Network (CNN) respectively for visual and auditory information and inputting the recognition results to the Neural Network that fusion the recognition results.
机译:最近,传感技术得到了极大的发展。伴随着此,在单个系统中使用了多种传感器,例如自动驾驶技术和机器人行业。但是,随着系统中传感器的数量增加,用于从传感器获得的信息的融合方法成为问题。当人类从环境中识别出信息时,从五种感官中获得的信息便会在诸如大脑的视觉和听觉区域之类的感觉区域中进行传输和处理。之后,在感觉区域中处理的信息被发送到关联区域,并且执行信息融合。同样在机器人的传感器融合系统中,期望开发这种人类传感器融合系统。在本文中,我们提出了一种使用深度学习为每个传感器提取特征值并融合特征值的方法。在该系统中,是通过使用视觉和听觉信息将唇读和语音识别相结合而构建的系统。我们旨在通过分别使用卷积神经网络(CNN)提取视觉识别和听觉信息的特征值并识别单词,并将识别结果输入到融合识别结果的神经网络中,来实现传感器融合。最近,传感技术得到了极大的发展。伴随着此,在单个系统中使用了多种传感器,例如自动驾驶技术和机器人行业。然而,随着系统中传感器的数量增加,用于从传感器获得的信息的融合方法成为问题。当人类从环境中识别出信息时,从五种感官中获得的信息便会在诸如大脑的视觉和听觉区域之类的感觉区域中进行传输和处理。之后,在感觉区域中处理的信息被发送到关联区域,并且执行信息融合。同样在机器人的传感器融合系统中,期望开发这种人类传感器融合系统。在本文中,我们提出了一种使用深度学习为每个传感器提取特征值并融合特征值的方法。在该系统中,是通过使用视觉和听觉信息将唇读和语音识别相结合而构建的系统。我们旨在通过分别使用卷积神经网络(CNN)提取视觉识别和听觉信息的特征值并识别单词,并将识别结果输入到融合识别结果的神经网络中,来实现传感器融合。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号