...
首页> 外文期刊>Journal of neural engineering >Brain2Char: a deep architecture for decoding text from brain recordings
【24h】

Brain2Char: a deep architecture for decoding text from brain recordings

机译:Brain2char:用于从脑录制中解码文本的深度架构

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Decoding language representations directly from the brain can enable new brain-computer interfaces (BCIs) for high bandwidth human-human and human-machine communication. Clinically, such technologies can restore communication in people with neurological conditions affecting their ability to speak. Approach. In this study, we propose a novel deep network architecture Brain2Char, for directly decoding text (specifically character sequences) from direct brain recordings (called electrocorticography, ECoG). Brain2Char framework combines state-of-the-art deep learning modules-3D Inception layers for multiband spatiotemporal feature extraction from neural data and bidirectional recurrent layers, dilated convolution layers followed by language model weighted beam search to decode character sequences, and optimizing a connectionist temporal classification loss. Additionally, given the highly non-linear transformations that underlie the conversion of cortical function to character sequences, we perform regularizations on the network’s latent representations motivated by insights into cortical encoding of speech production and artifactual aspects specific to ECoG data acquisition. To do this, we impose auxiliary losses on latent representations for articulatory movements, speech acoustics and session specific non-linearities. Main results. In three (out of four) participants reported here, Brain2Char achieves 10.6%, 8.5%, and 7.0% word error rates respectively on vocabulary sizes ranging from 1200 to 1900 words. Significance. These results establish a new end-to-end approach on decoding text from brain signals and demonstrate the potential of Brain2Char as a high-performance communication BCI.
机译:直接从大脑解码语言表示可以为高带宽人员和人机通信启用新的脑电脑界面(BCI)。临床上,这种技术可以恢复有影响其讲话能力的神经系统的沟通。方法。在这项研究中,我们提出了一种新颖的深度网络架构Brain2char,用于直接解码来自直接脑录制的文本(特别是字符序列)(称为电加线,ECOG)。 Brain2Char框架结合了最先进的深度学习模块-3D inception层 - 从神经数据和双向复发层提取多频带时空特征,扩张卷积层,后跟语言模型加权光束搜索解码字符序列,并优化连接员时间分类损失。此外,考虑到提高皮质函数对字符序列的高度线性变换,我们对网络的潜在表示,在网络的潜在表示上进行了定期提明,进入语音编码的语音编码和特定于ECOG数据采集的艺术方面。为此,我们施加辅助损失对明晰的表现,言论动作,语音声学和会议特定的非线性。主要结果。在此报告的三个(四分之一)的参与者中,Brain2char分别在1200到1900字的词汇量范围内实现10.6%,8.5%和7.0%的错误速率。意义。这些结果在从脑信号解码文本上建立了新的端到端方法,并证明了Brain2Char的潜力作为高性能通信BCI。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号