首页> 外文期刊>Audio, Speech, and Language Processing, IEEE/ACM Transactions on >Collaborative Joint Training With Multitask Recurrent Model for Speech and Speaker Recognition
【24h】

Collaborative Joint Training With Multitask Recurrent Model for Speech and Speaker Recognition

机译:多任务递归模型的协同联合训练,用于语音和说话者识别

获取原文
获取原文并翻译 | 示例

摘要

Automatic speech and speaker recognition are traditionally treated as two independent tasks and are studied separately. The human brain in contrast deciphers the linguistic content, and the speaker traits from the speech in a collaborative manner. This key observation motivates the work presented in this paper. A collaborative joint training approach based on multitask recurrent neural network models is proposed, where the output of one task is backpropagated to the other tasks. This is a general framework for learning collaborative tasks and fits well with the goal of joint learning of automatic speech and speaker recognition. Through a comprehensive study, it is shown that the multitask recurrent neural net models deliver improved performance on both automatic speech and speaker recognition tasks as compared to single-task systems. The strength of such multitask collaborative learning is analyzed, and the impact of various training configurations is investigated.
机译:传统上,自动语音和说话者识别被视为两个独立的任务,需要分别进行研究。相比之下,人脑则以协作的方式从语音中解读语言内容和说话人特征。这项关键的观察激励了本文提出的工作。提出了一种基于多任务递归神经网络模型的协同联合训练方法,其中一个任务的输出反向传播到其他任务。这是学习协作任务的通用框架,非常适合联合学习自动语音和说话者识别的目标。通过全面的研究表明,与单任务系统相比,多任务循环神经网络模型在自动语音和说话者识别任务上均提供了改进的性能。分析了这种多任务协作学习的优势,并研究了各种培训配置的影响。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号