首页> 外文会议>International Conference on Speech and Computer >A Perceptually Inspired Data Augmentation Method for Noise Robust CNN Acoustic Models
【24h】

A Perceptually Inspired Data Augmentation Method for Noise Robust CNN Acoustic Models

机译:一种感知激发噪声鲁棒CNN声学模型的数据增强方法

获取原文

摘要

Here, we present a data augmentation method that improves the robustness of convolutional neural network-based speech recognizers to additive noise. The proposed technique has its roots in the input dropout method because it discards a subset of the input features. However, instead of doing this in a completely random fashion, we introduce two simple heuristics that select the less reliable components of the spectrum of the speech signal as candidates for dropout. The first heuristic retains spectro-temporal maxima, while the second is based on a crude estimation of spectral dominance. The selected components are always retained, while the dropout step discards or retains the unselected ones in a probabilistic manner. Due to the randomness involved in dropout, the whole process may be interpreted as a data augmentation method that perturbs the data by creating new data instances from the existing ones on the fly. We evaluated the method on the Aurora-4 corpus just using the clean training data set, and we got relative word error rate reductions between 22% and 25%.
机译:在这里,我们提出了一种数据增强方法,其提高了基于卷积神经网络的语音识别器的鲁棒性与加性噪声。所提出的技术在输入丢弃方法中具有其根源,因为它会丢弃输入功能的子集。然而,我们不是以完全随机的方式执行此操作,我们介绍了两个简单的启发式方法,可以选择语音信号频谱的可靠组件作为辍学的候选者。第一启发式保留光谱时间最大值,而第二种启发式是基于光谱优势的粗略估计。始终保留所选择的组件,而辍学步骤以概率的方式丢弃或保留未选择的组件。由于丢失中所涉及的随机性,整个过程可以被解释为数据增强方法,它通过从现有的现有数据实例开始使用来自现有的新数据实例。我们在Aurora-4语料库上评估了Method的方法,只需使用清洁训练数据集,我们得到了22%和25%之间的相对字错误率。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号