首页> 外文会议>International Conference on speech and computer >A Perceptually Inspired Data Augmentation Method for Noise Robust CNN Acoustic Models
【24h】

A Perceptually Inspired Data Augmentation Method for Noise Robust CNN Acoustic Models

机译:噪声鲁棒的CNN声学模型的感知启发数据增强方法

获取原文

摘要

Here, we present a data augmentation method that improves the robustness of convolutional neural network-based speech recognizers to additive noise. The proposed technique has its roots in the input dropout method because it discards a subset of the input features. However, instead of doing this in a completely random fashion, we introduce two simple heuristics that select the less reliable components of the spectrum of the speech signal as candidates for dropout. The first heuristic retains spectro-temporal maxima, while the second is based on a crude estimation of spectral dominance. The selected components are always retained, while the dropout step discards or retains the unselected ones in a probabilistic manner. Due to the randomness involved in dropout, the whole process may be interpreted as a data augmentation method that perturbs the data by creating new data instances from the existing ones on the fly. We evaluated the method on the Aurora-4 corpus just using the clean training data set, and we got relative word error rate reductions between 22% and 25%.
机译:在这里,我们提出了一种数据增强方法,该方法可提高基于卷积神经网络的语音识别器对加性噪声的鲁棒性。所提出的技术源于输入丢弃方法,因为它丢弃了输入特征的子集。但是,不是以完全随机的方式执行此操作,而是引入了两种简单的启发式方法,它们选择语音信号频谱中较不可靠的分量作为丢失的候选者。第一种启发式方法保留了频谱时间最大值,而第二种启发式方法则基于对频谱优势的粗略估计。选定的组件始终保留,而退出步骤以概率方式丢弃或保留未选定的组件。由于辍学涉及的随机性,整个过程可以解释为一种数据扩充方法,该方法通过动态地从现有实例中创建新的数据实例来扰乱数据。我们仅使用干净的训练数据集就Aurora-4语料库评估了该方法,并且相对单词错误率降低了22%至25%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号