首页> 外文学位 >Ensemble feature selection for multi-stream automatic speech recognition.
【24h】

Ensemble feature selection for multi-stream automatic speech recognition.

机译:集成特征选择,用于多流自动语音识别。

获取原文
获取原文并翻译 | 示例

摘要

Multi-stream automatic speech recognition (ASR) systems consisting of an ensemble of classifiers working together, each with its own feature vector, are popular in the research literature. Published work on feature selection for such systems has dealt with indivisible blocks of features. I break from this tradition by investigating feature selection at the level of individual features. I use the OGI ISOLET and Numbers speech corpora, including noisy versions I created using a variety of noises and signal-to-noise ratios. I have made these noisy versions available for use by other researchers, along with my ASR and feature selection scripts.;I start with the random subspace method of ensemble feature selection, in which each feature vector is simply chosen randomly from the feature pool. Using ISOLET, I obtain performance improvements over baseline in almost every case where there is a statistically significant performance difference, but there are many cases with no such difference.;I then try hill-climbing, a wrapper approach that changes a single feature at a time when the change improves a performance score. With ISOLET, hill-climbing gives performance improvements in most cases for noisy data, but no improvement for clean data. I then move to Numbers, for which much more data is available to guide hill-climbing. When using either the clean or noisy Numbers data, hill-climbing gives performance improvements over multi-stream baselines in almost all cases, although it does not improve over the best single-stream baseline. For noisy data, these performance improvements are present even for noise types that were not seen during the hill-climbing process. In mismatched condition tests involving mismatch between clean and noisy data, hill-climbing outperforms all baselines when Opitz's scoring formula is used. I find that this scoring formula, which blends single-classifier accuracy and ensemble diversity, works better for me than ensemble accuracy as a performance score for guiding hill-climbing.
机译:在研究文献中,多流自动语音识别(ASR)系统由一起工作的一组分类器组成,每个分类器都有自己的特征向量。有关此类系统的特征选择的已发布工作涉及不可分割的特征块。我通过在单个要素的层次上研究要素选择来打破这种传统。我使用OGI ISOLET和Numbers语音语料库,包括我使用各种噪声和信噪比创建的嘈杂版本。我将这些嘈杂的版本以及我的ASR和特征选择脚本供其他研究人员使用。;我从整体特征选择的随机子空间方法开始,在该方法中,简单地从特征池中随机选择每个特征向量。使用ISOLET,几乎在每种情况下在统计上都有显着的性能差异,但在许多情况下却没有这种差异,我可以获得比基线更高的性能;然后我尝试爬山(一种爬山的方法),该方法可以在一个更改可改善效果得分的时间。使用ISOLET,在大多数情况下,爬山可以改善嘈杂数据的性能,但对于干净数据则无法提高性能。然后,我转到Numbers,那里有更多数据可用来指导爬山。使用“干净”或“嘈杂的数字”数据时,爬山在几乎所有情况下都比多流基准线性能有所提高,尽管并没有比最佳单流基准线有所改善。对于嘈杂的数据,即使对于在爬山过程中未发现的噪声类型,也存在这些性能改进。在涉及干净数据与嘈杂数据之间不匹配的不匹配条件测试中,使用Opitz评分公式时,爬坡性能优于所有基线。我发现,将单一分类器准确性和整体多样性相结合的评分公式对我来说比整体准确性作为指导爬山的性能得分更好。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号