首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >Generating Human Readable Transcript for Automatic Speech Recognition with Pre-Trained Language Model
【24h】

Generating Human Readable Transcript for Automatic Speech Recognition with Pre-Trained Language Model

机译:用预先接受训练的语言模型生成人类可读成绩单,用于自动语音识别

获取原文

摘要

Modern Automatic Speech Recognition (ASR) systems can achieve high performance in terms of recognition accuracy. However, a perfectly accurate transcript still can be challenging to read due to disfluency, filter words, and other errata common in spoken communication. Many downstream tasks and human readers rely on the output of the ASR system; therefore, errors introduced by the speaker and ASR system alike will be propagated to the next task in the pipeline. In this work, we propose an ASR post-processing model that aims to transform the incorrect and noisy ASR output into a readable text for humans and downstream tasks. We leverage the Metadata Extraction (MDE) corpus to construct a task-specific dataset for our study. Since the dataset is small, we propose a novel data augmentation method and use a two-stage training strategy to fine-tune the RoBERTa pre-trained model. On the constructed test set, our model outperforms a production two-step pipeline-based post-processing method by a large margin of 13.26 on readability-aware WER (RA-WER) and 17.53 on BLEU metrics. Human evaluation also demonstrates that our method can generate more human-readable transcripts than the baseline method.
机译:现代自动语音识别(ASR)系统可以在识别准确性方面实现高性能。然而,由于在口语通信中常见的传出,滤波器和其他勘误表,完美准确的成绩单仍然可能具有挑战性。许多下游任务和人类读者依赖于ASR系统的输出;因此,由扬声器和ASR系统引入的错误将传播到管道中的下一个任务。在这项工作中,我们提出了一个ASR后处理模型,该模型旨在将输出的不正确和嘈杂的ASR转换为人类和下游任务的可读文本。我们利用元数据提取(MDE)语料库来构建专用的数据集进行学习。由于数据集很小,我们提出了一种新颖的数据增强方法,并使用两级训练策略来微调罗伯塔预训练模型。在构造的测试集上,我们的模型优于一个在Bleu指标上的可读性感知WER(RA-WER)和17.53上的大幅度为13.26的生产两步管道的后处理方法。人类评估还证明我们的方法可以产生比基线方法更具人性可读的转录物。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号