首页> 外文会议>IEEE Automatic Speech Recognition and Understanding Workshop >Knowledge Distillation from Bert in Pre-Training and Fine-Tuning for Polyphone Disambiguation
【24h】

Knowledge Distillation from Bert in Pre-Training and Fine-Tuning for Polyphone Disambiguation

机译:Bert的语音预训练和微调中的知识提炼以消除歧义词

获取原文

摘要

Polyphone disambiguation aims to select the correct pronunciation for a polyphonic word from several candidates, which is important for text-to-speech synthesis. Since the pronunciation of a polyphonic word is usually decided by its context, polyphone disambiguation can be regarded as a language understanding task. Inspired by the success of BERT for language understanding, we propose to leverage pre-trained BERT models for polyphone disambiguation. However, BERT models are usually too heavy to be served online, in terms of both memory cost and inference speed. In this work, we focus on efficient model for polyphone disambiguation and propose a two-stage knowledge distillation method that transfers the knowledge from a heavy BERT model in both pre-training and fine-tuning stages to a lightweight BERT model, in order to reduce online serving cost. Experiments on Chinese and English polyphone disambiguation datasets demonstrate that our method reduces model parameters by a factor of 5 and improves inference speed by 7 times, while nearly matches the classification accuracy (95.4% on Chinese and 98.1% on English) to the original BERT model.
机译:复音歧义消除的目的是从多个候选者中为复音词选择正确的发音,这对于文本到语音的合成很重要。由于复音词的发音通常取决于其上下文,因此,将复音词消除歧义可以看作是一种语言理解任务。受BERT在语言理解方面的成功启发,我们建议利用经过预训练的BERT模型来消除多音素歧义。但是,就内存成本和推理速度而言,BERT模型通常太重而无法在线提供。在这项工作中,我们将重点放在用于消除多音素歧义的有效模型上,并提出一种两阶段的知识提炼方法,该方法将知识从预训练和微调阶段的重型BERT模型转移到轻型BERT模型,以减少在线服务费用。在中英文多音素歧义消除数据集上的实验表明,我们的方法将模型参数减少了5倍,推理速度提高了7倍,而分类准确率(中文为95.4%,英文为98.1%)几乎与原始BERT模型相匹配。 。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号