首页> 外文会议>SIGMORPHON workshop on computational research in phonetics phonology, and morphology >One Model to Pronounce Them All: Multilingual Grapheme-to-Phoneme Conversion With a Transformer Ensemble
【24h】

One Model to Pronounce Them All: Multilingual Grapheme-to-Phoneme Conversion With a Transformer Ensemble

机译:一个模型来全部发音:带有变压器组的多语言音素到音素转换

获取原文

摘要

The task of grapheme-to-phoneme (G2P) conversion is important for both speech recognition and synthesis. Similar to other speech and language processing tasks, in a scenario where only small-sized training data are available, learning G2P models is challenging. We describe a simple approach of exploiting model ensembles, based on multilingual Transformers and self-training, to develop a highly effective G2P solution for 15 languages. Our models are developed as part of our participation in the SIGMORPHON 2020 Shared Task 1 focused at G2P. Our best models achieve 14.99 word error rate (WER) and 3.30 phoneme error rate (PER), a sizeable improvement over the shared task competitive baselines.
机译:字素到音素(G2P)转换的任务对于语音识别和合成都很重要。与其他语音和语言处理任务类似,在只有小型培训数据可用的情况下,学习G2P模型具有挑战性。我们描述了一种基于模型的简单方法,该方法基于多语言的Transformers和自我训练,为15种语言开发了高效的G2P解决方案。我们的模型是作为我们参与针对G2P的SIGMORPHON 2020共享任务1的一部分而开发的。我们的最佳模型可实现14.99字错误率(WER)和3.30音素错误率(PER),比共享任务竞争基准大幅度提高。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号