首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >Exploring the use of Common Label Set to Improve Speech Recognition of Low Resource Indian Languages
【24h】

Exploring the use of Common Label Set to Improve Speech Recognition of Low Resource Indian Languages

机译:探索共同标签集的使用,提高低资源印度语言的语音识别

获取原文

摘要

In many Indian languages, written characters are organized on sound phonetic principles, and the ordering of characters is the same across many of them. However, while training conventional end-to-end (E2E) Multilingual speech recognition systems, we treat characters or target subword units from different languages as separate entities. Since the visual rendering of these characters is different, in this paper, we explore the benefits of representing such similar target subword units (e.g., Byte Pair Encoded(BPE) units) through a Common Label Set (CLS). The CLS can be very easily created using automatic methods since the ordering of characters is the same in many Indian Languages. E2E models are trained using a transformer-based encoder-decoder architecture. During testing, given the Mel-filterbank features as input, the system outputs a sequence of BPE units in CLS representation. Depending on the language, we then map the recognized CLS units back to the language-specific grapheme representation. Results show that models trained using CLS improve over monolingual baseline and a multilingual framework with separate symbols for each language. Similar experiments on a subset of the Voxforge dataset also confirm the benefits of CLS. An extension of this idea is to decode an unseen language (Zero-resource) using CLS trained model.
机译:在许多印度语言中,书面字符是在声音语音原理上组织的,并且在其中许多人中的字符排序是相同的。然而,在培训常规端到端(E2E)的多语言语音识别系统的同时,我们将不同语言的字符或目标子字单元视为单独的实体。由于这些字符的视觉渲染是不同的,因此在本文中,我们探讨了代表这种类似目标子字单元(例如,字节对编码(BPE)单元)的好处通过公共标签集(CLS)。可以使用自动方法非常容易地创建CLS,因为字符的排序是许多印度语言的顺序。 E2E模型使用基于变压器的编码器解码器架构进行培训。在测试期间,给出MEL-FilterBank特征作为输入,系统在CLS表示中输出一系列BPE单元。根据语言,我们将识别的CLS单元映射到特定于语言的图形表示。结果表明,使用CLS培训的型号通过针对单晶体基线和多语言框架进行培训,以及每个语言的单独符号。在Voxforge数据集的子集上的类似实验也证实了CLS的好处。此想法的扩展是使用CLS培训的模型解码未经语言(零资源)。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号