首页> 外文期刊>Telecommunication systems: Modeling, Analysis, Design and Management >Single-ended parametric voicing-aware models for live assessment of packetized VoIP conversations
【24h】

Single-ended parametric voicing-aware models for live assessment of packetized VoIP conversations

机译:单端参量语音感知模型,用于实时评估打包的VoIP对话

获取原文
获取原文并翻译 | 示例
           

摘要

The perceptual quality of VoIP conversations depends tightly on the pattern of packet losses, i.e., the distribution and duration of packet loss runs. The wider (resp. smaller) the inter-loss gap (resp. loss gap) duration, the lower is the quality degradation. Moreover, a set of speech sequences impaired using an identical packet loss pattern results in a different degree of perceptual quality degradation because dropped voice packets have unequal impact on the perceived quality. Therefore, we consider the voicing feature of speech wave included in lost packets in addition to packet loss pattern to estimate speech quality scores. We distinguish between voiced, unvoiced, and silence packets. This enables to achieve better correlation and accuracy between human-based subjective and machine-calculated objective scores. This paper proposes novel no-reference parametric speech quality estimate models which account for the voicing feature of signal wave included in missing packets. Precisely, we develop separate speech quality estimate models, which capture the perceptual effect of removed voiced or unvoiced packets, using elaborated simple and multiple regression analyses. A new speech quality estimate model, which mixes voiced and unvoiced quality scores to compute the overall speech quality score at the end of an assessment interval, is developed following a rigorous multiple linear regression analysis. The input parameters of proposed voicing-aware speech quality estimate models, namely Packet Loss Ratio (PLR) and Effective Burstiness Probability (EBP), are extracted based on a novel Markov model of voicing-aware packet loss which captures properly the feature of packet loss process as well as the voicing property of speech wave included in lost packets. The conceived voicing-aware packet loss model is calibrated at run time using an efficient packet loss event driven algorithm. The performance evaluation study shows that our voicing-aware speech quality estimate models outperform voicing-unaware speech quality estimate models, especially in terms of accuracy over a wide range of conditions. Moreover, it validates the accuracy of the developed parametric no-reference speech quality models. In fact, we found that predicted scores using our speech quality models achieve an excellent correlation with measured scores (>0.95) and a small mean absolute deviation (<0.25) for ITU-T G.729 and G.711 speech CODECs.
机译:VoIP会话的感知质量紧密取决于数据包丢失的模式,即数据包丢失的分布和持续时间。损失间的差距(损失的差距)持续时间越宽(越小),质量下降越低。而且,由于丢弃的语音分组对感知质量具有不平等的影响,使用相同的分组丢失模式而受损的一组语音序列导致不同程度的感知质量下降。因此,除了丢包模式之外,我们还考虑丢包中包含的语音波发声特征,以估计语音质量得分。我们区分浊音,清音和静音数据包。这使得可以在基于人的主观评分和机器计算的客观评分之间实现更好的相关性和准确性。本文提出了新颖的无参考参量语音质量估计模型,该模型考虑了丢失数据包中包含的信号波发声特征。精确地,我们开发了单独的语音质量估计模型,该模型使用精心设计的简单和多元回归分析来捕获已删除浊音或清音数据包的感知效果。在严格的多元线性回归分析之后,开发了一种新的语音质量估计模型,该模型将浊音和清音质量分数混合以计算评估间隔结束时的总体语音质量分数。基于新的清晰语音包丢失的马尔可夫模型,提取了建议的清晰语音语音质量估计模型的输入参数,即丢包率(PLR)和有效突发概率(EBP)。处理以及丢包中包含的语音波发声特性。使用有效的数据包丢失事件驱动算法,可以在运行时校准构思的语音识别数据包丢失模型。性能评估研究表明,我们的可识别语音的语音质量估计模型要优于不识别语音的语音质量估计模型,尤其是在各种条件下的准确性方面。此外,它验证了所开发的参量无参考语音质量模型的准确性。实际上,我们发现,对于ITU-T G.729和G.711语音编解码器,使用我们的语音质量模型预测的分数与测得的分数(> 0.95)表现出极好的相关性,并且平均绝对偏差小(<0.25)。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号