...
【24h】

Reinterpreting CTC training as iterative fitting

机译:重新解释CTC培训作为迭代配件

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

The connectionist temporal classification (CTC) enables end-to-end sequence learning by maximizing the probability of correctly recognizing sequences during training. The outputs of a CTC-trained model tend to form a series of spikes separated by strongly predicted blanks, know as the spiky problem. To figure out the reason for it, we reinterpret the CTC training process as an iterative fitting task that is based on frame-wise cross-entropy loss. It offers us an intuitive way to compare target probabilities with model outputs for each iteration, and explain how the model outputs gradually turns spiky. Inspired by it, we put forward two ways to modify the CTC training. The experiments demonstrate that our method can well solve the spiky problem and moreover, lead to faster convergence over various training settings. Beside this, the reinterpretation of CTC, as a brand new perspective, may be potentially useful in other situations. The code is publicly available at https://github.com/hzli-ucas/caffe/tree/ctc. (C) 2020 Elsevier Ltd. All rights reserved.
机译:通过在训练期间最大化正确识别序列的概率来最大限度地,连接员时间分类(CTC)能够实现端到端序列学习。 CTC培训的模型的输出倾向于形成一系列由强烈预测的空白分开的尖峰,知道是尖刺的问题。要弄清楚它的原因,我们将CTC培训过程重新诠释为基于帧展跨熵损失的迭代拟合任务。它为我们提供了一种直观的方式来比较每个迭代的模型输出的目标概率,并解释模型输出如何逐渐变为尖峰。灵感来自于,我们提出了两种方法来修改CTC培训。实验表明,我们的方法可以很好地解决尖峰问题,而且,导致各种训练环境更快地收敛。除此之外,CTC作为全新角色的重新解释可能在其他情况下可能有用。该代码在https://github.com/hzli-ucas/caffe/tree/ctc上公开使用。 (c)2020 elestvier有限公司保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号