【24h】

Lattice Transformer for Speech Translation

机译:用于语音翻译的莱迪思变压器

获取原文

摘要

Recent advances in sequence modeling have highlighted the strengths of the transformer architecture, especially in achieving state-of-the-art machine translation results. However, depending on the up-stream systems, e.g., speech recognition, or word segmentation, the input to translation system can vary greatly. The goal of this work is to extend the attention mechanism of the transformer to naturally consume the lattice in addition to the traditional sequential input. We first propose a general lattice transformer for speech translation where the input is the output of the automatic speech recognition (ASR) which contains multiple paths and posterior scores. To leverage the extra information from the lattice structure, we develop a novel controllable lattice attention mechanism to obtain latent representations. On the LDC Spanish-English speech translation corpus, our experiments show that lattice transformer generalizes significantly better and outperforms both a transformer baseline and a lattice LSTM. Additionally, we validate our approach on the WMT 2017 Chinese-English translation task with lattice inputs from different BPE segmentations. In this task, we also observe the improvements over strong baselines.
机译:序列建模的最新进展凸显了变压器架构的优势,特别是在实现最新的机器翻译结果方面。然而,取决于上游系统,例如语音识别或单词分段,翻译系统的输入可以有很大的变化。这项工作的目的是扩展变压器的注意力机制,以使除传统的顺序输入外自然消耗晶格。我们首先提出一种用于语音翻译的通用晶格变换器,其中输入是自动语音识别(ASR)的输出,其中包含多个路径和后验分数。为了利用来自晶格结构的额外信息,我们开发了一种新颖的可控晶格注意机制来获得潜在表示。在最不发达国家(LDC)西班牙语-英语语音翻译语料库上,我们的实验表明,晶格变换器的泛化效果明显更好,并且优于变换器基线和晶格LSTM。此外,我们使用来自不同BPE细分的格点输入验证了我们在WMT 2017汉英翻译任务中的方法。在此任务中,我们还将观察到在强大基准上的改进。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号