...
首页> 外文期刊>Data technologies and applications >Robust cross-lingual knowledge base question answering via knowledge distillation
【24h】

Robust cross-lingual knowledge base question answering via knowledge distillation

机译:强大的跨语言知识库的问题通过知识回答蒸馏

获取原文
获取原文并翻译 | 示例
           

摘要

Previous knowledge base question answering (KBQA) models only consider the monolingual scenario and cannot be directly extended to the cross-lingual scenario, in which the language of questions and that of knowledge base (KB) are different. Although a machine translation (MT) model can bridge the gap through translating questions to the language of KB, the noises of translated questions could accumulate and further sharply impair the final performance. Therefore, the authors propose a method to improve the robustness of KBQA models in the cross-lingual scenario. Design/methodology/approach: The authors propose a knowledge distillation-based robustness enhancement (KDRE) method. Specifically, first a monolingual model (teacher) is trained by ground truth (GT) data. Then to imitate the practical noises, a noise-generating model is designed to inject two types of noise into questions: general noise and translation-aware noise. Finally, the noisy questions are input into the student model. Meanwhile, the student model is jointly trained by GT data and distilled data, which are derived from the teacher when feeding GT questions. Findings: The experimental results demonstrate that KDRE can improve the performance of models in the cross-lingual scenario. The performance of each module in KBQA model is improved by KDRE. The knowledge distillation (KD) and noise-generating model in the method can complementarily boost the robustness of models. Originality/value: The authors first extend KBQA models from monolingual to cross-lingual scenario. Also, the authors first implement KD for KBQA to develop robust cross-lingual models.
机译:以前的基础知识问答(KBQA)模型只考虑单语和场景不能直接扩展到跨语言场景中,语言的问题知识库(KB)是不同的。虽然机器翻译(MT)模型通过翻译问题的桥梁KB的语言,翻译的声音问题可能会进一步积累和尖锐影响最终的性能。作者提出一个方法来改善在跨语言KBQA模型的鲁棒性场景。作者提出一个distillation-based知识鲁棒性增强(KDRE)方法。具体地说,第一单语模型(老师)是由地面真理(GT)训练数据。模仿实际噪声,噪声产生模型设计注入两种类型的噪音问题:噪音和将军translation-aware噪音。问题是输入学生模型。与此同时,学生模型共同训练通过GT数据和蒸馏数据,派生从老师给GT提问。结果:实验结果证明KDRE可以提高模型的性能在跨语言的场景。每个模块在KBQA模型由KDRE改进。蒸馏(KD)的知识噪声产生模型的方法互补提高模型的鲁棒性。创意/价值:作者首先扩展KBQA模型从单语到跨语言场景。KBQA开发健壮的跨语言模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号