首页> 外文会议>International Workshop on Search-Oriented Conversational AI >Multi-Task Learning using Dynamic Task Weighting for Conversational Question Answering
【24h】

Multi-Task Learning using Dynamic Task Weighting for Conversational Question Answering

机译:使用动态任务加权对会话问题应答的多任务学习

获取原文
获取外文期刊封面目录资料

摘要

Conversational Question Answering (Con-vQA) is a Conversational Search task in a simplified setting, where an answer must be extracted from a given passage. Neural language models, such as BERT, fine-tuned on large-scale ConvQA datasets such as CoQA and QuAC have been used to address this task. Recently, Multi-Task Learning (MTL) has emerged as a particularly interesting approach for developing ConvQA models, where the objective is to enhance the performance of a primary task by sharing the learned structure across several related auxiliary tasks. However, existing ConvQA models that leverage MTL have not investigated the dynamic adjustment of the relative importance of the different tasks during learning, nor the resulting impact on the performance of the learned models. In this paper, we first study the effectiveness and efficiency of dynamic MTL methods including Evolving Weighting, Uncertainty Weighting, and Loss-Balanced Task Weighting, compared to static MTL methods such as the uniform weighting of tasks. Furthermore, we propose a novel hybrid dynamic method combining Abridged Linear for the main task with a Loss-Balanced Task Weighting (LBTW) for the auxiliary tasks, so as to automatically fine-tune task weighting during learning, ensuring that each of the tasks' weights is adjusted by the relative importance of the different tasks. We conduct experiments using QuAC, a large-scale ConvQA dataset. Our results demonstrate the effectiveness of our proposed method, which significantly outperforms both the single-task learning and static task weighting methods with improvements ranging from +2.72% to +3.20% in Fl scores. Finally, our findings show that the performance of using MTL in developing ConvQA model is sensitive to the correct selection of the auxiliary tasks as well as to an adequate balancing of the loss rates of these tasks during training by using LBTW.
机译:会话问题应答(Con-VQA)是简化设置中的会话搜索任务,其中必须从给定段落中提取答案。神经语言模型,如BERT,微调在大型CONVQA数据集(如COQA和QUAC)上用于解决此任务。最近,多任务学习(MTL)被出现为开发ConvQA模型的特别有趣的方法,其中目标是通过在多个相关辅助任务中共享学习结构来增强主要任务的性能。然而,利用MTL的现有ConvQA模型尚未调查学习期间不同任务的相对重要性的动态调整,也不会产生对学习模型的性能的影响。在本文中,我们首先研究动态MTL方法的有效性和效率,包括演变的加权,不确定性加权和损失 - 平衡任务加权,例如静态MTL方法,例如任务的均匀加权。此外,我们提出了一种新颖的混合动态方法,将销售线性与辅助任务的丢失均衡任务加权(LBTW)结合起来的混合动态方法,以便在学习期间自动微调任务加权,确保每个任务通过不同任务的相对重要性来调整重量。我们使用Quac进行实验,这是一个大规模的ConvQA数据集。我们的结果表明了我们提出的方法的有效性,这显着优于单一任务学习和静态任务加权方法,其改善范围从+ 2.72%到+ 3.20%的流量。最后,我们的研究结果表明,在开发CONDQA模型中使用MTL的性能对辅助任务的正确选择敏感,以及通过使用LBTW在训练期间对这些任务的损失率进行充分平衡。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号