...
首页> 外文期刊>Electronics Letters >Twice fine-tuning deep neural networks for paraphrase identification
【24h】

Twice fine-tuning deep neural networks for paraphrase identification

机译:两次微调深神经网络,用于解释识别

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

In this Letter, the authors introduce a novel approach to learn representations for sentence-level paraphrase identification (PI) using BERT and ten natural language processing tasks. Their method trains an unsupervised model called BERT with two different tasks to detect whether two sentences are in paraphrase relation or not. Unlike conventional BERT, which fine tunes the target task such as PI to pre-trained BERT, twice fine-tuning deep neural networks first fine tune each task (e.g. general language understanding evaluation tasks, question answering, and paraphrase adversaries from word scrambling task) and second fine tune target PI task. As a result, the multi-fine-tuned BERT model outperformed the fine-tuned model only with Microsoft Research Paraphrase Corpus (MRPC), which is paraphrase data, except for one case of Stanford Sentiment Treebank - 2 (SST-2). Multi-task fine-tuning is a simple idea but experimentally powerful. Experiments show that fine-tuning just PI tasks to the BERT already gives enough performance, but additionally, fine-tuning similar tasks can affect performance (3.4% point absolute improvement) and be competitive with the state-of-the-art systems.
机译:在这封信中,作者介绍了一种使用BERT和十种自然语言处理任务来学习句子级解释识别(PI)的表示的新方法。他们的方法列举了一个叫做伯特的无人监督模型,用两个不同的任务来检测两个句子是否在释义关系中。与传统的伯特不同,这很好地调整PI的目标任务,例如PI预训练的伯特,两次微调深神经网络首先微调每个任务(例如,从Word Scrambling任务中的一般语言认识到评估任务,问题应答和解释者对手)和第二次微调目标PI任务。结果,多细小调谐BERT模型仅与Microsoft Research acraphrase语料库(MRPC)相表达了微调模型,除了一个斯坦福情绪TreeBank - 2(SST-2)的一个例外。多任务微调是一个简单的想法,但实验强大。实验表明,微调只是PI任务到伯特已经提供了足够的性能,但另外,微调类似的任务可能会影响性能(3.4%的点绝对改善)并与最先进的系统竞争。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号