首页> 外文会议>Workshop on Innovative use of NLP for Building Educational Applications >Training and Domain Adaptation for Supervised Text Segmentation
【24h】

Training and Domain Adaptation for Supervised Text Segmentation

机译:培训和域适应监督案文分割

获取原文

摘要

Unlike traditional unsupervised text segmentation methods, recent supervised segmentation models rely on Wikipedia as the source of large-scale segmentation supervision. These models have, however, predominantly been evaluated on the in-domain (Wikipedia-based) test sets, preventing conclusions about their general segmentation efficacy. In this work, we focus on the domain transfer performance of supervised neural text segmentation in the educational domain. To this end, we first introduce K12SEG, a new dataset for evaluation of supervised segmentation, created from educational reading material for grade-1 to college-level students. We then benchmark a hierarchical text segmentation model (HITS), based on RoBERTa, in both in-domain and domain-transfer segmentation experiments. While HITS produces state-of-the-art in-domain performance (on three Wikipedia-based test sets), we show that, subject to the standard fullblown fine-tuning, it is susceptible to domain overfitting. We identify adapter-based fine-tuning as a remedy that substantially improves transfer performance.
机译:与传统无监督的文本分割方法不同,最近监督分割模型依靠维基百科作为大规模分割监管的来源。然而,这些模型主要在域名(基于维基百科的)测试集上进行了评估,防止了其总结疗效的结论。在这项工作中,我们专注于教育领域监督神经文本细分的域转移性能。为此,我们首先介绍K12SeG,一个新数据集进行评估,用于评估监督分割,从教育阅读材料到大学学生创建。然后,我们基于域和域传输分割实验的基于Roberta基于Roberta来基准测试分层文本分段模型(HITS)。虽然点击产生最先进的域表现(基于三个维基百科的测试集),但我们认为,在标准的普通精细调整后,它易于域过度装备。我们将基于适配器的微调识别为基本上提高转移性​​能的补救措施。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号