首页> 外文会议>Conference on empirical methods in natural language processing >Neural Transductive Learning and Beyond: Morphological Generation in the Minimal-Resource Setting
【24h】

Neural Transductive Learning and Beyond: Morphological Generation in the Minimal-Resource Setting

机译:神经转导学习及超越:最小资源设置中的形态学

获取原文

摘要

Neural state-of-the-art sequence-to-sequence (seq2seq) models often do not perform well for small training sets. We address paradigm completion, the morphological task of, given a partial paradigm, generating all missing forms. We propose two new methods for the minimal-resource setting: (ⅰ) Paradigm transduction: Since we assume only few paradigms available for training, neural seq2seq models are able to capture relationships between paradigm cells, but are tied to the idiosyncracies of the training set. Paradigm transduction mitigates this problem by exploiting the input subset of inflected forms at test time, (ⅱ) Source selection with high precision (SHIP): Multi-source models which learn to automatically select one or multiple sources to predict a target inflection do not perform well in the minimal-resource setting. SHIP is an alternative to identify a reliable source if training data is limited. On a 52-language benchmark dataset, we outperform the previous state of the art by up to 9.71% absolute accuracy.
机译:神经最新的序列到序列(SEQ2Seq)模型通常对小型训练集不起作用。我们解决了范式完成,给予部分范式的形态任务,产生所有缺失的表格。我们提出了两种新的资源设置方法:(Ⅰ)范式转导:由于我们假设只有少数可用于培训的范式,神经SEQ2SEQ模型能够捕获范式细胞之间的关系,但与培训集的特质相关联。范例转导通过利用测试时间的输入子集来减轻该问题,(Ⅱ)源选择具有高精度(船舶):多源模型,用于自动选择一个或多个来源以预测目标拐点不会表现井在最小资源设置。如果培训数据有限,船是识别可靠源的替代方案。在52语言的基准数据集上,我们优于先前的最先进的绝对精度为9.71%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号