首页> 外文会议>CCF international conference on natural language processing and Chinese computing >Learning Domain Invariant Word Representations for Parsing Domain Adaptation
【24h】

Learning Domain Invariant Word Representations for Parsing Domain Adaptation

机译:学习领域不变词表示法以解析领域适应

获取原文

摘要

We show that strong domain adaptation results for dependency parsing can be achieved using a conceptually simple method that learns domain-invariant word representations. Lacking labeled resources, dependency parsing for low-resource domains has been a challenging task. Existing work considers adapting a model trained on a resource-rich domain to low-resource domains. A mainstream solution is to find a set of shared features across domains. For neural network models, word embeddings are a fundamental set of initial features. However, little work has been done investigating this simple aspect. We propose to learn domain-invariant word representations by fine-tuning pretrained word representations adversarially. Our parser achieves error reductions of 5.6% UAS, 7.9% LAS on PTB respectively, and 4.2% UAS, 3.2% LAS on Genia respectively, showing the effectiveness of domain invariant word representations for alleviating lexical bias between source and target data.
机译:我们表明,可以使用学习域不变的词表示形式的概念上简单的方法来获得强大的域适应结果,以进行相关性分析。缺少标记资源,用于低资源域的依存关系解析一直是一项艰巨的任务。现有工作考虑将在资源丰富的域上训练的模型改编为资源匮乏的域。主流解决方案是跨域查找一组共享功能。对于神经网络模型,词嵌入是一组基本的初始功能。但是,研究此简单方面的工作很少。我们建议通过对抗性地微调预训练的单词表示来学习领域不变的单词表示。我们的解析器实现的错误减少分别为PTB的5.6%UAS,7.9%LAS和Genia的4.2%UAS,3.2%LAS,显示了领域不变词表示在减轻源数据和目标数据之间的词汇偏向方面的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号