首页> 外文会议>International conference on statistical language and speech processing >Weakly Supervised Discriminative Training of Linear Models for Natural Language Processing
【24h】

Weakly Supervised Discriminative Training of Linear Models for Natural Language Processing

机译:用于自然语言处理的线性模型的弱监督判别训练

获取原文

摘要

This work explores weakly supervised training of discriminative linear classifiers. Such features-rich classifiers have been widely adopted by the Natural Language processing (NLP) community because of their powerful modeling capacity and their support for correlated features, which allow separating the expert task of designing features from the core learning method. However, unsupervised training of discriminative models is more challenging than with generative models. We adapt a recently proposed approximation of the classifier risk and derive a closed-form solution that greatly speeds-up its convergence time. This method is appealing because it provably converges towards the minimum risk without any labeled corpus, thanks to only two reasonable assumptions about the rank of class marginal and Gaussianity of class-conditional linear scores. We also show that the method is a viable, interesting alternative to achieve weakly supervised training of linear classifiers in two NLP tasks: predicate and entity recognition.
机译:这项工作探索了区分线性分类器的弱监督训练。这种功能丰富的分类器由于其强大的建模能力和对相关特征的支持,已被自然语言处理(NLP)社区广泛采用,从而可以将设计特征的专家任务与核心学习方法分开。但是,与生成模型相比,对歧视模型的无监督训练更具挑战性。我们采用了最近提出的分类器风险的近似方法,并得出了一种封闭形式的解决方案,该解决方案大大加快了其收敛速度。这种方法之所以具有吸引力,是因为只有两个关于类边际的等级和类条件线性得分的高斯性的合理假设,它可证明地收敛到最小风险而没有任何标记语料库。我们还表明,该方法是在两个NLP任务:谓词和实体识别中实现线性分类器的弱监督训练的可行且有趣的替代方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号