首页> 外文会议>International Conference on Statistical Language and Speech Processing >Weakly Supervised Discriminative Training of Linear Models for Natural Language Processing
【24h】

Weakly Supervised Discriminative Training of Linear Models for Natural Language Processing

机译:用于自然语言处理的线性模型的弱敏感判断

获取原文

摘要

This work explores weakly supervised training of discriminative linear classifiers. Such features-rich classifiers have been widely adopted by the Natural Language processing (NLP) community because of their powerful modeling capacity and their support for correlated features, which allow separating the expert task of designing features from the core learning method. However, unsupervised training of discriminative models is more challenging than with generative models. We adapt a recently proposed approximation of the classifier risk and derive a closed-form solution that greatly speeds-up its convergence time. This method is appealing because it provably converges towards the minimum risk without any labeled corpus, thanks to only two reasonable assumptions about the rank of class marginal and Gaussianity of class-conditional linear scores. We also show that the method is a viable, interesting alternative to achieve weakly supervised training of linear classifiers in two NLP tasks: predicate and entity recognition.
机译:这项工作探讨了歧视性线性分类器的弱监督培训。自然语言处理(NLP)社区已被强大的建模能力及其对相关特征的支持,这些功能丰富的分类器已被广泛采用,这允许将专家任务从核心学习方法中分离。然而,对歧视性模型的无监督培训比生成模型更具挑战性。我们调整最近提出的分类器风险的近似,并得出了封闭式解决方案,其大大加速其收敛时间。这种方法令人吸引人,因为它可以达到没有任何标记的语料库的最低风险,因为只有两个关于类条件线性分数的阶级边际和高斯的级别的合理假设。我们还表明,该方法是一种可行,有趣的替代方案,可以在两个NLP任务中实现线性分类器的弱监督培训:谓词和实体识别。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号