首页> 外文会议>Advances in natural language processing >Robust Semi-supervised and Ensemble-Based Methods in Word Sense Disambiguation
【24h】

Robust Semi-supervised and Ensemble-Based Methods in Word Sense Disambiguation

机译:词义消歧的鲁棒半监督和基于集合的方法

获取原文
获取原文并翻译 | 示例

摘要

Mihalcea [1] discusses self-training and co-training in the context of word sense disambiguation and shows that parameter optimization on individual words was important to obtain good results. Using smoothed co-training of a naive Bayes classifier she obtains a 9.8% error reduction on Senseval-2 data with a fixed parameter setting. In this paper we test a semi-supervised learning algorithm with no parameters, namely tri-training [2]. We also test the random subspace method [3] for building committees out of stable learners. Both techniques lead to significant error reductions with different learning algorithms, but improvements do not accumulate. Our best error reduction is 7.4%, and our best absolute average over Senseval-2 data, though not directly comparable, is 12% higher than the results reported in Mihalcea [1]. Keywords: co-training, tri-training, word sense disambiguation.
机译:Mihalcea [1]在单词义歧义化的背景下讨论了自我训练和共同训练,并表明对单个单词进行参数优化对于获得良好结果非常重要。通过使用朴素的贝叶斯分类器的平滑共训练,她在使用固定参数设置的Senseval-2数据上获得了9.8%的误差减少。在本文中,我们测试了一种无参数的半监督学习算法,即三训练[2]。我们还测试了随机子空间方法[3],用于从稳定的学习者中组建委员会。两种技术都可以使用不同的学习算法来显着减少错误,但是并不能累积改进。我们的最大误差减少为7.4%,而我们与Senseval-2数据相比的最佳绝对平均值,尽管没有直接可比性,但比Mihalcea [1]中报告的结果高12%。关键字:共同训练,三重训练,词义消歧。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号