首页> 外文会议>19th International Conference on Computational Linguistics Coling 2002 Vol.1 Aug 26-30, 2002 Taipei, Taiwan >Improved Iterative Scaling can yield multiple globally optimal models with radically differing performance levels
【24h】

Improved Iterative Scaling can yield multiple globally optimal models with radically differing performance levels

机译:改进的迭代缩放比例可以产生具有根本不同性能水平的多个全局最优模型

获取原文
获取原文并翻译 | 示例

摘要

Log-linear models can be efficiently estimated using algorithms such as Improved Iterative Scaling (IIS) (Lafferty et al, 1997). Under certain conditions and for a particular class of problems, IIS is guaranteed to approach both the maximum-likelihood and maximum entropy solution. This solution, in likelihood space, is unique. Unfortunately, in realistic situations, multiple solutions may exist, all of which are equivalent to each other in terms of likelihood, but radically different from each other in terms of performance. We show that this behaviour can occur when a model contains overlapping features and the training material is sparse. Experimental results, from the domain of parse selection for stochastic attribute value grammars, shows the wide variation in performance that can be found when estimating models using IIS. Further results show that the influence of the initial model can be diminished by selecting either uniform weights, or else by model averaging.
机译:对数线性模型可以使用诸如改进的迭代缩放(IIS)(Lafferty等,1997)之类的算法进行有效估计。在某些情况下,对于特定类别的问题,可以保证IIS能够同时解决最大似然和最大熵的问题。在可能性空间中,此解决方案是唯一的。不幸的是,在现实情况下,可能存在多个解决方案,所有这些解决方案在可能性方面彼此相等,但在性能方面却彼此根本不同。我们表明,当模型包含重叠特征并且培训材料稀疏时,就会发生这种现象。从针对随机属性值语法的语法选择的分析领域得出的实验结果表明,使用IIS估计模型时,可以发现性能上的巨大差异。进一步的结果表明,可以通过选择均匀权重或通过模型平均来减小初始模型的影响。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号