【24h】

Benchmarking Approximate Inference Methods for Neural Structured Prediction

机译:神经结构预测的基准近似推理方法

获取原文

摘要

Exact structured inference with neural network scoring functions is computationally challenging but several methods have been proposed for approximating inference. One approach is to perform gradient descent with respect to the output structure directly (Belanger and McCallum, 2016). Another approach, proposed recently, is to train a neural network (an "inference network") to perform inference (Tu and Gimpel, 2018). In this paper, we compare these two families of inference methods on three sequence labeling datasets. We choose sequence labeling because it permits us to use exact inference as a benchmark in terms of speed, accuracy, and search error. Across datasets, we demonstrate that inference networks achieve a better speed/accuracy/search error trade-off than gradient descent, while also being faster than exact inference at similar accuracy levels. We find further benefit by combining inference networks and gradient descent, using the former to provide a warm start for the latter.
机译:具有神经网络评分功能的精确结构化推论在计算上具有挑战性,但已提出了几种近似推论的方法。一种方法是直接对输出结构执行梯度下降(Belanger和McCallum,2016年)。最近提出的另一种方法是训练神经网络(“推理网络”)以执行推理(Tu和Gimpel,2018年)。在本文中,我们在三个序列标记数据集上比较了这两种推理方法。我们选择序列标记是因为它允许我们在速度,准确性和搜索错误方面使用精确推断作为基准。在整个数据集中,我们证明了与梯度下降相比,推理网络实现了更好的速度/精度/搜索误差折衷,同时在相似的精度水平上也比精确推理要快。通过将推理网络和梯度下降相结合,使用前者为后者提供了一个良好的开端,我们发现了更多的好处。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号