首页> 外文会议>Annual conference on Neural Information Processing Systems >Learned Prioritization for Trading Off Accuracy and Speed
【24h】

Learned Prioritization for Trading Off Accuracy and Speed

机译:权衡准确性和速度的学习优先级

获取原文

摘要

Users want inference to be both fast and accurate, but quality often comes at the cost of speed. The field has experimented with approximate inference algorithms that make different speed-accuracy tradeoffs (for particular problems and datasets). We aim to explore this space automatically, focusing here on the case of agenda-based syntactic parsing [12]. Unfortunately, off-the-shelf reinforcement learning techniques fail to learn good policies: the state space is simply too large to explore naively. An attempt to counteract this by applying imitation learning algorithms also fails: the "teacher" follows a far better policy than anything in our learner's policy space, free of the speed-accuracy tradeoff that arises when oracle information is unavailable, and thus largely insensitive to the known reward functfion. We propose a hybrid reinforcement/apprenticeship learning algorithm that learns to speed up an initial policy, trading off accuracy for speed according to various settings of a speed term in the loss function.
机译:用户希望推理既快速又准确,但是质量通常是以速度为代价的。该领域已经尝试了近似推理算法,这些算法做出了不同的速度精度权衡(针对特定问题和数据集)。我们旨在自动探索该空间,这里重点讨论基于议程的句法分析[12]。不幸的是,现成的强化学习技术无法学习良好的政策:状态空间太大而无法天真地探索。通过应用模仿学习算法来抵消这种情况的尝试也失败了:“教师”遵循的策略要比我们学习者的策略空间中的任何方法都要好得多,没有在无法获得oracle信息时出现的速度准确性权衡,因此对已知的奖励功能。我们提出了一种混合式强化/学徒学习算法,该算法学习加速初始策略,并根据损失函数中速度项的各种设置来权衡速度的准确性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号