首页> 外文会议>IEEE Data Science Workshop >Distance-Penalized Active Learning via Markov Decision Processes
【24h】

Distance-Penalized Active Learning via Markov Decision Processes

机译:马尔可夫决策过程的距离惩罚主动学习

获取原文
获取外文期刊封面目录资料

摘要

We consider the problem of active learning in the context of spatial sampling, where the measurements are obtained by a mobile sampling unit. The goal is to localize the change point of a one-dimensional threshold classifier while minimizing the total sampling time, a function of both the cost of sampling and the distance traveled. In this paper, we present a general framework for active learning by modeling the search problem as a Markov decision process. Using this framework, we present time-optimal algorithms for the spatial sampling problem when there is a uniform prior on the change point, a known non-uniform prior on the change point, and a need to return to the origin for intermittent battery recharging. We demonstrate through simulations that our proposed algorithms significantly outperform existing methods while maintaining a low computational cost.
机译:我们考虑在空间采样的背景下的主动学习的问题,其中通过移动采样单元获得测量。目标是本地化一维阈值分类器的变更点,同时最小化总采样时间,采样成本和行驶距离的函数。在本文中,我们通过将搜索问题建模为Markov决策过程,提出了一个用于主动学习的一般框架。使用该框架,我们在改变点之前存在均匀时,我们为空间采样问题提供时间最佳算法,在改变点上已知的不均匀,并且需要返回到间歇电池充电的原点。我们通过模拟展示我们所提出的算法显着优于现有方法,同时保持低计算成本。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号