首页> 外文会议>Conference on uncertainty in artificial intelligence >Approximation of Lorenz-Optimal Solutions in Multiobjective Markov Decision Processes
【24h】

Approximation of Lorenz-Optimal Solutions in Multiobjective Markov Decision Processes

机译:多目标马尔可夫决策过程中的Lorenz最优解的逼近

获取原文
获取外文期刊封面目录资料

摘要

This paper is devoted to fair optimization in Multiobjective Markov Decision Processes (MOMDPs). A MOMDP is an extension of the MDP model for planning under uncertainty while trying to optimize several reward functions simultaneously. This applies to multiagent problems when rewards define individual utility functions, or in multicrite-ria problems when rewards refer to different features. In this setting, we study the determination of policies leading to Lorenz-non-dominated tradeoffs. Lorenz dominance is a refinement of Pareto dominance that was introduced in Social Choice for the measurement of inequalities. In this paper, we introduce methods to efficiently approximate the sets of Lorenz-non-dominated solutions of infinite-horizon, discounted MOMDPs. The approximations are polynomial-sized subsets of those solutions.
机译:本文致力于多目标马尔可夫决策过程(MOMDP)中的公平优化。 MOMDP是MDP模型的扩展,用于在不确定性下进行计划,同时尝试同时优化多个奖励功能。当奖励定义了单个效用函数时,这适用于多主体问题;当奖励指的是不同的功能时,这适用于多记录问题。在这种情况下,我们研究导致Lorenz非主导权衡的政策确定。 Lorenz主导地位是对Pareto主导地位的改进,它是在Social Choice中引入的,用于衡量不平等。在本文中,我们介绍了有效逼近无限水平折扣MOMDP的Lorenz非支配解集的方法。近似值是这些解的多项式大小的子集。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号