首页> 外文会议>European Workshop on Reinforcement Learning >A Near Optimal Policy for Channel Allocation in Cognitive Radio
【24h】

A Near Optimal Policy for Channel Allocation in Cognitive Radio

机译:认知无线电频道分配近的最佳政策

获取原文

摘要

Several tasks of interest in digital communications can be cast into the framework of planning in Partially Observable Markov Decision Processes (POMDP). In this contribution, we consider a previously proposed model for a channel allocation task and develop an approach to compute a near optimal policy. The proposed method is based on approximate (point based) value iteration in a continuous state Markov Decision Process (MDP) which uses a specific internal state as well as an original discretization scheme for the internal points. The obtained results provide interesting insights into the behavior of the optimal policy in the channel allocation model.
机译:在部分可观察到的马尔可夫决策过程(POMDP)中,可以将数字通信兴趣的几个任务施入规划框架中。在此贡献中,我们考虑先前提出的频道分配任务模型,并开发一种方法来计算近最佳政策。该方法基于在使用特定内部状态的连续状态马尔可夫决策过程(MDP)中的近似(基于点)值迭代,其使用特定的内部状态以及内部点的原始离散化方案。所获得的结果提供了有趣的见解,以渠道分配模型中最佳政策的行为。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号