首页> 外文期刊>Mathematical Problems in Engineering: Theory, Methods and Applications >A Case Study on Air Combat Decision Using Approximated Dynamic Programming
【24h】

A Case Study on Air Combat Decision Using Approximated Dynamic Programming

机译:近似动态规划空战决策的案例研究

获取原文
           

摘要

As a continuous state space problem, air combat is difficult to be resolved by traditional dynamic programming (DP) with discretized state space. The approximated dynamic programming (ADP) approach is studied in this paper to build a high performance decision model for air combat in 1 versus 1 scenario, in which the iterative process for policy improvement is replaced by mass sampling from history trajectories and utility function approximating, leading to high efficiency on policy improvement eventually. A continuous reward function is also constructed to better guide the plane to find its way to “winner” state from any initial situation. According to our experiments, the plane is more offensive when following policy derived from ADP approach other than the baseline Min-Max policy, in which the “time to win” is reduced greatly but the cumulated probability of being killed by enemy is higher. The reason is analyzed in this paper.
机译:作为连续的状态空间问题,传统的动态编程(DP)难以通过离散状态空间来解决空战。 在本文中研究了近似的动态编程(ADP)方法,为1与1场景中的空战构建高性能决策模型,其中历史轨迹和近似的历史轨迹和实用函数的质量采样取代了迭代过程。 最终导致政策改进的高效率。 还构建了连续奖励功能,以更好地指导飞机找到它的方式“赢家” 任何初始情况的国家。 根据我们的实验,当遵循基线MIN-MAX政策以外的ADP方法的遵循策略时,飞机更令人反感,其中“时间赢得” 大大减少,但被敌人杀死的累积概率更高。 本文分析了原因。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号