首页> 外文学位 >An Approximate Dynamic Programming Approach to Financial Execution for Weapon System Programs.
【24h】

An Approximate Dynamic Programming Approach to Financial Execution for Weapon System Programs.

机译:武器系统程序财务执行的一种近似动态编程方法。

获取原文
获取原文并翻译 | 示例

摘要

During each twelve month fiscal year (FY) cycle weapon system programs across the Department of Defense (DoD) are expected to execute their allocated budgets in an expedient and timely manner. As the FY progresses, a weapon system's cash flow state at any given moment is primarily measured by the cumulative amounts of their budget that are either committed, obligated, accrued, or expended. Regulatory and oversight initiatives such as midyear financial execution reviews and published monthly execution goals serve as measures that are designed to ensure that there is in fact high utilization of a weapon system's allocated yearly budget. The challenge of finding an ideal monthly commitment cash flow policy that achieves a high level of utilization can be expressed as a sequential decision making problem over time. The mathematical area known as Markovian analysis is dedicated to modeling and finding solution methods that focus on such problems with emphasis on understanding how the system moves from state to state throughout the decision process. The complexity of the problem examined in this research stems from the size of the multimillion dollar budgets in question and the numerous projects they fund. In turn, weapon system offices must make hundreds of commitment action determinations over any given fiscal year in an environment of uncertainty. This intricate decision system necessitates that decision makers have good mathematical tools that can assist them with determining an optimal commitment policy.;The research described in this thesis uses approximate dynamic programming (ADP) techniques as a solution method to the financial execution commitment problem for DoD weapon system programs. ADP ideas and concepts are extensions of Markovian analysis principles. The modeling effort uses a simulation based optimization method specifically geared towards solving sequential decision making problems. The more traditional dynamic programming (DP) approaches are variants on the implementation of Bellman's recursive optimality equation. Unfortunately, as a result of the "curse of dimensionality" and the "curse of modeling" these classical methods tend to breakdown when applied within the more complex problem structure scenarios. The ADP approach expands upon the original recursive idea embedded in Bellman's optimality equation and addresses the difficulties associated with the "curse of dimensionality" and the "curse of modeling".;As part of this research, two types of ADP models were built around the use of a post decision state (PDS) variable. The application of the models was tested against a collection of theoretical financial execution project scenarios. The initial model leveraged a Q-learning design, while the second model used value function learning. In each approach, the formulation of an optimal policy was dependent upon three modeling phases. The three phases are referred to as exploration, exploitation (learning), and learnt. The exploration phase of the model relaxes the driving optimality conditions while simulating the execution decision system. The exploitation or learning phase incorporates the optimality conditions within the simulation environment. Lastly, the learnt phase leverages the outputs produced by exploration and exploitation to provide the recommended optimal policy. Additionally, the learnt phase of the models was designed to provide a means for conducting various sensitivity analysis and financial execution drill excursions.;The research resulted in a unique application of ADP as a simulation and problem solving method for generating financial execution commitment policies. The generated ADP polices or commitment plans were compared against an alternative myopic policy approach referred to as a stubby pencil policy. The learnt modeling phase examined and tested the reaction of both the ADP and stubby pencil policies under various expenditure conditions. The analysis showed that the ADP commitment strategy was often either equal or less than that of the myopic stubby pencil strategy. These results suggest that a decision maker following an ADP strategy would either reach full commitment of the budget at a later date or would not reach full commitment of the budget prior to the end of the FY. In the latter case, the remaining uncommitted dollars serve as an indication that improved cash utilization could be obtained by incorporating more work or projects into the budget.
机译:在国防部(DoD)的每个十二个月财政年度(FY)周期中,预计武器系统程序将以方便,及时的方式执行分配的预算。随着FY的进行,武器系统在任何给定时刻的现金流量状态主要由其预算的已承诺,已承付,应计或已支出的累计金额来衡量。诸如年中财务执行情况审查和已发布的每月执行目标之类的监管和监督措施旨在确保实际上充分利用武器系统分配的年度预算。寻找理想的每月承诺现金流量策略以实现较高利用率的挑战可以表示为随着时间的推移而出现的顺序决策问题。称为马尔可夫分析的数学领域致力于建模和寻找解决方法,这些解决方法着重于此类问题,重点在于理解系统在整个决策过程中如何从一个状态移动到另一个状态。本研究中研究的问题的复杂性源于相关数百万美元预算的规模以及他们资助的众多项目。反过来,在不确定的环境下,武器系统办公室必须在任何给定的财政年度做出数百项承诺行动决定。这种复杂的决策系统需要决策者拥有良好的数学工具,可以帮助他们确定最佳承诺策略。本论文中描述的研究使用近似动态规划(ADP)技术作为DoD财务执行承诺问题的解决方法。武器系统程序。 ADP思想和概念是马尔可夫分析原理的扩展。建模工作使用基于仿真的优化方法,专门用于解决顺序决策问题。更传统的动态规划(DP)方法是Bellman递归最优方程实现的变体。不幸的是,由于“维数的诅咒”和“建模的诅咒”,这些经典方法在更复杂的问题结构场景中应用时往往会崩溃。 ADP方法是在Bellman最优性方程中嵌入的原始递归思想的基础上扩展的,并解决了与“维数诅咒”和“建模诅咒”相关的难题。;作为本研究的一部分,围绕ADP模型建立了两种类型的ADP模型。使用决策后状态(PDS)变量。针对一系列理论财务执行项目方案测试了模型的应用。最初的模型利用了Q学习设计,而第二个模型则使用了价值函数学习。在每种方法中,最优政策的制定都取决于三个建模阶段。这三个阶段称为探索,开发(学习)和学习。该模型的探索阶段在模拟执行决策系统时放宽了驾驶最佳条件。开发或学习阶段将最佳条件纳入仿真环境中。最后,学习阶段利用勘探和开发产生的输出来提供建议的最佳策略。此外,模型的学习阶段旨在为进行各种敏感性分析和财务执行演习游览提供一种手段。研究结果是ADP在生成财务执行承诺政策的模拟和问题解决方法方面的独特应用。将生成的ADP政策或承诺计划与另一种称为近视铅笔政策的近视政策方法进行了比较。学习的建模阶段检查并测试了在各种支出条件下ADP和粗笔政策的反应。分析表明,ADP承诺策略通常等于或小于近视短铅笔策略。这些结果表明,采用ADP策略的决策者要么在以后的日期达到预算的全部承诺,要么在FY结束之前未达到预算的全部承诺。在后一种情况下,剩余的未承付美元表示可以通过将更多工作或项目纳入预算来提高现金利用率。

著录项

  • 作者

    Morman, Erich D.;

  • 作者单位

    George Mason University.;

  • 授予单位 George Mason University.;
  • 学科 Operations Research.;Economics Finance.
  • 学位 Ph.D.
  • 年度 2013
  • 页码 168 p.
  • 总页数 168
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号