首页> 外文会议>National Conferences on Aritificial Intelligence >Reinforcement Learning for Trading Systems and Portfolios
【24h】

Reinforcement Learning for Trading Systems and Portfolios

机译:交易系统和投资组合的加强学习

获取原文

摘要

We propose to train trading systems by optimizing financial objective functions via reinforcement learning. The performance functions that we consider as value functions are profit or wealth, the Sharpe ratio and our recently proposed differential Sharpe ratio for online learning. In Moody & Wu (1997), we presented empirical results in controlled experiments that demonstrated the advantages of reinforcement learning relative to supervised learning. Here we extend our previous work to compare Q-Learning to a reinforcement learning technique based on real-time recurrent learning (RTRL) that maximizes immediate reward. Our simulation results include a spectacular demonstration of the presence of predictability in the monthly Standard and Poors 500 stock index for the 25 year period 1970 through 1994. Our reinforcement trader achieves a simulated out-of-sample profit of over 4000% for this period, compared to the return for a buy and hold strategy of about 1300% (with dividends rein-vested). This superior result is achieved with substantially lower risk.
机译:我们建议通过加强学习优化金融目标职能来培训贸易系统。我们认为作为价值函数的绩效函数是利润或财富,锐利比率和我们最近提出的在线学习的差异尺度比率。在穆迪和吴(1997年),我们提出在受控实验证明该学习相对增强的优势,监督学习实证结果。在这里,我们将先前的工作扩展到基于实时复发学习(RTRL)的加强学习技术进行比较,以最大化即时奖励。我们的仿真结果包括在1970年至1994年25年期间的月度标准和诗500股指数中存在可预测性的壮观演示。我们的加固交易员在此期间实现了模拟的营业利润超过4000%,与买卖的回报相比,持有约1300%的策略(股息缰绳救了归属)。这种卓越的结果是通过大大降低的风险实现的。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号