首页> 外文期刊>Expert systems with applications >Deep reinforcement learning based trading agents: Risk curiosity driven learning for financial rules-based policy
【24h】

Deep reinforcement learning based trading agents: Risk curiosity driven learning for financial rules-based policy

机译:基于深度强化学习的交易代理:金融规则的危险效力驱动学习

获取原文
获取原文并翻译 | 示例

摘要

Financial markets are complex dynamic systems influenced by a high number of active agents, which produce a behavior with high randomness and noise. Trading strategies are well depicted as an online decision-making problem involving imperfect information and aiming to maximize the return while restraining the risk. However, it is challenging to obtain an optimal strategy in the complex and dynamic stock market. Therefore, recent developments in similar environments have pushed researchers towards exciting new horizons.In this paper, a novel rule-based policy approach is proposed to train a deep reinforcement learning agent for automated financial trading. Precisely, a continuous virtual environment has been created, with different versions of agents trading against one another. During this multiplex process, the agents which are trained on 504 risky datasets, use the fundamental concepts of proximal policy optimization to improve their own decision making by adjusting their action choice against the uncertainty of states. Risk curiosity-driven learning acts as an intrinsic reward function and is heavily laden with signals to find salient relationships between actions and market behaviors. The trained agent based on curiosity-driven risk has steadily and progressively improved actions quality. The self-learned rules driven by the agent curiosity push the policy towards actions that yield a high performance over the environment. Experiments on 8 real-world stocks are given to verify the appropriateness and efficiency of the self-learned rules. The proposed system has achieved promising performances, made better trades using fewer transactions, and outperformed the state-of-the-art baselines.
机译:金融市场是受高量活性剂影响的复杂动态系统,其产生具有高随机性和噪声的行为。交易策略很好地描述为涉及不完美信息的在线决策问题,旨在最大限度地提高返回,同时限制风险。然而,在复杂和动态股市中获得最佳策略是挑战性的。因此,最近在类似环境中的发展已经推动了研究人员令人兴奋的新视野。本文提出了一种新的规则的政策方法,旨在培养一个自动化金融交易的深增强学习代理。精确地,已创建连续虚拟环境,具有不同版本的代理商互相交易。在此多路复用过程中,在504个风险数据集接受培训的代理商,使用近端政策优化的基本概念来通过调整他们的行动选择来改善自己的决策,以防止国家的不确定性。风险好奇程度的学习作为一个内在奖励功能,并且具有船长,信号在于,在行动和市场行为之间找到突出的关系。培训的代理基于好奇风险的风险稳步上逐步提高了动作质量。由代理人好奇地推动的自我学习规则推动了对环境产生高性能的行动的政策。给出了8个现实股票的实验,以验证自我审查规则的适当性和效率。拟议的系统取得了有希望的表现,使得更好的交易使用较少的交易,并且表现出最先进的基线。

著录项

获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号