R-DDQN: Optimizing Algorithmic Trading Strategies Using a Reward Network in a Double DQN

被引:0
|
作者
Zhou, Chujin [1 ]
Huang, Yuling [1 ]
Cui, Kai [1 ]
Lu, Xiaoping [1 ]
机构
[1] Macau Univ Sci & Technol, Sch Comp Sci & Engn, Taipa, Macao, Peoples R China
关键词
reinforcement learning; algorithmic trading; reward network; deep learning;
D O I
10.3390/math12111621
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Algorithmic trading is playing an increasingly important role in the financial market, achieving more efficient trading strategies by replacing human decision-making. Among numerous trading algorithms, deep reinforcement learning is gradually replacing traditional high-frequency trading strategies and has become a mainstream research direction in the field of algorithmic trading. This paper introduces a novel approach that leverages reinforcement learning with human feedback (RLHF) within the double DQN algorithm. Traditional reward functions in algorithmic trading heavily rely on expert knowledge, posing challenges in their design and implementation. To tackle this, the reward-driven double DQN (R-DDQN) algorithm is proposed, integrating human feedback via a reward function network trained on expert demonstrations. Additionally, a classification-based training method is employed for optimizing the reward function network. The experiments, conducted on datasets including HSI, IXIC, SP500, GOOGL, MSFT, and INTC, show that the proposed method outperforms all baselines across six datasets and achieves a maximum cumulative return of 1502% within 24 months.
引用
收藏
页数:22
相关论文
共 17 条
  • [1] Trading Bot Implementation and Performance Comparison Using DQN and DDQN
    Kim M.T.
    Kim B.W.
    [J]. Transactions of the Korean Institute of Electrical Engineers, 2021, 70 (01): : 158 - 167
  • [2] Optimizing Algorithmic Strategies for Trading Bitcoin
    Cohen, Gil
    [J]. COMPUTATIONAL ECONOMICS, 2021, 57 (02) : 639 - 654
  • [3] Optimizing Algorithmic Strategies for Trading Bitcoin
    Gil Cohen
    [J]. Computational Economics, 2021, 57 : 639 - 654
  • [4] Algorithmic Trading Behavior Identification using Reward Learning Method
    Yang, Steve Y.
    Qiao, Qifeng
    Beling, Peter A.
    Scherer, William T.
    [J]. PROCEEDINGS OF THE 2014 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2014, : 3807 - 3814
  • [5] Options Trading using Artificial Neural Network and Algorithmic Trading
    Ghosh, Sayandeep
    Kumar, Sudhanshu
    Deshmukh, Atharva
    Kurve, Akshay
    Welekar, Rashmi
    [J]. INTERNATIONAL JOURNAL OF NEXT-GENERATION COMPUTING, 2022, 13 (05): : 1286 - 1293
  • [6] Algorithmic trading using machine learning and neural network
    Agarwal D.
    Sheth R.
    Shekokar N.
    [J]. Lecture Notes on Data Engineering and Communications Technologies, 2021, 66 : 407 - 421
  • [7] Optimization of Cryptocurrency Algorithmic Trading Strategies Using the Decomposition Approach
    Omran, Sherin M.
    El-Behaidy, Wessam H.
    Youssif, Aliaa A. A.
    [J]. BIG DATA AND COGNITIVE COMPUTING, 2023, 7 (04)
  • [8] Robust technical trading strategies using GP for algorithmic portfolio selection
    Berutich, Jose Manuel
    Lopez, Francisco
    Luna, Francisco
    Quintana, David
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2016, 46 : 307 - 315
  • [9] Multi-peak Algorithmic Trading Strategies Using Grey Wolf Optimizer
    Mazumdar, Kingshuk
    Zhang, Dongmo
    Guo, Yi
    [J]. PRICAI 2019: TRENDS IN ARTIFICIAL INTELLIGENCE, PT III, 2019, 11672 : 748 - 754
  • [10] Algorithmic Trading Using Double Deep Q-Networks and Sentiment Analysis
    Tabaro, Leon
    Kinani, Jean Marie Vianney
    Rosales-Silva, Alberto Jorge
    Salgado-Ramirez, Julio Cesar
    Mujica-Vargas, Dante
    Escamilla-Ambrosio, Ponciano Jorge
    Ramos-Diaz, Eduardo
    [J]. INFORMATION, 2024, 15 (08)