Statistical arbitrage trading on the intraday market using the asynchronous advantage actor-critic method

被引:10
|
作者
Demir, Sumeyra
Stappers, Bart
Kok, Koen
Paterakis, Nikolaos G.
机构
[1] Department of Electrical Engineering, Eindhoven University of Technology
[2] Scholt Energy, Valkensvard
关键词
Algorithmic trading; Actor-critic; A3C; Behaviour cloning; Deep reinforcement learning; Intraday markets; Non-physical virtual trader; Single intraday coupled market; Statistical arbitrage; ELECTRICITY; IMPACT;
D O I
10.1016/j.apenergy.2022.118912
中图分类号
TE [石油、天然气工业]; TK [能源与动力工程];
学科分类号
0807 ; 0820 ;
摘要
In this paper, we focus on statistical arbitrage trading opportunities involving the continuous exploitation of price differences arising during an intraday trading period with the option of closing positions on the balancing market. We aim to maximise the reward-risk ratio of an autonomous trading strategy. To find an optimal trading policy, we propose utilising the asynchronous advantage actor-critic (A3C) algorithm, a deep reinforcement learning method, with function approximators of two-headed shared deep neural networks. We enforce a risk-constrained trading strategy by limiting the maximum allowed position, and conduct state engineering and selection processes. We introduce a novel reward function and goal-based exploration, i.e. behaviour cloning. Our methodology is evaluated on a case study using the limit order book of the European single intraday coupled market (SIDC) available for the Dutch market area. The majority of hourly products on the test set return a profit. We expect our study to benefit electricity traders, renewable electricity producers and researchers who seek to implement state-of-art intelligent trading strategies.
引用
收藏
页数:10
相关论文
共 50 条
  • [41] Supervised actor-critic reinforcement learning with action feedback for algorithmic trading
    Qizhou Sun
    Yain-Whar Si
    Applied Intelligence, 2023, 53 : 16875 - 16892
  • [42] An Actor-Critic Method Using Least Squares Temporal Difference Learning
    Paschalidis, Ioannis Ch
    Li, Keyong
    Estanjini, Reza Moazzez
    PROCEEDINGS OF THE 48TH IEEE CONFERENCE ON DECISION AND CONTROL, 2009 HELD JOINTLY WITH THE 2009 28TH CHINESE CONTROL CONFERENCE (CDC/CCC 2009), 2009, : 2564 - 2569
  • [43] Active Queue Management in L4S with Asynchronous Advantage Actor-Critic: A FreeBSD Networking Stack Perspective
    Satish, Deol
    Kua, Jonathan
    Pokhrel, Shiva Raj
    FUTURE INTERNET, 2024, 16 (08)
  • [44] VMP-A3C: Virtual machines placement in cloud computing based on asynchronous advantage actor-critic algorithm
    Wei, Pengcheng
    Zeng, Yushan
    Yan, Bei
    Zhou, Jiahui
    Nikougoftar, Elaheh
    JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2023, 35 (05)
  • [45] A unified NDP method for MDPs by actor-critic networks
    Tang Hao
    Chen Dong
    Zhou Lei
    PROCEEDINGS OF THE 24TH CHINESE CONTROL CONFERENCE, VOLS 1 AND 2, 2005, : 1012 - 1016
  • [46] Merging with Extraction Method for Transfer Learning in Actor-Critic
    Takano, Toshiaki
    Takase, Haruhiko
    Kawanaka, Hiroharu
    Tsuruoka, Shinji
    JOURNAL OF ADVANCED COMPUTATIONAL INTELLIGENCE AND INTELLIGENT INFORMATICS, 2011, 15 (07) : 814 - 821
  • [47] Learning-Based Resource Allocation in Cloud Data Center Using Advantage Actor-Critic
    Chen, Zheyi
    Hu, Jia
    Min, Geyong
    ICC 2019 - 2019 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2019,
  • [48] Cooperative traffic signal control using Multi-step return and Off-policy Asynchronous Advantage Actor-Critic Graph algorithm
    Yang, Shantian
    Yang, Bo
    Wong, Hau-San
    Kang, Zhongfeng
    KNOWLEDGE-BASED SYSTEMS, 2019, 183
  • [49] An Actor-Critic Method for Simulation-Based Optimization
    Li, Kuo
    Jia, Qing-Shan
    Yan, Jiaqi
    IFAC PAPERSONLINE, 2022, 55 (11): : 7 - 12
  • [50] Swarm Reinforcement Learning Method Based on an Actor-Critic Method
    Iima, Hitoshi
    Kuroe, Yasuaki
    SIMULATED EVOLUTION AND LEARNING, 2010, 6457 : 279 - 288