Q-Learning model for selfish miners with optional stopping theorem for honest miners

被引:0
|
作者
Rakkini, M. J. Jeyasheela [1 ]
Geetha, K. [1 ]
机构
[1] SASTRA Deemed Univ, Sch Comp, Tiruchirappalli 620014, India
关键词
difficulty adjustment algorithms; gambler ruin; honest mining; prediction; reinforcement learning; selfish mining;
D O I
10.1111/itor.13359
中图分类号
C93 [管理学];
学科分类号
12 ; 1201 ; 1202 ; 120202 ;
摘要
Bitcoin, the most popular cryptocurrency used in the blockchain, has miners join mining pools and get rewarded for the proportion of hash rate they have contributed to the mining pool. This work proposes the prediction of the relativegain of the miners by machine learning and deep learning models, the miners' selection of higher relativegain by the Q-learning model, and an optional stopping theorem for honest miners in the presence of selfish mining attacks. Relativegain is the ratio of the number of blocks mined by selfish miners in the main canonical chain to the blocks of other miners. A Q-learning agent with & epsilon;-greedy value iteration, which seeks to increase the relativegain for the selfish miners, that takes into account all the other quintessential parameters, including the hash rate of miners, time warp, the height of the blockchain, the number of times the blockchain was reorganized, and the adjustment of the timestamp of the block, is implemented. Next, the ruin of the honest miners and the optional stopping theorem are analyzed so that the honest miners can quit the mining process before their complete ruin. We obtain a low mean square error of 0.0032 and a mean absolute error of 0.0464 in our deep learning model. Our Q-learning model exhibits a linearly increasing curve, which denotes the increase in the relativegain caused by the selection of the action of performing the reorganization attack.
引用
收藏
页码:3975 / 3998
页数:24
相关论文
共 50 条
  • [1] One Weird Trick to Stop Selfish Miners: Fresh Bitcoins, A Solution for the Honest Miner
    Heilman, Ethan
    FINANCIAL CRYPTOGRAPHY AND DATA SECURITY: FC 2014 WORKSHOPS, BITCOIN AND WAHC 2014, 2014, 8438 : 161 - 162
  • [2] Zap Q-Learning for Optimal Stopping
    Chen, Shuhang
    Devraj, Adithya M.
    Busic, Ana
    Meyn, Sean
    2020 AMERICAN CONTROL CONFERENCE (ACC), 2020, : 3920 - 3925
  • [3] An Accurate Analytical Model for A Proof-of-Work Blockchain with Multiple Selfish Miners
    Wang, Sheng-Wei
    Tzeng, Show-Shiow
    ICC 2024 - IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2024, : 1053 - 1059
  • [4] Interactive model building for Q-learning
    Laber, Eric B.
    Linn, Kristin A.
    Stefanski, Leonard A.
    BIOMETRIKA, 2014, 101 (04) : 831 - 847
  • [5] Deep learning classification of bitcoin miners and exploration of upper confidence bound algorithm with less regret for the selection of honest mining
    Rakkini, M. J. Jeyasheela
    Geetha, K.
    JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2021, 14 (6) : 6545 - 6561
  • [6] Deep learning classification of bitcoin miners and exploration of upper confidence bound algorithm with less regret for the selection of honest mining
    M. J. Jeyasheela Rakkini
    K. Geetha
    Journal of Ambient Intelligence and Humanized Computing, 2023, 14 : 6545 - 6561
  • [7] Q-learning agents in a Cournot oligopoly model
    Waltman, Ludo
    Kaymak, Uzay
    JOURNAL OF ECONOMIC DYNAMICS & CONTROL, 2008, 32 (10): : 3275 - 3293
  • [8] Information Theoretic Model Predictive Q-Learning
    Bhardwaj, Mohak
    Handa, Ankur
    Fox, Dieter
    Boots, Byron
    LEARNING FOR DYNAMICS AND CONTROL, VOL 120, 2020, 120 : 840 - 850
  • [9] A New Learning Model for Swarm Intelligence Based on Q-Learning
    Li, Fuming
    He, Xiaoxian
    Xu, Jingjing
    2010 8TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION (WCICA), 2010, : 2769 - 2775
  • [10] Model based path planning using Q-Learning
    Sharma, Avinash
    Gupta, Kanika
    Kumar, Anirudha
    Sharma, Aishwarya
    Kumar, Rajesh
    2017 IEEE INTERNATIONAL CONFERENCE ON INDUSTRIAL TECHNOLOGY (ICIT), 2017, : 837 - 842