Q-Learning model for selfish miners with optional stopping theorem for honest miners

被引:0
|
作者
Rakkini, M. J. Jeyasheela [1 ]
Geetha, K. [1 ]
机构
[1] SASTRA Deemed Univ, Sch Comp, Tiruchirappalli 620014, India
关键词
difficulty adjustment algorithms; gambler ruin; honest mining; prediction; reinforcement learning; selfish mining;
D O I
10.1111/itor.13359
中图分类号
C93 [管理学];
学科分类号
12 ; 1201 ; 1202 ; 120202 ;
摘要
Bitcoin, the most popular cryptocurrency used in the blockchain, has miners join mining pools and get rewarded for the proportion of hash rate they have contributed to the mining pool. This work proposes the prediction of the relativegain of the miners by machine learning and deep learning models, the miners' selection of higher relativegain by the Q-learning model, and an optional stopping theorem for honest miners in the presence of selfish mining attacks. Relativegain is the ratio of the number of blocks mined by selfish miners in the main canonical chain to the blocks of other miners. A Q-learning agent with & epsilon;-greedy value iteration, which seeks to increase the relativegain for the selfish miners, that takes into account all the other quintessential parameters, including the hash rate of miners, time warp, the height of the blockchain, the number of times the blockchain was reorganized, and the adjustment of the timestamp of the block, is implemented. Next, the ruin of the honest miners and the optional stopping theorem are analyzed so that the honest miners can quit the mining process before their complete ruin. We obtain a low mean square error of 0.0032 and a mean absolute error of 0.0464 in our deep learning model. Our Q-learning model exhibits a linearly increasing curve, which denotes the increase in the relativegain caused by the selection of the action of performing the reorganization attack.
引用
收藏
页码:3975 / 3998
页数:24
相关论文
共 50 条
  • [41] A Q-learning agent-based model for the analysis of the power market dynamics
    Tellidou, Athina
    Bakirtzis, Anastasios
    PROCEEDINGS OF THE SIXTH IASTED INTERNATIONAL CONFERENCE ON EUROPEAN POWER AND ENERGY SYSTEMS, 2006, : 228 - +
  • [42] The economics analysis of a Q-learning model of cooperation with punishment and risk taking preferences
    Nazaria Solferino
    Viviana Solferino
    Serena F. Taurino
    Journal of Economic Interaction and Coordination, 2018, 13 : 601 - 613
  • [43] Combination Optimization Model of Urban Key Intersections Based on Q-Learning Algorithm
    Dong, Dan-Ping
    Wei, Fu-Lu
    Chen, Ming-Tao
    Guo, Yong-Qing
    Yang, Chang-Hai
    Han, Yu-Xin
    CICTP 2023: INNOVATION-EMPOWERED TECHNOLOGY FOR SUSTAINABLE, INTELLIGENT, DECARBONIZED, AND CONNECTED TRANSPORTATION, 2023, : 849 - 859
  • [44] Hyperparameter Optimization for the LSTM Method of AUV Model Identification Based on Q-Learning
    Wang, Dianrui
    Wan, Junhe
    Shen, Yue
    Qin, Ping
    He, Bo
    JOURNAL OF MARINE SCIENCE AND ENGINEERING, 2022, 10 (08)
  • [45] Adaptive and Coordinated Traffic Signal Control Based on Q-Learning and MULTIBAND Model
    Lu, Shoufeng
    Liu, Ximin
    Dai, Shiqiang
    2008 IEEE CONFERENCE ON CYBERNETICS AND INTELLIGENT SYSTEMS, VOLS 1 AND 2, 2008, : 446 - +
  • [46] A Double Deep Q-Learning Model for Energy-Efficient Edge Scheduling
    Zhang, Qingchen
    Lin, Man
    Yang, Laurence T.
    Chen, Zhikui
    Khan, Samee U.
    Li, Peng
    IEEE TRANSACTIONS ON SERVICES COMPUTING, 2019, 12 (05) : 739 - 749
  • [47] Simulation Model for the AGC System of Isolated Microgrid Based on Q-learning Method
    Wang, Penghu
    Tang, Hao
    Lv, Kai
    PROCEEDINGS OF 2018 IEEE 7TH DATA DRIVEN CONTROL AND LEARNING SYSTEMS CONFERENCE (DDCLS), 2018, : 1213 - 1217
  • [48] QT-TDM: Planning With Transformer Dynamics Model and Autoregressive Q-Learning
    Kotb, Mostafa
    Weber, Cornelius
    Hafez, Muhammad Burhan
    Wermter, Stefan
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2025, 10 (01): : 112 - 119
  • [49] A multi-stage group decision model based on improved Q-learning
    Zhang F.
    Liu L.-Y.
    Guo X.-X.
    Kongzhi yu Juece/Control and Decision, 2019, 34 (09): : 1917 - 1922
  • [50] Switching-model Cooperative Control for Two Intersections Applying Q-learning
    Zhang, Jie
    Zhao, Xiao-hua
    Yu, Quan
    Li, Zhen-long
    2008 7TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-23, 2008, : 82 - 85