Using deep Q-learning to understand the tax evasion behavior of risk-averse firms

被引:14
|
作者
Goumagias, Nikolaos D. [1 ]
Hristu-Varsakelis, Dimitrios [2 ]
Assael, Yannis M. [3 ]
机构
[1] Northumbria Univ, Newcastle Business Sch, Cent Campus East 1, Newcastle Upon Tyne NE1 8ST, Tyne & Wear, England
[2] Univ Macedonia, Dept Appl Informat, Egnatia 156, Thessaloniki 54006, Greece
[3] Univ Oxford, Dept Comp Sci, Wolfson Bldg,Parks Rd, Oxford OX1 3QD, England
关键词
Markov decision processes; Tax evasion; Q-learning; Deep learning; NEURAL-NETWORKS; AMNESTIES;
D O I
10.1016/j.eswa.2018.01.039
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Designing tax policies that are effective in curbing tax evasion and maximize state revenues requires a rigorous understanding of taxpayer behavior. This work explores the problem of determining the strategy a self-interested, risk-averse tax entity is expected to follow, as it "navigates" - in the context of a Markov Decision Process - a government-controlled tax environment that includes random audits, penalties and occasional tax amnesties. Although simplified versions of this problem have been previously explored, the mere assumption of risk-aversion (as opposed to risk-neutrality) raises the complexity of finding the optimal policy well beyond the reach of analytical techniques. Here, we obtain approximate solutions via a combination of Q-learning and recent advances in Deep Reinforcement Learning. By doing so, we (i) determine the tax evasion behavior expected of the taxpayer entity, (ii) calculate the degree of risk aversion of the "average" entity given empirical estimates of tax evasion, and (iii) evaluate sample tax policies, in terms of expected revenues. Our model can be useful as a testbed for "in-vitro" testing of tax policies, while our results lead to various policy recommendations. (C) 2018 Elsevier Ltd. All rights reserved.
引用
收藏
页码:258 / 270
页数:13
相关论文
共 50 条
  • [21] Collaborative Traffic Signal Automation Using Deep Q-Learning
    Hassan, Muhammad Ahmed
    Elhadef, Mourad
    Khan, Muhammad Usman Ghani
    [J]. IEEE ACCESS, 2023, 11 : 136015 - 136032
  • [22] Faster Deep Q-learning using Neural Episodic Control
    Nishio, Daichi
    Yamane, Satoshi
    [J]. 2018 IEEE 42ND ANNUAL COMPUTER SOFTWARE AND APPLICATIONS CONFERENCE (COMPSAC), VOL 1, 2018, : 486 - 491
  • [23] Deep Q-Learning using Redundant Outputs in Visual Doom
    Park, Hyunsoo
    Kim, Kyung-Joong
    [J]. 2016 IEEE CONFERENCE ON COMPUTATIONAL INTELLIGENCE AND GAMES (CIG), 2016,
  • [24] Air-Combat Strategy Using Deep Q-Learning
    Ma, Xiaoteng
    Xia, Li
    Zhao, Qianchuan
    [J]. 2018 CHINESE AUTOMATION CONGRESS (CAC), 2018, : 3952 - 3957
  • [25] Deep Convolution Q-Learning for emulation of human behavior patterns in gaming bots
    Kalra, Ishmeet Singh
    Pandita, Sarthak
    Soni, Sagar
    Alam, Md Kazmi
    [J]. 2019 IEEE PUNE SECTION INTERNATIONAL CONFERENCE (PUNECON), 2019,
  • [26] A scheduling scheme in the cloud computing environment using deep Q-learning
    Tong, Zhao
    Chen, Hongjian
    Deng, Xiaomei
    Li, Kenli
    Li, Keqin
    [J]. INFORMATION SCIENCES, 2020, 512 (512) : 1170 - 1191
  • [27] Adaptive-Precision Framework for SGD Using Deep Q-Learning
    Zhang, Wentai
    Huang, Hanxian
    Zhang, Jiaxi
    Jiang, Ming
    Luo, Guojie
    [J]. 2018 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER-AIDED DESIGN (ICCAD) DIGEST OF TECHNICAL PAPERS, 2018,
  • [28] An intelligent financial portfolio trading strategy using deep Q-learning
    Park, Hyungjun
    Sim, Min Kyu
    Choi, Dong Gu
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2020, 158 (158)
  • [29] Path Planning Using Wasserstein Distributionally Robust Deep Q-learning
    Alpturk, Cem
    Renganathan, Venkatraman
    [J]. 2023 EUROPEAN CONTROL CONFERENCE, ECC, 2023,
  • [30] Battlefield Situation Deduction and Maneuver Decision Using Deep Q-Learning
    Shi, Minghui
    Dong, Xiwang
    Han, Liang
    Li, Qingdong
    Ren, Zhang
    [J]. 2021 PROCEEDINGS OF THE 40TH CHINESE CONTROL CONFERENCE (CCC), 2021, : 3651 - 3656