Ordering-Based Causal Discovery with Reinforcement Learning

被引:0
|
作者
Wang, Xiaoqiang [1 ]
Du, Yali [2 ]
Zhu, Shengyu [3 ]
Ke, Liangjun [1 ]
Chen, Zhitang [3 ]
Hao, Jianye [3 ,4 ]
Wang, Jun [2 ]
机构
[1] Xi An Jiao Tong Univ, Sch Automat Sci & Engn, State Key Lab Mfg Syst Engn, Xian, Peoples R China
[2] UCL, London, England
[3] Huawei Noahs Ark Lab, Quebec City, PQ, Canada
[4] Tianjin Univ, Coll Intelligence & Comp, Tianjin, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
It is a long-standing question to discover causal relations among a set of variables in many empirical sciences. Recently, Reinforcement Learning (RL) has achieved promising results in causal discovery from observational data. However, searching the space of directed graphs and enforcing acyclicity by implicit penalties tend to be inefficient and restrict the existing RL-based method to small scale problems. In this work, we propose a novel RL-based approach for causal discovery, by incorporating RL into the ordering-based paradigm. Specifically, we formulate the ordering search problem as a multi-step Markov decision process, implement the ordering generating process with an encoder-decoder architecture, and finally use RL to optimize the proposed model based on the reward mechanisms designed for each ordering. A generated ordering would then be processed using variable selection to obtain the final causal graph. We analyze the consistency and computational complexity of the proposed method, and empirically show that a pretrained model can be exploited to accelerate training. Experimental results on both synthetic and real data sets shows that the proposed method achieves a much improved performance over existing RL-based method.
引用
收藏
页码:3566 / 3573
页数:8
相关论文
共 50 条
  • [31] Causal discovery approach with reinforcement learning for risk factors of type II diabetes mellitus
    Xiu-E. Gao
    Jian-Gang Hu
    Bo Chen
    Yun-Ming Wang
    Sheng-Bin zhou
    BMC Bioinformatics, 24
  • [32] An efficient ordering-based ensemble pruning algorithm via dynamic programming
    Qun Dai
    Xiaomeng Han
    Applied Intelligence, 2016, 44 : 816 - 830
  • [33] An efficient ordering-based ensemble pruning algorithm via dynamic programming
    Dai, Qun
    Han, Xiaomeng
    APPLIED INTELLIGENCE, 2016, 44 (04) : 816 - 830
  • [34] Enhanced causal effects estimation based on offline reinforcement learning
    Xia, Huan
    Jiang, Chaozhe
    Zhang, Chenyang
    APPLIED INTELLIGENCE, 2025, 55 (04)
  • [35] LINKORD: link ordering-based data gathering protocol for wireless sensor networks
    Marjan Radi
    Behnam Dezfouli
    Kamalrulnizam Abu Bakar
    Shukor Abd Razak
    Malrey Lee
    Computing, 2015, 97 : 205 - 236
  • [36] STREAMS: Towards Spatio-Temporal Causal Discovery with Reinforcement Learning for Streamflow Rate Prediction
    Sheth, Paras
    Mosallanezhad, Ahmadreza
    Ding, Kaize
    Shah, Reepal
    Sabo, John
    Liu, Huan
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 4815 - 4821
  • [37] LINKORD: link ordering-based data gathering protocol for wireless sensor networks
    Radi, Marjan
    Dezfouli, Behnam
    Abu Bakar, Kamalrulnizam
    Abd Razak, Shukor
    Lee, Malrey
    COMPUTING, 2015, 97 (03) : 205 - 236
  • [38] Online food ordering delivery strategies based on deep reinforcement learning
    Guangyu Zou
    Jiafu Tang
    Levent Yilmaz
    Xiangyu Kong
    Applied Intelligence, 2022, 52 : 6853 - 6865
  • [39] Reinforcement Learning Based Query Vertex Ordering Model for Subgraph Matching
    Wang, Hanchen
    Zhang, Ying
    Qin, Lu
    Wang, Wei
    Zhang, Wenjie
    Lin, Xuemin
    2022 IEEE 38TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE 2022), 2022, : 245 - 258
  • [40] Online food ordering delivery strategies based on deep reinforcement learning
    Zou, Guangyu
    Tang, Jiafu
    Yilmaz, Levent
    Kong, Xiangyu
    APPLIED INTELLIGENCE, 2022, 52 (06) : 6853 - 6865