An accurate and interpretable deep learning model for environmental properties prediction using hybrid molecular representations

被引:25
|
作者
Zhang, Jun [1 ]
Wang, Qin [2 ]
Su, Yang [3 ]
Jin, Saimeng [1 ]
Ren, Jingzheng [4 ]
Eden, Mario [5 ]
Shen, Weifeng [1 ]
机构
[1] Chongqing Univ, Sch Chem & Chem Engn, Chongqing 400044, Peoples R China
[2] Chongqing Univ Sci & Technol, Sch Chem & Chem Engn, Chongqing, Peoples R China
[3] Chongqing Univ Sci & Technol, Sch Intelligent Technol & Engn, Chongqing, Peoples R China
[4] Hong Kong Polytech Univ, Dept Ind & Syst Engn, Hong Kong, Peoples R China
[5] Auburn Univ, Dept Chem Engn, Auburn, AL 36849 USA
基金
中国国家自然科学基金;
关键词
deep learning network; interpretability; lipophilicity; message-passing neural network; QSPR; WATER PARTITION-COEFFICIENTS; OCTANOL-WATER; EXTRACTIVE DISTILLATION; IONIC LIQUIDS; GREEN CHEMISTRY; LIPOPHILICITY; DESIGN; APPLICABILITY;
D O I
10.1002/aic.17634
中图分类号
TQ [化学工业];
学科分类号
0817 ;
摘要
Lipophilicity, as quantified by the decimal logarithm of the octanol-water partition coefficient (log K-OW), is an essential environmental property. Deep neural networks (DNNs) based quantitative structure-property relationship (QSPR) studies have received more and more attention because of their excellent performance for prediction. However, the black-box nature of DNNs limits the application range where interpretability is essential. Hence, this study aims to develop an accurate and interpretable deep neural network (AI-DNN) model for log K-OW prediction. A hybrid method of molecular representation was employed to guarantee the accuracy of the proposed AI-DNN model. The hybrid molecular representations are able to integrate the directed message passing neural networks (D-MPNNs) learned molecular representations and the fixed molecule-level features of CDK descriptors, and can capture both the local and the global features of overall molecule. The performance analysis shows that the proposed QSPR model exhibits promising predictive accuracy and discriminative power in the structural isomers and stereoisomers. Moreover, the Monte Carlo Tree Search (MCTS) approach was used to interpret the proposed AI-DNN model by identifying the molecular substructures contributed to the lipophilicity. This interpretability can be applied to critical fields where there is a high demand for interpretable deep networks, such as green solvent design and drug discovery.
引用
收藏
页数:13
相关论文
共 50 条
  • [31] Interpretable Deep Learning for Probabilistic MJO Prediction
    Delaunay, Antoine
    Christensen, Hannah M.
    [J]. GEOPHYSICAL RESEARCH LETTERS, 2022, 49 (16)
  • [32] Interpretable molecular encodings and representations for machine learning tasks
    Weckbecker, Moritz
    Anzela, Aleksandar
    Yang, Zewen
    Hattab, Georgesm
    [J]. COMPUTATIONAL AND STRUCTURAL BIOTECHNOLOGY JOURNAL, 2024, 23 : 2326 - 2336
  • [33] A Novel Hybrid Deep Learning Method for Accurate Exchange Rate Prediction
    Iqbal, Farhat
    Koutmos, Dimitrios
    Ahmed, Eman A.
    Al-Essa, Lulwah M.
    [J]. RISKS, 2024, 12 (09)
  • [34] Hybrid deep learning and evolutionary algorithms for accurate cloud workload prediction
    Ali, Tassawar
    Khan, Hikmat Ullah
    Alarfaj, Fawaz Khaled
    Alreshoodi, Mohammed
    [J]. COMPUTING, 2024, : 3905 - 3944
  • [35] An accurate pediatric bone age prediction model using deep learning and contrast conversion
    Choi, Dong Hyeok
    Ahn, So Hyun
    Lee, Rena
    [J]. EWHA MEDICAL JOURNAL, 2024, 47 (02):
  • [36] Learning interpretable representations of entanglement in quantum optics experiments using deep generative models
    Flam-Shepherd, Daniel
    Wu, Tony C.
    Gu, Xuemei
    Cervera-Lierta, Alba
    Krenn, Mario
    Aspuru-Guzik, Alan
    [J]. NATURE MACHINE INTELLIGENCE, 2022, 4 (06) : 544 - 554
  • [37] Cyanobacteria cell prediction using interpretable deep learning model with observed, numerical, and sensing data assemblage
    Pyo, JongCheol
    Cho, Kyung Hwa
    Kim, Kyunghyun
    Baek, Sang-Soo
    Nam, Gibeom
    Park, Sanghyun
    [J]. WATER RESEARCH, 2021, 203
  • [38] Learning interpretable representations of entanglement in quantum optics experiments using deep generative models
    Daniel Flam-Shepherd
    Tony C. Wu
    Xuemei Gu
    Alba Cervera-Lierta
    Mario Krenn
    Alán Aspuru-Guzik
    [J]. Nature Machine Intelligence, 2022, 4 : 544 - 554
  • [39] A novel deep learning prediction model for concrete dam displacements using interpretable mixed attention mechanism
    Ren, Qiubing
    Li, Mingchao
    Li, Heng
    Shen, Yang
    [J]. ADVANCED ENGINEERING INFORMATICS, 2021, 50
  • [40] An interpretable hybrid deep learning model for flood forecasting based on Transformer and LSTM
    Li, Wenzhong
    Liu, Chengshuai
    Xu, Yingying
    Niu, Chaojie
    Li, Runxi
    Li, Ming
    Hu, Caihong
    Tian, Lu
    [J]. JOURNAL OF HYDROLOGY-REGIONAL STUDIES, 2024, 54