Goal-directed molecule generation with fine-tuning by policy gradient

被引:2
|
作者
Sha, Chunli [1 ]
Zhu, Fei [1 ]
机构
[1] Soochow Univ, Sch Comp Sci & Technol, Suzhou 215006, Peoples R China
基金
中国国家自然科学基金;
关键词
Drug design; Graph neural network; Reinforcement learning; Policy gradient; Molecule generation; NETWORK; DESIGN;
D O I
10.1016/j.eswa.2023.123127
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Graph -structured drug molecule representations often struggle to generate molecules by particular intentions, which results in generated molecules without pharmacological properties. To address the problem, we propose a de novo molecular generation method that utilizes the policy gradient algorithm of reinforcement learning to fine-tune the molecular graph generation model. The training process of the method is divided into the pre -training stage and the fine-tuning stage. During the pre -training stage, it uses graph neural networks and multilayer perceptrons to train a molecule graph generation model. During the fine-tuning stage, scoring functions are devised for multiple goal -directed generation tasks, and subsequently, the policy loss function is formulated based on the reward shaping mechanism. A value network is designed to calculate the value of taking an action based on the current graph state during agent sampling to guide policy updates. To mitigate the issue of molecular uniqueness decline during the learning process, we dynamically adjust the weights of the two learning processes in the policy loss function, aiming to generate desirable molecules with high probability and reduce the descent of uniqueness. The experiments showed that after fine-tuning, the generative model exhibits a higher probability of generating molecules with desired properties compared to other models. Furthermore, our method effectively mitigates the issue of molecular uniqueness declining during the learning process when compared to alternative fine-tuning methods.
引用
收藏
页数:12
相关论文
共 50 条
  • [31] Robustness Against Gradient based Attacks through Cost Effective Network Fine-Tuning
    Agarwal, Akshay
    Ratha, Nalini
    Singh, Richa
    Vatsa, Mayank
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW, 2023, : 28 - 37
  • [32] Fine-Tuning a Large Language Model with Reinforcement Learning for Educational Question Generation
    Lamsiyah, Salima
    El Mahdaouy, Abdelkader
    Nourbakhsh, Aria
    Schommer, Christoph
    ARTIFICIAL INTELLIGENCE IN EDUCATION, PT I, AIED 2024, 2024, 14829 : 424 - 438
  • [33] PAC-tuning: Fine-tuning Pretrained Language Models with PAC-driven Perturbed Gradient Descent
    Liu, Guangliang
    Xue, Zhiyu
    Zhang, Xitong
    Johnson, Kristen Marie
    Wang, Rongrong
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 12178 - 12189
  • [34] Atomic Design and Fine-Tuning of Subnanometric Pt Catalysts to Tame Hydrogen Generation
    Yang, Jie
    Fu, Wenzhao
    Chen, Chaoqiu
    Chen, Wenyao
    Huang, Wugen
    Yang, Ruoou
    Kong, Qingqiang
    Zhang, Baiyan
    Zhao, Jixiao
    Chen, Chengmeng
    Luo, Jun
    Yang, Fan
    Duan, Xuezhi
    Jiang, Zheng
    Qin, Yong
    ACS CATALYSIS, 2021, 11 (07) : 4146 - 4156
  • [35] Bridging the Gap between Pre-Training and Fine-Tuning for Commonsense Generation
    Yang, Haoran
    Wang, Yan
    Li, Piji
    Bi, Wei
    Lam, Wai
    Xu, Chen
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 376 - 383
  • [36] Knowledge Enhanced Fine-Tuning for Better Handling Unseen Entities in Dialogue Generation
    Cui, Leyang
    Wu, Yu
    Liu, Shujie
    Zhang, Yue
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 2328 - 2337
  • [37] Tuning of olfactory cortex ventral tenia tecta neurons to distinct task elements of goal-directed behavior
    Shiotani, Kazuki
    Tanisumi, Yuta
    Murata, Koshi
    Hirokawa, Junya
    Sakurai, Yoshio
    Manabe, Hiroyuki
    ELIFE, 2020, 9
  • [38] Short Answer Questions Generation by Fine-Tuning BERT and GPT-2
    Tsai, Danny C. L.
    Chang, Willy J. W.
    Yang, Stephen J. H.
    29TH INTERNATIONAL CONFERENCE ON COMPUTERS IN EDUCATION (ICCE 2021), VOL II, 2021, : 508 - 514
  • [39] Extreme Risk Averse Policy for Goal-Directed Risk-Sensitive Markov Decision Process
    Freire, Valdinei
    Delgado, Karina Valdivia
    PROCEEDINGS OF 2016 5TH BRAZILIAN CONFERENCE ON INTELLIGENT SYSTEMS (BRACIS 2016), 2016, : 79 - 84
  • [40] Diverse Hits in De Novo Molecule Design: Diversity-Based Comparison of Goal-Directed Generators
    Renz, Philipp
    Luukkonen, Sohvi
    Klambauer, Guenter
    JOURNAL OF CHEMICAL INFORMATION AND MODELING, 2024, 64 (15) : 5756 - 5761