CrysGNN: Distilling Pre-trained Knowledge to Enhance Property Prediction for Crystalline Materials

被引:0
|
作者
Das, Kishalay [1 ]
Samanta, Bidisha [1 ]
Goyal, Pawan [1 ]
Lee, Seung-Cheol [2 ]
Bhattacharjee, Satadeep [2 ]
Ganguly, Niloy [1 ,3 ]
机构
[1] Indian Inst Technol Kharagpur, Kharagpur, India
[2] Indo Korea Sci & Technol Ctr, Bangalore, Karnataka, India
[3] Leibniz Univ Hannover, L3S, Hannover, Germany
关键词
NETWORKS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, graph neural network (GNN) based approaches have emerged as a powerful technique to encode complex topological structure of crystal materials in an enriched representation space. These models are often supervised in nature and using the property-specific training data, learn relationship between crystal structure and different properties like formation energy, bandgap, bulk modulus, etc. Most of these methods require a huge amount of property-tagged data to train the system which may not be available for different properties. However, there is an availability of a huge amount of crystal data with its chemical composition and structural bonds. To leverage these untapped data, this paper presents CrysGNN, a new pre-trained GNN framework for crystalline materials, which captures both node and graph level structural information of crystal graphs using a huge amount of unlabelled material data. Further, we extract distilled knowledge from CrysGNN and inject into different state of the art property predictors to enhance their property prediction accuracy. We conduct extensive experiments to show that with distilled knowledge from the pre-trained model, all the SOTA algorithms are able to outperform their own vanilla version with good margins. We also observe that the distillation process provides a significant improvement over the conventional approach of finetuning the pre-trained model. We have re-leased the pre-trained model along with the large dataset of 800K crystal graph which we carefully curated; so that the pretrained model can be plugged into any existing and upcoming models to enhance their prediction accuracy.
引用
收藏
页码:7323 / 7331
页数:9
相关论文
共 50 条
  • [1] Using Noise and External Knowledge to Enhance Chinese Pre-trained Model
    Ma, Haoyang
    Li, Zeyu
    Guo, Hongyu
    2022 IEEE 34TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, ICTAI, 2022, : 476 - 480
  • [2] TinyMIM: An Empirical Study of Distilling MIM Pre-trained Models
    Ren, Sucheng
    Wei, Fangyun
    Zhang, Zhang
    Hu, Han
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 3687 - 3697
  • [3] Distilling Relation Embeddings from Pre-trained Language Models
    Ushio, Asahi
    Camacho-Collados, Jose
    Schockaert, Steven
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 9044 - 9062
  • [4] NewsBERT: Distilling Pre-trained Language Model for Intelligent News Application
    Wu, Chuhan
    Wu, Fangzhao
    Yu, Yang
    Qi, Tao
    Huang, Yongfeng
    Liu, Qi
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 3285 - 3295
  • [5] A pre-trained multi-representation fusion network for molecular property prediction
    Zhang, Haohui
    Wu, Juntong
    Liu, Shichao
    Han, Shen
    INFORMATION FUSION, 2024, 103
  • [6] GenDistiller: Distilling Pre-trained Language Models based on an Autoregressive Generative Model
    Gao, Yingying
    Zhang, Shilei
    Deng, Chao
    Feng, Junlan
    INTERSPEECH 2024, 2024, : 3325 - 3329
  • [7] Knowledge Rumination for Pre-trained Language Models
    Yao, Yunzhi
    Wang, Peng
    Mao, Shengyu
    Tan, Chuanqi
    Huang, Fei
    Chen, Huajun
    Zhang, Ningyu
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 3387 - 3404
  • [8] Knowledge Inheritance for Pre-trained Language Models
    Qin, Yujia
    Lin, Yankai
    Yi, Jing
    Zhang, Jiajie
    Han, Xu
    Zhang, Zhengyan
    Su, Yusheng
    Liu, Zhiyuan
    Li, Peng
    Sun, Maosong
    Zhou, Jie
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 3921 - 3937
  • [9] Toxicity Prediction Using Pre-trained Autoencoder
    Galushka, Mykola
    Browne, Fiona
    Mulvenna, Maurice
    Bond, Raymond
    Lightbody, Gaye
    PROCEEDINGS 2018 IEEE INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND BIOMEDICINE (BIBM), 2018, : 299 - 304
  • [10] CollRec: Pre-Trained Language Models and Knowledge Graphs Collaborate to Enhance Conversational Recommendation System
    Liu, Shuang
    Ao, Zhizhuo
    Chen, Peng
    Kolmanic, Simon
    IEEE ACCESS, 2024, 12 : 104663 - 104675