Synthetic pre-training for neural-network interatomic potentials

被引:9
|
作者
Gardner, John L. A. [1 ]
Baker, Kathryn T. [1 ]
Deringer, Volker L. [1 ]
机构
[1] Univ Oxford, Dept Chem, Inorgan Chem Lab, Oxford OX1 3QR, England
来源
基金
英国科研创新办公室; 英国工程与自然科学研究理事会;
关键词
machine learning; neural networks; synthetic data; atomistic simulations; molecular dynamics;
D O I
10.1088/2632-2153/ad1626
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Machine learning (ML) based interatomic potentials have transformed the field of atomistic materials modelling. However, ML potentials depend critically on the quality and quantity of quantum-mechanical reference data with which they are trained, and therefore developing datasets and training pipelines is becoming an increasingly central challenge. Leveraging the idea of 'synthetic' (artificial) data that is common in other areas of ML research, we here show that synthetic atomistic data, themselves obtained at scale with an existing ML potential, constitute a useful pre-training task for neural-network (NN) interatomic potential models. Once pre-trained with a large synthetic dataset, these models can be fine-tuned on a much smaller, quantum-mechanical one, improving numerical accuracy and stability in computational practice. We demonstrate feasibility for a series of equivariant graph-NN potentials for carbon, and we carry out initial experiments to test the limits of the approach.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] SIMPLE-NN: An efficient package for training and executing neural-network interatomic potentials
    Lee, Kyuhyun
    Yoo, Dongsun
    Jeong, Wonseok
    Han, Seungwu
    COMPUTER PHYSICS COMMUNICATIONS, 2019, 242 : 95 - 103
  • [2] Synthetic Pre-Training Tasks for Neural Machine Translation
    He, Zexue
    Blackwood, Graeme
    Panda, Rameswar
    McAuley, Julian
    Feris, Rogerio
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 8080 - 8098
  • [3] Unsupervised Pre-training on Improving the Performance of Neural Network in Regression
    Salida, Pallabi
    Vij, Prateek
    Baruah, Rashmi Dutta
    2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,
  • [4] Pre-training Graph Neural Network for Cross Domain Recommendation
    Wang, Chen
    Liang, Yueqing
    Liu, Zhiwei
    Zhang, Tao
    Yu, Philip S.
    2021 IEEE THIRD INTERNATIONAL CONFERENCE ON COGNITIVE MACHINE INTELLIGENCE (COGMI 2021), 2021, : 140 - 145
  • [5] Pre-Training of an Artificial Neural Network for Software Fault Prediction
    Owhadi-Kareshk, Moein
    Sedaghat, Yasser
    Akbarzadeh-T, Mohammad-R
    PROCEEDINGS OF THE 2017 7TH INTERNATIONAL CONFERENCE ON COMPUTER AND KNOWLEDGE ENGINEERING (ICCKE), 2017, : 223 - 228
  • [6] Specialized Pre-Training of Neural Networks on Synthetic Data for Improving Paraphrase Generation
    Skurzhanskyi, O. H.
    Marchenko, O. O.
    Anisimov, A. V.
    CYBERNETICS AND SYSTEMS ANALYSIS, 2024, 60 (02) : 167 - 174
  • [7] Neural Grammatical Error Correction Systems with Unsupervised Pre-training on Synthetic Data
    Grundkiewicz, Roman
    Junczys-Dowmunt, Marcin
    Heafield, Kenneth
    INNOVATIVE USE OF NLP FOR BUILDING EDUCATIONAL APPLICATIONS, 2019, : 252 - 263
  • [8] Specialized Pre-Training of Neural Networks on Synthetic Data for Improving Paraphrase Generation
    O. H. Skurzhanskyi
    O. O. Marchenko
    A. V. Anisimov
    Cybernetics and Systems Analysis, 2024, 60 : 167 - 174
  • [9] GCC: Graph Contrastive Coding for Graph Neural Network Pre-Training
    Qiu, Jiezhong
    Chen, Qibin
    Dong, Yuxiao
    Zhang, Jing
    Yang, Hongxia
    Ding, Ming
    Wang, Kuansan
    Tang, Jie
    KDD '20: PROCEEDINGS OF THE 26TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2020, : 1150 - 1160
  • [10] Investigating of Disease Name Normalization Using Neural Network and Pre-Training
    Lou, Yinxia
    Qian, Tao
    Li, Fei
    Zhou, Junxiang
    Ji, Donghong
    Cheng, Ming
    IEEE ACCESS, 2020, 8 : 85729 - 85739