Scalable training of graph convolutional neural networks for fast and accurate predictions of HOMO-LUMO gap in molecules

被引:10
|
作者
Choi, Jong Youl [1 ]
Zhang, Pei [2 ]
Mehta, Kshitij [1 ]
Blanchard, Andrew [2 ]
Pasini, Massimiliano Lupo [2 ]
机构
[1] Oak Ridge Natl Lab, Comp Sci & Math Div, 1 Bethel Valley Rd, Oak Ridge, TN 37831 USA
[2] Oak Ridge Natl Lab, Computat Sci & Engn Div, 1 Bethel Valley Rd, Oak Ridge, TN 37831 USA
关键词
Graph neural networks; Distributed data parallelism; Surrogate models; Atomic modeling; Molecular dynamics; HOMO-LUMO gap;
D O I
10.1186/s13321-022-00652-1
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Graph Convolutional Neural Network (GCNN) is a popular class of deep learning (DL) models in material science to predict material properties from the graph representation of molecular structures. Training an accurate and comprehensive GCNN surrogate for molecular design requires large-scale graph datasets and is usually a time-consuming process. Recent advances in GPUs and distributed computing open a path to reduce the computational cost for GCNN training effectively. However, efficient utilization of high performance computing (HPC) resources for training requires simultaneously optimizing large-scale data management and scalable stochastic batched optimization techniques. In this work, we focus on building GCNN models on HPC systems to predict material properties of millions of molecules. We use HydraGNN, our in-house library for large-scale GCNN training, leveraging distributed data parallelism in PyTorch. We use ADIOS, a high-performance data management framework for efficient storage and reading of large molecular graph data. We perform parallel training on two open-source large-scale graph datasets to build a GCNN predictor for an important quantum property known as the HOMO-LUMO gap. We measure the scalability, accuracy, and convergence of our approach on two DOE supercomputers: the Summit supercomputer at the Oak Ridge Leadership Computing Facility (OLCF) and the Perlmutter system at the National Energy Research Scientific Computing Center (NERSC). We present our experimental results with HydraGNN showing (i) reduction of data loading time up to 4.2 times compared with a conventional method and (ii) linear scaling performance for training up to 1024 GPUs on both Summit and Perlmutter.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] DistGNN: Scalable Distributed Training for Large -Scale Graph Neural Networks
    Md, Vasimuddin
    Misra, Sanchit
    Ma, Guixiang
    Mohanty, Ramanarayan
    Georganas, Evangelos
    Heinecke, Alexander
    Kalamkar, Dhiraj
    Ahmed, Nesreen K.
    Avancha, Sasikanth
    SC21: INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS, 2021,
  • [22] Training Matters: Unlocking Potentials of Deeper Graph Convolutional Neural Networks
    Luan, Sitao
    Zhao, Mingde
    Chang, Xiao-Wen
    Precup, Doina
    COMPLEX NETWORKS & THEIR APPLICATIONS XII, VOL 1, COMPLEX NETWORKS 2023, 2024, 1141 : 49 - 60
  • [23] Crystal Graph Convolutional Neural Networks for an Accurate and Interpretable Prediction of Material Properties
    Xie, Tian
    Grossman, Jeffrey C.
    PHYSICAL REVIEW LETTERS, 2018, 120 (14)
  • [24] A fast and scalable architecture to run convolutional neural networks in low density FPGAs
    Vestias, Mario P.
    Duarte, Rui P.
    de Sousa, Jose T.
    Neto, Horacio C.
    MICROPROCESSORS AND MICROSYSTEMS, 2020, 77
  • [25] Stress predictions in polycrystal plasticity using graph neural networks with subgraph training
    Zhai, Hanfeng
    COMPUTATIONAL MECHANICS, 2025,
  • [26] Sketch-GNN: Scalable Graph Neural Networks with Sublinear Training Complexity
    Ding, Mucong
    Rabbani, Tahseen
    An, Bang
    Wang, Evan Z.
    Huang, Furong
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [27] Scalable and Adaptive Graph Neural Networks with Self-Label-Enhanced Training
    Sun, Chuxiong
    Hu, Jie
    Gu, Hongming
    Chen, Jinpeng
    Liang, Wei
    Yang, Mingchuan
    PATTERN RECOGNITION, 2025, 160
  • [28] Regularization and Iterative Initialization of Softmax for Fast Training of Convolutional Neural Networks
    Rao, Qiang
    Yu, Bing
    He, Kun
    Feng, Bailan
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [29] Sketch-GNN: Scalable Graph Neural Networks with Sublinear Training Complexity
    Ding, Mucong
    Rabbani, Tahseen
    An, Bang
    Wang, Evan Z.
    Huang, Furong
    Advances in Neural Information Processing Systems, 2022, 35
  • [30] A Framework for Recommending Accurate and Diverse Items Using Bayesian Graph Convolutional Neural Networks
    Sun, Jianing
    Guo, Wei
    Zhang, Dengcheng
    Zhang, Yingxue
    Regol, Florence
    Hu, Yaochen
    Guo, Huifeng
    Tang, Ruiming
    Yuan, Han
    He, Xiuqiang
    Coates, Mark
    KDD '20: PROCEEDINGS OF THE 26TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2020, : 2030 - 2039