Large-Scale Training in Neural Compact Models for Accurate and Adaptable MOSFET Simulation

被引:0
|
作者
Park, Chanwoo [1 ]
Lee, Seungjun [1 ]
Park, Junghwan [1 ]
Rim, Kyungjin [1 ]
Park, Jihun [1 ]
Cho, Seonggook [1 ]
Jeon, Jongwook [2 ]
Cho, Hyunbo [1 ]
机构
[1] Alsemy Inc, Res & Dev Ctr, Seoul 06154, South Korea
[2] Sungkyunkwan Univ, Sch Elect & Elect Engn, Suwon 03063, South Korea
关键词
Integrated circuit modeling; Adaptation models; Data models; Mathematical models; Capacitance-voltage characteristics; Predictive models; MOSFET; Compact model; DTCO; foundation model; neural network;
D O I
10.1109/JEDS.2024.3417521
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
We address the challenges associated with traditional analytical models, such as BSIM, in semiconductor device modeling. These models often face limitations in accurately representing the complex behaviors of miniaturized devices. As an alternative, Neural Compact Models (NCMs) offer improved modeling capabilities, but their effectiveness is constrained by a reliance on extensive datasets for accurate performance. In real-world scenarios, where measurements for device modeling are often limited, this dependence becomes a significant hindrance. In response, this work presents a large-scale pre-training approach for NCMs. By utilizing extensive datasets across various technology nodes, our method enables NCMs to develop a more detailed understanding of device behavior, thereby enhancing the accuracy and adaptability of MOSFET device simulations, particularly when data availability is limited. Our study illustrates the potential benefits of large-scale pre-training in enhancing the capabilities of NCMs, offering a practical solution to one of the key challenges in current device modeling practices.
引用
收藏
页码:745 / 751
页数:7
相关论文
共 50 条
  • [41] Accelerating Large-Scale Graph Neural Network Training on Crossbar Diet
    Ogbogu, Chukwufumnanya
    Arka, Aqeeb Iqbal
    Joardar, Biresh Kumar
    Doppa, Janardhan Rao
    Li, Hai
    Chakrabarty, Krishnendu
    Pande, Partha Pratim
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2022, 41 (11) : 3626 - 3637
  • [42] Rethinking residual connection in training large-scale spiking neural networks
    Li, Yudong
    Lei, Yunlin
    Yang, Xu
    NEUROCOMPUTING, 2025, 616
  • [43] An implicit solvent model for accurate simulation of large-scale protein conformational transitions
    Gong, Xiping
    Chen, Jianhan
    BIOPHYSICAL JOURNAL, 2022, 121 (03) : 274A - 274A
  • [44] Training large-scale language models with limited GPU memory: a survey
    Yu TANG
    Linbo QIAO
    Lujia YIN
    Peng LIANG
    Ao SHEN
    Zhilin YANG
    Lizhi ZHANG
    Dongsheng LI
    Frontiers of Information Technology & Electronic Engineering, 2025, 26 (03) : 309 - 331
  • [45] Training large-scale language models with limited GPU memory: a survey
    Tang, Yu
    Qiao, Linbo
    Yin, Lujia
    Liang, Peng
    Shen, Ao
    Yang, Zhilin
    Zhang, Lizhi
    Li, Dongsheng
    FRONTIERS OF INFORMATION TECHNOLOGY & ELECTRONIC ENGINEERING, 2025, : 309 - 331
  • [46] PipeTransformer: Automated Elastic Pipelining for Distributed Training of Large-scale Models
    He, Chaoyang
    Li, Shen
    Soltanolkotabi, Mahdi
    Avestimehr, Salman
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [47] Training Large-Scale News Recommenders with Pretrained Language Models in the Loop
    Xiao, Shitao
    Liu, Zheng
    Shao, Yingxia
    Di, Tao
    Middha, Bhuvan
    Wu, Fangzhao
    Xie, Xing
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 4215 - 4225
  • [48] Visual Diagnostics of Parallel Performance in Training Large-Scale DNN Models
    Wei, Yating
    Wang, Zhiyong
    Wang, Zhongwei
    Dai, Yong
    Ou, Gongchang
    Gao, Han
    Yang, Haitao
    Wang, Yue
    Cao, Caleb Chen
    Weng, Luoxuan
    Lu, Jiaying
    Zhu, Rongchen
    Chen, Wei
    IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2024, 30 (07) : 3915 - 3929
  • [49] MixPipe: Efficient Bidirectional Pipeline Parallelism for Training Large-Scale Models
    Zhang, Weigang
    Zhou, Biyu
    Tang, Xuehai
    Wang, Zhaoxing
    Hu, Songlin
    2023 60TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC, 2023,
  • [50] Large-scale simulation platform
    Institute of Cybernetics, Tallinn Technical University, Akadeemia tee 21, 12618 Tallinn, Estonia
    WSEAS Trans. Comput., 2007, 1 (65-71):