Mixed-Precision Continual Learning Based on Computational Resistance Random Access Memory

被引:11
|
作者
Li, Yi [1 ,2 ]
Zhang, Woyu [1 ,2 ]
Xu, Xiaoxin [1 ]
He, Yifan [3 ]
Dong, Danian [1 ]
Jiang, Nanjia [1 ]
Wang, Fei [1 ,2 ]
Guo, Zeyu [1 ,2 ]
Wang, Shaocong [4 ]
Dou, Chunmeng [1 ]
Liu, Yongpan [3 ]
Wang, Zhongrui [4 ]
Shang, Dashan [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Microelect, Key Lab Microelect Devices & Integrated Technol, Beijing 100029, Peoples R China
[2] Univ Chinese Acad Sci, Beijing 101408, Peoples R China
[3] Tsinghua Univ, Dept Elect Engn, Beijing 100084, Peoples R China
[4] Univ Hong Kong, Dept Elect & Elect Engn, Pok Fu Lam Rd, Hong Kong 999077, Peoples R China
基金
中国国家自然科学基金;
关键词
continual learning; in-memory computing; mixed precision; resistance random access memory;
D O I
10.1002/aisy.202200026
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Artificial neural networks have acquired remarkable achievements in the field of artificial intelligence. However, it suffers from catastrophic forgetting when dealing with continual learning problems, i.e., the loss of previously learned knowledge upon learning new information. Although several continual learning algorithms have been proposed, it remains a challenge to implement these algorithms efficiently on conventional digital systems due to the physical separation between memory and processing units. Herein, a software-hardware codesigned in-memory computing paradigm is proposed, where a mixed-precision continual learning (MPCL) model is deployed on a hybrid analogue-digital hardware system equipped with resistance random access memory chip. Software-wise, the MPCL effectively alleviates catastrophic forgetting and circumvents the requirement for high-precision weights. Hardware-wise, the hybrid analogue-digital system takes advantage of the colocation of memory and processing units, greatly improving energy efficiency. By combining the MPCL with an in situ fine-tuning method, high classification accuracies of 94.9% and 95.3% (software baseline 97.0% and 97.7%) on the 5-split-MNIST and 5-split-FashionMNIST are achieved, respectively. The proposed system reduces approximate to 200 times energy consumption of the multiply-and-accumulation operations during the inference phase compared to the conventional digital systems. This work paves the way for future autonomous systems at the edge.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] Mixed-Precision Deep Learning Based on Computational Memory
    Nandakumar, S. R.
    Le Gallo, Manuel
    Piveteau, Christophe
    Joshi, Vinay
    Mariani, Giovanni
    Boybat, Irem
    Karunaratne, Geethan
    Khaddam-Aljameh, Riduan
    Egger, Urs
    Petropoulos, Anastasios
    Antonakopoulos, Theodore
    Rajendran, Bipin
    Sebastian, Abu
    Eleftheriou, Evangelos
    FRONTIERS IN NEUROSCIENCE, 2020, 14
  • [2] Mixed-precision architecture based on computational memory for training deep neural networks
    Nandakumar, S. R.
    Le Gallo, Manuel
    Boybat, Irem
    Rajendran, Bipin
    Sebastian, Abu
    Eleftheriou, Evangelos
    2018 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2018,
  • [3] Mixed-precision in-memory computing
    Manuel Le Gallo
    Abu Sebastian
    Roland Mathis
    Matteo Manica
    Heiner Giefers
    Tomas Tuma
    Costas Bekas
    Alessandro Curioni
    Evangelos Eleftheriou
    Nature Electronics, 2018, 1 : 246 - 253
  • [4] Mixed-precision in-memory computing
    Le Gallo, Manuel
    Sebastian, Abu
    Mathis, Roland
    Manica, Matteo
    Giefers, Heiner
    Tuma, Tomas
    Bekas, Costas
    Curioni, Alessandro
    Eleftheriou, Evangelos
    NATURE ELECTRONICS, 2018, 1 (04): : 246 - 253
  • [5] Mixed-Precision Random Projection for RandNLA on Tensor Cores
    Ootomo, Hiroyuki
    Yokota, Rio
    PROCEEDINGS OF THE PLATFORM FOR ADVANCED SCIENTIFIC COMPUTING CONFERENCE, PASC 2023, 2023,
  • [6] MINT: Mixed-precision RRAM-based IN-memory Training Architecture
    Jiang, Hongwu
    Huang, Shanshi
    Peng, Xiaochen
    Yu, Shimeng
    2020 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2020,
  • [7] Mixed-Precision Partial Differential Equation Solver Design Based on Nonvolatile Memory
    Yang, Haozhang
    Huang, Peng
    Zhou, Zheng
    Zhang, Yizhou
    Han, Runze
    Liu, Xiaoyan
    Kang, Jinfeng
    IEEE TRANSACTIONS ON ELECTRON DEVICES, 2022, 69 (07) : 3708 - 3715
  • [8] SQNR-based Layer-wise Mixed-Precision Schemes with Computational Complexity Consideration
    Kim, Ha-Na
    Eun, Hyun
    Choi, Jung Hwan
    Kim, Ji-Hoon
    2022 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS 22), 2022, : 234 - 235
  • [9] Learning Generalizable Mixed-Precision Quantization via Attribution Imitation
    Wang, Ziwei
    Xiao, Han
    Zhou, Jie
    Lu, Jiwen
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (11) : 5101 - 5123
  • [10] A mixed-precision memristor and SRAM compute-in-memory AI processor
    Khwa, Win-San
    Wen, Tai-Hao
    Hsu, Hung-Hsi
    Huang, Wei-Hsing
    Chang, Yu-Chen
    Chiu, Ting-Chien
    Ke, Zhao-En
    Chin, Yu-Hsiang
    Wen, Hua-Jin
    Hsu, Wei-Ting
    Lo, Chung-Chuan
    Liu, Ren-Shuo
    Hsieh, Chih-Cheng
    Tang, Kea-Tiong
    Ho, Mon-Shu
    Lele, Ashwin Sanjay
    Teng, Shih-Hsin
    Chou, Chung-Cheng
    Chih, Yu-Der
    Chang, Tsung-Yung Jonathan
    Chang, Meng-Fan
    NATURE, 2025, : 617 - 623