Probabilistic Weight Fixing: Large-scale training of neural network weight uncertainties for quantization

被引:0
|
作者
Subia-Waud, Christopher [1 ]
Dasmahapatra, Srinandan [1 ]
机构
[1] Univ Southampton, Sch Elect & Comp Sci, Southampton, Hants, England
基金
英国科研创新办公室;
关键词
TUTORIAL;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Weight-sharing quantization has emerged as a technique to reduce energy expenditure during inference in large neural networks by constraining their weights to a limited set of values. However, existing methods often assume weights are treated solely based on value, neglecting the unique role of weight position. This paper proposes a probabilistic framework based on Bayesian neural networks (BNNs) and a variational relaxation to identify which weights can be moved to which cluster center and to what degree based on their individual position-specific learned uncertainty distributions. We introduce a new initialization setting and a regularization term, enabling the training of BNNs with complex dataset-model combinations. Leveraging the flexibility of weight values from probability distributions, we enhance noise resilience and compressibility. Our iterative clustering procedure demonstrates superior compressibility and higher accuracy compared to state-of-the-art methods on both ResNet models and the more complex transformer-based architectures. In particular, our method outperforms the state-of-the-art quantization method top-1 accuracy by 1.6% on ImageNet using DeiT-Tiny, with its 5 million+ weights now represented by only 296 unique values. Code available at https://github.com/subiawaud/PWFN.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] A Large-Scale Study of Probabilistic Calibration in Neural Network Regression
    Dheur, Victor
    Ben Taieb, Souhaib
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 202, 2023, 202
  • [2] MUSCLE HYPERTROPHY WITH LARGE-SCALE WEIGHT-LOSS AND RESISTANCE TRAINING
    DONNELLY, JE
    SHARP, T
    HOUMARD, J
    CARLSON, MG
    HILL, JO
    WHATLEY, JE
    ISRAEL, RG
    [J]. AMERICAN JOURNAL OF CLINICAL NUTRITION, 1993, 58 (04): : 561 - 565
  • [3] Parallel Large-Scale Neural Network Training For Online Advertising
    Qi, Quanchang
    Lu, Guangming
    Zhang, Jun
    Yang, Lichun
    Liu, Haishan
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2018, : 343 - 350
  • [4] Neural network weight training by mutation
    Jenkins, W. M.
    [J]. COMPUTERS & STRUCTURES, 2006, 84 (31-32) : 2107 - 2112
  • [5] A large-scale in-memory computing for deep neural network with trained quantization
    Cheng, Yuan
    Wang, Chao
    Chen, Hai-Bao
    Yu, Hao
    [J]. INTEGRATION-THE VLSI JOURNAL, 2019, 69 : 345 - 355
  • [6] Weight Quantization in Spiking Neural Network for Hardware Implementation
    Sulaiman, Muhammad Bintang Gemintang
    Juang, Kai-Cheung
    Lu, Chih-Cheng
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS - TAIWAN (ICCE-TAIWAN), 2020,
  • [7] Training Convolutional Neural Network for Sketch Recognition on Large-Scale Dataset
    Zhou, Wen
    Jia, Jinyuan
    [J]. INTERNATIONAL ARAB JOURNAL OF INFORMATION TECHNOLOGY, 2020, 17 (01) : 82 - 89
  • [8] Accelerating Large-Scale Distributed Neural Network Training with SPMD Parallelism
    Zhang, Shiwei
    Diao, Lansong
    Wu, Chuan
    Wang, Siyu
    Lin, Wei
    [J]. PROCEEDINGS OF THE 13TH SYMPOSIUM ON CLOUD COMPUTING, SOCC 2022, 2022, : 403 - 418
  • [9] Accelerating Large-Scale Graph Neural Network Training on Crossbar Diet
    Ogbogu, Chukwufumnanya
    Arka, Aqeeb Iqbal
    Joardar, Biresh Kumar
    Doppa, Janardhan Rao
    Li, Hai
    Chakrabarty, Krishnendu
    Pande, Partha Pratim
    [J]. IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2022, 41 (11) : 3626 - 3637
  • [10] ON LARGE-SCALE QUANTIZATION
    CARVALHO, JC
    [J]. LETTERE AL NUOVO CIMENTO, 1985, 44 (06): : 337 - 342