Probabilistic Weight Fixing: Large-scale training of neural network weight uncertainties for quantization

被引:0
|
作者
Subia-Waud, Christopher [1 ]
Dasmahapatra, Srinandan [1 ]
机构
[1] Univ Southampton, Sch Elect & Comp Sci, Southampton, Hants, England
基金
英国科研创新办公室;
关键词
TUTORIAL;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Weight-sharing quantization has emerged as a technique to reduce energy expenditure during inference in large neural networks by constraining their weights to a limited set of values. However, existing methods often assume weights are treated solely based on value, neglecting the unique role of weight position. This paper proposes a probabilistic framework based on Bayesian neural networks (BNNs) and a variational relaxation to identify which weights can be moved to which cluster center and to what degree based on their individual position-specific learned uncertainty distributions. We introduce a new initialization setting and a regularization term, enabling the training of BNNs with complex dataset-model combinations. Leveraging the flexibility of weight values from probability distributions, we enhance noise resilience and compressibility. Our iterative clustering procedure demonstrates superior compressibility and higher accuracy compared to state-of-the-art methods on both ResNet models and the more complex transformer-based architectures. In particular, our method outperforms the state-of-the-art quantization method top-1 accuracy by 1.6% on ImageNet using DeiT-Tiny, with its 5 million+ weights now represented by only 296 unique values. Code available at https://github.com/subiawaud/PWFN.
引用
收藏
页数:15
相关论文
共 50 条
  • [21] A variable-weight neural network combined predicting model to the trend predicting of the state development of the large-scale rotary sets
    Xu, Xiao-Li
    Zuo, Yun-Bo
    Zhu, Chun-Inei
    [J]. WMSCI 2006: 10TH WORLD MULTI-CONFERENCE ON SYSTEMICS, CYBERNETICS AND INFORMATICS, VOL III, PROCEEDINGS, 2006, : 304 - +
  • [22] TIGER: Training Inductive Graph Neural Network for Large-scale Knowledge Graph Reasoning
    Wang, Kai
    Xu, Yuwei
    Luo, Siqiang
    [J]. PROCEEDINGS OF THE VLDB ENDOWMENT, 2024, 17 (10): : 2459 - 2472
  • [23] POSTER: ParGNN: Efficient Training for Large-Scale Graph Neural Network on GPU Clusters
    Li, Shunde
    Gu, Junyu
    Wang, Jue
    Yao, Tiechui
    Liang, Zhiqiang
    Shi, Yumeng
    Li, Shigang
    Xi, Weiting
    Li, Shushen
    Zhou, Chunbao
    Wang, Yangang
    Chi, Xuebin
    [J]. PROCEEDINGS OF THE 29TH ACM SIGPLAN ANNUAL SYMPOSIUM ON PRINCIPLES AND PRACTICE OF PARALLEL PROGRAMMING, PPOPP 2024, 2024, : 469 - 471
  • [25] Preventive Network Protection in Probabilistic Large-Scale Failure Scenarios
    Izaddoost, Alireza
    Heydari, Shahram Shah
    [J]. 2012 IEEE GLOBECOM WORKSHOPS (GC WKSHPS), 2012, : 858 - 862
  • [26] Probabilistic neural network with the concept of edge weight-based entropy
    Alharbi, Rehab
    Ahmad, Ali
    Azeem, Muhammad
    Koam, Ali N. A.
    [J]. MOLECULAR PHYSICS, 2023, 121 (16)
  • [27] Large-Scale Emulation Network Topology Partition Based on Community Detection With the Weight of Vertex Similarity
    Yan, Jianen
    Xu, Haiyan
    Li, Ning
    Zhang, Zhaoxin
    [J]. Computer Journal, 2023, 66 (08): : 1817 - 1828
  • [28] Large-Scale Emulation Network Topology Partition Based on Community Detection With the Weight of Vertex Similarity
    Yan, Jianen
    Xu, Haiyan
    Li, Ning
    Zhang, Zhaoxin
    [J]. COMPUTER JOURNAL, 2023, 66 (08): : 1817 - 1828
  • [29] Deep Scaling Factor Quantization Network for Large-scale Image Retrieval
    Deng, Ziqing
    Lai, Zhihui
    Ding, Yujuan
    Kong, Heng
    Wu, Xu
    [J]. PROCEEDINGS OF THE 4TH ANNUAL ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2024, 2024, : 851 - 859
  • [30] LEARNING VECTOR QUANTIZATION FOR THE PROBABILISTIC NEURAL NETWORK
    BURRASCANO, P
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 1991, 2 (04): : 458 - 461