Long-range zero-shot generative deep network quantization

被引:2
|
作者
Luo, Yan [1 ]
Gao, Yangcheng [1 ]
Zhang, Zhao [1 ,3 ]
Fan, Jicong [2 ,3 ]
Zhang, Haijun [4 ]
Xu, Mingliang [5 ]
机构
[1] Hefei Univ Technol, Sch Comp Sci & Informat Engn, Hefei, Peoples R China
[2] Chinese Univ Hong Kong, Sch Data Sci, Shenzhen, Peoples R China
[3] Shenzhen Res Inst Big Data, Shenzhen, Peoples R China
[4] Harbin Inst Technol, Sch Comp Sci, Shenzhen, Peoples R China
[5] Zhengzhou Univ, Sch Informat Engn, Zhengzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep network quantization; Long-range generator; Adversarial margin add; Synthetic data generation;
D O I
10.1016/j.neunet.2023.07.042
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Quantization approximates a deep network model with floating-point numbers by the model with low bit width numbers, thereby accelerating inference and reducing computation. Zero-shot quantization, which aims to quantize a model without access to the original data, can be achieved by fitting the real data distribution through data synthesis. However, it has been observed that zero-shot quantization leads to inferior performance compared to post-training quantization with real data for two primary reasons: 1) a normal generator has difficulty obtaining a high diversity of synthetic data since it lacks long-range information to allocate attention to global features, and 2) synthetic images aim to simulate the statistics of real data, which leads to weak intra-class heterogeneity and limited feature richness. To overcome these problems, we propose a novel deep network quantizer called long-range zero-shot generative deep network quantization (LRQ). Technically, we propose a long-range generator (LRG) to learn long-range information instead of simple local features. To incorporate more global features into the synthetic data, we use long-range attention with large-kernel convolution in the generator. In addition, we also present an adversarial margin add (AMA) module to force intra-class angular enlargement between the feature vector and class center. The AMA module forms an adversarial process that increases the convergence difficulty of the loss function, which is opposite to the training objective of the original loss function. Furthermore, to transfer knowledge from the full-precision network, we also utilize decoupled knowledge distillation. Extensive experiments demonstrate that LRQ obtains better performance than other competitors.(c) 2023 Elsevier Ltd. All rights reserved.
引用
收藏
页码:683 / 691
页数:9
相关论文
共 50 条
  • [31] Hard Sample Matters a Lot in Zero-Shot Quantization
    Li, Huantong
    Wu, Xiangmiao
    Lv, Fanbing
    Liao, Daihai
    Li, Thomas H.
    Zhang, Yonggang
    Han, Bo
    Tan, Mingkui
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 24417 - 24426
  • [32] Zero-Shot Learning with Joint Generative Adversarial Networks
    Zhang, Minwan
    Wang, Xiaohua
    Shi, Yueting
    Ren, Shiwei
    Wang, Weijiang
    ELECTRONICS, 2023, 12 (10)
  • [33] Alleviating Feature Confusion for Generative Zero-shot Learning
    Li, Jingjing
    Jing, Mengmeng
    Lu, Ke
    Zhu, Lei
    Yang, Yang
    Huang, Zi
    PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA (MM'19), 2019, : 1587 - 1595
  • [34] Investigating the Bilateral Connections in Generative Zero-Shot Learning
    Li, Jingjing
    Jing, Mengmeng
    Lu, Ke
    Zhu, Lei
    Shen, Heng Tao
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (08) : 8167 - 8178
  • [35] Gradient Matching Generative Networks for Zero-Shot Learning
    Sariyildiz, Mert Bulent
    Cinbis, Ramazan Gokberk
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 2163 - 2173
  • [36] Zero-Shot Learning of SAR Target Feature Space With Deep Generative Neural Networks
    Song, Qian
    Xu, Feng
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2017, 14 (12) : 2245 - 2249
  • [37] A Generative Approach to Zero-Shot and Few-Shot Action Recognition
    Mishra, Ashish
    Verma, Vinay Kumar
    Reddy, M. Shiva Krishna
    Arulkumar, S.
    Rai, Piyush
    Mittal, Anurag
    2018 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2018), 2018, : 372 - 380
  • [38] Ranking Synthetic Features for Generative Zero-Shot Learning
    Ramazi, Shayan
    Nadian-Ghomsheh, Ali
    2021 26TH INTERNATIONAL COMPUTER CONFERENCE, COMPUTER SOCIETY OF IRAN (CSICC), 2021,
  • [39] Leveraging the Invariant Side of Generative Zero-Shot Learning
    Li, Jingjing
    Jing, Mengmeng
    Lu, Ke
    Ding, Zhengming
    Zhu, Lei
    Huang, Zi
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 7394 - 7403
  • [40] Generative Multi-Label Zero-Shot Learning
    Gupta, Akshita
    Narayan, Sanath
    Khan, Salman
    Khan, Fahad Shahbaz
    Shao, Ling
    van de Weijer, Joost
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (12) : 14611 - 14624