Long-range zero-shot generative deep network quantization

被引:2
|
作者
Luo, Yan [1 ]
Gao, Yangcheng [1 ]
Zhang, Zhao [1 ,3 ]
Fan, Jicong [2 ,3 ]
Zhang, Haijun [4 ]
Xu, Mingliang [5 ]
机构
[1] Hefei Univ Technol, Sch Comp Sci & Informat Engn, Hefei, Peoples R China
[2] Chinese Univ Hong Kong, Sch Data Sci, Shenzhen, Peoples R China
[3] Shenzhen Res Inst Big Data, Shenzhen, Peoples R China
[4] Harbin Inst Technol, Sch Comp Sci, Shenzhen, Peoples R China
[5] Zhengzhou Univ, Sch Informat Engn, Zhengzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep network quantization; Long-range generator; Adversarial margin add; Synthetic data generation;
D O I
10.1016/j.neunet.2023.07.042
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Quantization approximates a deep network model with floating-point numbers by the model with low bit width numbers, thereby accelerating inference and reducing computation. Zero-shot quantization, which aims to quantize a model without access to the original data, can be achieved by fitting the real data distribution through data synthesis. However, it has been observed that zero-shot quantization leads to inferior performance compared to post-training quantization with real data for two primary reasons: 1) a normal generator has difficulty obtaining a high diversity of synthetic data since it lacks long-range information to allocate attention to global features, and 2) synthetic images aim to simulate the statistics of real data, which leads to weak intra-class heterogeneity and limited feature richness. To overcome these problems, we propose a novel deep network quantizer called long-range zero-shot generative deep network quantization (LRQ). Technically, we propose a long-range generator (LRG) to learn long-range information instead of simple local features. To incorporate more global features into the synthetic data, we use long-range attention with large-kernel convolution in the generator. In addition, we also present an adversarial margin add (AMA) module to force intra-class angular enlargement between the feature vector and class center. The AMA module forms an adversarial process that increases the convergence difficulty of the loss function, which is opposite to the training objective of the original loss function. Furthermore, to transfer knowledge from the full-precision network, we also utilize decoupled knowledge distillation. Extensive experiments demonstrate that LRQ obtains better performance than other competitors.(c) 2023 Elsevier Ltd. All rights reserved.
引用
收藏
页码:683 / 691
页数:9
相关论文
共 50 条
  • [1] Generative Zero-shot Network Quantization
    He, Xiangyu
    Lu, Jiahao
    Xu, Weixiang
    Hu, Qinghao
    Wang, Peisong
    Cheng, Jian
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 2994 - 3005
  • [2] ZERO-SHOT LEARNING OF A CONDITIONAL GENERATIVE ADVERSARIAL NETWORK FOR DATA-FREE NETWORK QUANTIZATION
    Choi, Yoojin
    El-Khamy, Mostafa
    Lee, Jungwon
    2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 3552 - 3556
  • [3] Zero-shot Adversarial Quantization
    Liu, Yuang
    Zhang, Wei
    Wang, Jun
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 1512 - 1521
  • [4] Deep quantization network with visual-semantic alignment for zero-shot image retrieval
    Liu, Huixia
    Qin, Zhihong
    ELECTRONIC RESEARCH ARCHIVE, 2023, 31 (07): : 4232 - 4247
  • [5] ROBUST BIDIRECTIONAL GENERATIVE NETWORK FOR GENERALIZED ZERO-SHOT LEARNING
    Xing, Yun
    Huang, Sheng
    Huangfu, Luwen
    Chen, Feiyu
    Ge, Yongxin
    2020 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2020,
  • [6] Cross-class generative network for zero-shot learning
    Liu, Jinlu
    Zhang, Zhaocheng
    Yang, Gang
    INFORMATION SCIENCES, 2021, 555 : 147 - 163
  • [7] Zero-shot image classification based on generative adversarial network
    Wei H.
    Zhang Y.
    Beijing Hangkong Hangtian Daxue Xuebao/Journal of Beijing University of Aeronautics and Astronautics, 2019, 45 (12): : 2345 - 2350
  • [8] GTNet: Generative Transfer Network for Zero-Shot Object Detection
    Zhao, Shizhen
    Gao, Changxin
    Shao, Yuanjie
    Li, Lerenhan
    Yu, Changqian
    Ji, Zhong
    Sang, Nang
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 12967 - 12974
  • [9] Generative Dual Adversarial Network for Generalized Zero-shot Learning
    Huang, He
    Wang, Changhu
    Yu, Philip S.
    Wang, Chang-Dong
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 801 - 810
  • [10] Generalized Zero-Shot Learning with Deep Calibration Network
    Liu, Shichen
    Long, Mingsheng
    Wang, Jianmin
    Jordan, Michael I.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31