AdderNet: Do We Really Need Multiplications in Deep Learning?

被引:127
|
作者
Chen, Hanting [1 ,2 ]
Wang, Yunhe [2 ]
Xu, Chunjing [2 ]
Shi, Boxin [3 ,4 ]
Xu, Chao [1 ]
Tian, Qi [2 ]
Xu, Chang [5 ]
机构
[1] Peking Univ, Dept Machine Intelligence, Key Lab Machine Percept MOE, Beijing, Peoples R China
[2] Huawei Technol, Noahs Ark Lab, Shenzhen, Peoples R China
[3] Peking Univ, Dept CS, NELVT, Beijing, Peoples R China
[4] Peng Cheng Lab, Shenzhen, Peoples R China
[5] Univ Sydney, Fac Engn, Sch Comp Sci, Sydney, NSW, Australia
基金
国家重点研发计划; 澳大利亚研究理事会; 中国国家自然科学基金;
关键词
D O I
10.1109/CVPR42600.2020.00154
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Compared with cheap addition operation, multiplication operation is of much higher computation complexity. The widely-used convolutions in deep neural networks are exactly cross-correlation to measure the similarity between input feature and convolution filters, which involves massive multiplications between float values. In this paper, we present adder networks (AdderNets) to trade these massive multiplications in deep neural networks, especially convolutional neural networks (CNNs), for much cheaper additions to reduce computation costs. In AdderNets, we take the L 1 -norm distance between filters and input feature as the output response. The influence of this new similarity measure on the optimization of neural network have been thoroughly analyzed. To achieve a better performance, we develop a special back-propagation approach for AdderNets by investigating the full-precision gradient. We then propose an adaptive learning rate strategy to enhance the training procedure of AdderNets according to the magnitude of each neuron's gradient. As a result, the proposed AdderNets can achieve 74.9% Top-I accuracy 91.7% Top-5 accuracy using ResNet-50 on the ImageNet dataset without any multiplication in convolutional layer. The codes are publicly available at: https://github.com/huaweinoah/AdderNet.
引用
收藏
页码:1465 / 1474
页数:10
相关论文
共 50 条
  • [1] Do we really need to do this?
    Ehnert, Jesse
    [J]. JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2023, 153 (03):
  • [2] Do we still need deep learning?
    Vasile, Cristian
    [J]. JOURNAL OF EDUCATIONAL SCIENCES & PSYCHOLOGY, 2024, 14 (01): : 1 - 3
  • [3] Do we really need this?
    Milo, P
    [J]. EE-EVALUATION ENGINEERING, 2002, 41 (08): : 8 - 8
  • [4] Do we really need deep CNN for plant diseases identification?
    Li, Yang
    Nie, Jing
    Chao, Xuewei
    [J]. COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2020, 178
  • [5] DO WE REALLY NEED THE FISH
    HINMAN, G
    [J]. JOURNAL OF ENERGY ENGINEERING-ASCE, 1995, 121 (02): : 49 - 51
  • [6] DO WE REALLY NEED LEADERSHIP
    HOLZMAN, M
    [J]. EDUCATIONAL LEADERSHIP, 1992, 49 (05) : 36 - 40
  • [7] DO WE REALLY NEED DOCTORS
    PORTER, R
    [J]. NEW SOCIETY, 1984, 69 (1129): : 87 - 89
  • [8] DO WE REALLY NEED THE UNIVERSE
    BIGGIN, S
    [J]. NEW SCIENTIST, 1994, 144 (1953) : 51 - 52
  • [9] Do we really need that test?
    Phillips, C. Douglas
    [J]. APPLIED RADIOLOGY, 2016, 45 (09) : 48 - 48
  • [10] QA - DO WE REALLY NEED IT
    HART, J
    [J]. MANUFACTURING CHEMIST, 1987, 58 (12): : 62 - 63