Parameter Transfer Unit for Deep Neural Networks

被引:12
|
作者
Zhang, Yinghua [1 ]
Zhang, Yu [1 ]
Yang, Qiang [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Dept Comp Sci & Engn, Kowloon, Hong Kong, Peoples R China
关键词
Transfer learning; Deep neural networks;
D O I
10.1007/978-3-030-16145-3_7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Parameters in deep neural networks which are trained on large-scale databases can generalize across multiple domains, which is referred as "transferability". Unfortunately, the transferability is usually defined as discrete states and it differs with domains and network architectures. Existing works usually heuristically apply parameter-sharing or fine-tuning, and there is no principled approach to learn a parameter transfer strategy. To address the gap, a Parameter Transfer Unit (PTU) is proposed in this paper. PTU learns a fine-grained nonlinear combination of activations from both the source domain network and the target domain network, and subsumes hand-crafted discrete transfer states. In the PTU, the transferability is controlled by two gates which are artificial neurons and can be learned from data. The PTU is a general and flexible module which can be used in both CNNs and RNNs. It can be also integrated with other transfer learning methods in a plug-and-play manner. Experiments are conducted with various network architectures and multiple transfer domain pairs. Results demonstrate the effectiveness of the PTU as it outperforms heuristic parameter-sharing and fine-tuning in most settings.
引用
收藏
页码:82 / 95
页数:14
相关论文
共 50 条
  • [1] Architecture of neural processing unit for deep neural networks
    Lee, Kyuho J.
    HARDWARE ACCELERATOR SYSTEMS FOR ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING, 2021, 122 : 217 - 245
  • [2] Transfer Entropy in Deep Neural Networks
    Andonie, R.
    Cataron, A.
    Moldovan, A.
    INTERNATIONAL JOURNAL OF COMPUTERS COMMUNICATIONS & CONTROL, 2025, 20 (01)
  • [3] Incremental Trainable Parameter Selection in Deep Neural Networks
    Thakur, Anshul
    Abrol, Vinayak
    Sharma, Pulkit
    Zhu, Tingting
    Clifton, David A.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (05) : 6478 - 6491
  • [4] Parameter inference with deep jointly informed neural networks
    Humbird, Kelli D.
    Peterson, J. Luc
    McClarren, Ryan G.
    STATISTICAL ANALYSIS AND DATA MINING, 2019, 12 (06) : 496 - 504
  • [5] DSNNs:learning transfer from deep neural networks to spiking neural networks
    张磊
    Du Zidong
    Li Ling
    Chen Yunji
    HighTechnologyLetters, 2020, 26 (02) : 136 - 144
  • [6] DSNNs: learning transfer from deep neural networks to spiking neural networks
    Zhang L.
    Du Z.
    Li L.
    Chen Y.
    High Technology Letters, 2020, 26 (02): : 136 - 144
  • [7] Neural Artistic Style Transfer Using Deep Neural Networks
    Gupta, Bharath
    Govinda, K.
    Rajkumar, R.
    Masih, Jolly
    PROCEEDINGS OF ACADEMIA-INDUSTRY CONSORTIUM FOR DATA SCIENCE (AICDS 2020), 2022, 1411 : 1 - 12
  • [8] Deep neural networks classifying transfer efficiency in complex networks
    Melnikov, Alexey A.
    Fedichkin, Leonid E.
    Lee, Ray-Kuang
    Alodjants, Alexander
    2020 OPTO-ELECTRONICS AND COMMUNICATIONS CONFERENCE (OECC 2020), 2020,
  • [9] Hierarchical deep convolution neural networks based on transfer learning for transformer rectifier unit fault diagnosis
    Chen, Shuwen
    Ge, Hongjuan
    Li, Huang
    Sun, Youchao
    Qian, Xiaoyan
    MEASUREMENT, 2021, 167
  • [10] Parameter-Efficient Deep Neural Networks With Bilinear Projections
    Yu, Litao
    Gao, Yongsheng
    Zhou, Jun
    Zhang, Jian
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (09) : 4075 - 4085