HiFi: High-Information Attention Heads Hold for Parameter-Efficient Model Adaptation

被引:0
|
作者
Gui, Anchun [1 ]
Xiao, Han [1 ]
机构
[1] Xiamen Univ, Sch Informat, Dept Artificial Intelligence, Xiamen, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To fully leverage the advantages of large-scale pre-trained language models (PLMs) on downstream tasks, it has become a ubiquitous adaptation paradigm to fine-tune the entire parameters of PLMs. However, this paradigm poses issues of inefficient updating and resource over-consuming for fine-tuning in data-scarce and resource-limited scenarios, because of the large scale of parameters in PLMs. To alleviate these concerns, in this paper, we propose a parameter-efficient fine-tuning method HiFi, that is, only the highly informative and strongly correlated attention heads for the specific task are fine-tuned. To search for those significant attention heads, we develop a novel framework to analyze the effectiveness of heads. Specifically, we first model the relationship between heads into a graph from two perspectives of information richness and correlation, and then apply PageRank algorithm to determine the relative importance of each head. Extensive experiments on the GLUE benchmark demonstrate the effectiveness of our method, and show that HiFi obtains state-of-the-art performance over the prior baselines.
引用
收藏
页码:8521 / 8537
页数:17
相关论文
共 50 条
  • [41] Towards Adaptive Prefix Tuning for Parameter-Efficient Language Model Fine-tuning
    Zhang, Zhen-Ru
    Tan, Chuanqi
    Xu, Haiyang
    Wang, Chengyu
    Huang, Jun
    Huang, Songfang
    61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, 2023, : 1239 - 1248
  • [42] ERAT-DLoRA: Parameter-efficient tuning with enhanced range adaptation in time and depth aware dynamic LoRA
    Luo, Dan
    Zheng, Kangfeng
    Wu, Chunhua
    Wang, Xiujuan
    Wang, Jvjie
    NEUROCOMPUTING, 2025, 614
  • [43] DyLoRA: Parameter-Efficient Tuning of Pretrained Models using Dynamic Search-Free Low Rank Adaptation
    Valipour, Mojtaba
    Rezagholizadeh, Mehdi
    Kobyzev, Ivan
    Ghodsi, Ali
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 3274 - 3287
  • [44] Separate the Wheat from the Chaff: Model Deficiency Unlearning via Parameter-Efficient Module Operation
    Hu, Xinshuo
    Li, Dongfang
    Hu, Baotian
    Zheng, Zihao
    Liu, Zhenyu
    Zhang, Min
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 16, 2024, : 18252 - 18260
  • [45] PRISM-Med: Parameter-Efficient Robust Interdomain Specialty Model for Medical Language Tasks
    Kang, Jieui
    Ryu, Hyungon
    Sim, Jaehyeong
    IEEE ACCESS, 2025, 13 : 4957 - 4965
  • [46] UPetu: A Unified Parameter-Efficient Fine-Tuning Framework for Remote Sensing Foundation Model
    Dong, Zhe
    Gu, Yanfeng
    Liu, Tianzhu
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62 : 1 - 13
  • [47] Discriminative Sample-Guided and Parameter-Efficient Feature Space Adaptation for Cross-Domain Few-Shot Learning
    Perera, Rashindrie
    Halgamuge, Saman
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 23794 - 23804
  • [48] Baize: An Open-Source Chat Model with Parameter-Efficient Tuning on Self-Chat Data
    Xu, Canwen
    Guo, Daya
    Duan, Nan
    McAuley, Julian
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 6268 - 6278
  • [49] HyperPELT: Unified Parameter-Efficient Language Model Tuning for Both Language and Vision-and-Language Tasks
    Zhang, Zhengkun
    Guo, Wenya
    Meng, Xiaojun
    Wang, Yasheng
    Wang, Yadao
    Jiang, Xin
    Liu, Qun
    Yang, Zhenglu
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 11442 - 11453
  • [50] Parameter-efficient fine-tuning large language model approach for hospital discharge paper summarization
    Goswami, Joyeeta
    Prajapati, Kaushal Kumar
    Saha, Ashim
    Saha, Apu Kumar
    APPLIED SOFT COMPUTING, 2024, 157