Prefix-Propagation: Parameter-Efficient Tuning for Long Sequences

被引:0
|
作者
Li, Jonathan [2 ]
Aitken, Will [1 ,2 ]
Bhambhoria, Rohan [1 ,2 ]
Zhu, Xiaodan [1 ,2 ]
机构
[1] Queens Univ, Dept Elect & Comp Engn, Kingston, ON, Canada
[2] Queens Univ, Ingenu Labs Res Inst, Kingston, ON, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Parameter-efficient tuning aims to mitigate the large memory requirements of adapting pretrained language models for downstream tasks. For example, one popular method, prefix-tuning (Li and Liang, 2021; Liu et al., 2022), prepends trainable tokens to sequences while freezing the rest of the model's parameters. Although such models attain comparable performance with fine-tuning when applied to sequences with short to moderate lengths, we show their inferior performance when modelling long sequences. To bridge this gap, we propose prefix-propagation, a simple but effective approach that conditions prefixes on previous hidden states. We empirically demonstrate that prefix-propagation outperforms prefix-tuning across long-document tasks, while using similar to 50% fewer parameters. To further investigate the proposed architecture, we also show its advantage in calibration, and perform additional study on its relationship with kernel attention. To the best of our knowledge, this work is the first to focus on parameter-efficient learning for long-sequence language tasks.(1)
引用
收藏
页码:1408 / 1419
页数:12
相关论文
共 50 条
  • [21] Bayesian Parameter-Efficient Fine-Tuning for Overcoming Catastrophic Forgetting
    Chen, Haolin
    Garner, Philip N.
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 4253 - 4262
  • [22] Scattered or Connected? An Optimized Parameter-efficient Tuning Approach for Information Retrieval
    Ma, Xinyu
    Guo, Jiafeng
    Zhang, Ruqing
    Fan, Yixing
    Cheng, Xueqi
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 1471 - 1480
  • [23] Parameter-efficient fine-tuning of large language models using semantic knowledge tuning
    Prottasha, Nusrat Jahan
    Mahmud, Asif
    Sobuj, Md. Shohanur Islam
    Bhat, Prakash
    Kowsher, Md
    Yousefi, Niloofar
    Garibay, Ozlem Ozmen
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [24] Automatic depression severity assessment with deep learning using parameter-efficient tuning
    Lau, Clinton
    Zhu, Xiaodan
    Chan, Wai-Yip
    FRONTIERS IN PSYCHIATRY, 2023, 14
  • [25] Token Embeddings Augmentation benefits Parameter-Efficient Fine-Tuning under long-tailed distribution
    Wang, Weiqiu
    Chen, Zining
    Zhao, Zhicheng
    Su, Fei
    NEUROCOMPUTING, 2025, 615
  • [26] One is Not Enough: Parameter-Efficient Fine-Tuning With Multiplicative Sparse Factorization
    Chen, Xuxi
    Chen, Tianlong
    Cheng, Yu
    Chen, Weizhu
    Awadallah, Ahmed Hassan
    Wang, Zhangyang
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2024, 18 (06) : 1059 - 1069
  • [27] Parameter-Efficient Low-Resource Dialogue State Tracking by Prompt Tuning
    Ma, Mingyu Derek
    Kao, Jiun-Yu
    Gao, Shuyang
    Gupta, Arpit
    Jin, Di
    Chung, Tagyoung
    Peng, Nanyun
    INTERSPEECH 2023, 2023, : 4653 - 4657
  • [28] Bridging Vision and Language Encoders: Parameter-Efficient Tuning for Referring Image Segmentation
    Xu, Zunnan
    Chen, Zhihong
    Zhang, Yong
    Song, Yibing
    Wan, Xiang
    Li, Guanbin
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 17457 - 17466
  • [29] Parameter-efficient fine-tuning in large language models: a survey of methodologies
    Luping Wang
    Sheng Chen
    Linnan Jiang
    Shu Pan
    Runze Cai
    Sen Yang
    Fei Yang
    Artificial Intelligence Review, 58 (8)
  • [30] Parameter-Efficient Tuning for Object Tracking by Migrating Pre-Trained Decoders
    Zhang, Ruijuan
    Wang, Li
    Yang, Song
    ELECTRONICS, 2024, 13 (23):