Prefix-Propagation: Parameter-Efficient Tuning for Long Sequences

被引:0
|
作者
Li, Jonathan [2 ]
Aitken, Will [1 ,2 ]
Bhambhoria, Rohan [1 ,2 ]
Zhu, Xiaodan [1 ,2 ]
机构
[1] Queens Univ, Dept Elect & Comp Engn, Kingston, ON, Canada
[2] Queens Univ, Ingenu Labs Res Inst, Kingston, ON, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Parameter-efficient tuning aims to mitigate the large memory requirements of adapting pretrained language models for downstream tasks. For example, one popular method, prefix-tuning (Li and Liang, 2021; Liu et al., 2022), prepends trainable tokens to sequences while freezing the rest of the model's parameters. Although such models attain comparable performance with fine-tuning when applied to sequences with short to moderate lengths, we show their inferior performance when modelling long sequences. To bridge this gap, we propose prefix-propagation, a simple but effective approach that conditions prefixes on previous hidden states. We empirically demonstrate that prefix-propagation outperforms prefix-tuning across long-document tasks, while using similar to 50% fewer parameters. To further investigate the proposed architecture, we also show its advantage in calibration, and perform additional study on its relationship with kernel attention. To the best of our knowledge, this work is the first to focus on parameter-efficient learning for long-sequence language tasks.(1)
引用
收藏
页码:1408 / 1419
页数:12
相关论文
共 50 条
  • [31] Parameter-Efficient Prompt Tuning Makes Generalized and Calibrated Neural Text Retrievers
    Tam, Weng Lam
    Liu, Xiao
    Ji, Kaixuan
    Xue, Lilong
    Zhang, Xingjian
    Dong, Yuxiao
    Lin, Jiahua
    Hu, Maodi
    Tang, Jie
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 13117 - 13130
  • [32] Strong Baselines for Parameter-Efficient Few-Shot Fine-Tuning
    Basu, Samyadeep
    Hu, Shell
    Massiceti, Daniela
    Feizi, Soheil
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 10, 2024, : 11024 - 11031
  • [33] Parameter-Efficient Fine-Tuning Large Speech Model Based on LoRA
    Ou, Ling
    Feng, Gen
    PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 36 - 41
  • [34] Parameter-efficient Tuning for Large Language Model without Calculating Its Gradients
    Jin, Feihu
    Zhang, Jiajun
    Zong, Chengqing
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 321 - 330
  • [35] Parameter-Efficient Masking Networks
    Bai, Yue
    Wang, Huan
    Ma, Xu
    Zhang, Yitian
    Tao, Zhiqiang
    Fu, Yun
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [36] Pass-Tuning: Towards Structure-Aware Parameter-Efficient Tuning for Code Representation Learning
    Chen, Nuo
    Sun, Qiushi
    Wang, Jianing
    Li, Xiang
    Gao, Ming
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 577 - 591
  • [37] SCT: A Simple Baseline for Parameter-Efficient Fine-Tuning via Salient Channels
    Zhao, Henry Hengyuan
    Wang, Pichao
    Zhao, Yuyang
    Luo, Hao
    Wang, Fan
    Shou, Mike Zheng
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (03) : 731 - 749
  • [38] PROTECT: Parameter-Efficient Tuning for Few-Shot Robust Chinese Text Correction
    Feng, Xuan
    Gu, Tianlong
    Chang, Liang
    Liu, Xiaoli
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 3270 - 3282
  • [39] XtremeCLIP: Extremely Parameter-efficient Tuning for Low-resource Vision Language Understanding
    Tang, Moming
    Wang, Chengyu
    Wang, Jianing
    Tan, Chuanqi
    Huang, Songfang
    Chen, Cen
    Qian, Weining
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 6368 - 6376
  • [40] Federated Learning of Large Language Models with Parameter-Efficient Prompt Tuning and Adaptive Optimization
    Che, Tianshi
    Liu, Ji
    Zhou, Yang
    Ren, Jiaxiang
    Zhou, Jiwen
    Sheng, Victor S.
    Dai, Huaiyu
    Dou, Dejing
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 7871 - 7888