Parameter-Efficient Sparse Retrievers and Rerankers Using Adapters

被引:2
|
作者
Pal, Vaishali [1 ,2 ]
Lassance, Carlos [2 ]
Dejean, Herve [2 ]
Clinchant, Stephane [2 ]
机构
[1] Univ Amsterdam, IRLab, Amsterdam, Netherlands
[2] Naver Labs Europe, Meylan, France
关键词
Adapters; Information Retrieval; Sparse neural retriever;
D O I
10.1007/978-3-031-28238-6_2
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Parameter-Efficient transfer learning with Adapters have been studied in Natural Language Processing (NLP) as an alternative to full fine-tuning. Adapters are memory-efficient and scale well with downstream tasks by training small bottle-neck layers added between transformer layers while keeping the large pretrained language model (PLMs) frozen. In spite of showing promising results in NLP, these methods are under-explored in Information Retrieval. While previous studies have only experimented with dense retriever or in a cross lingual retrieval scenario, in this paper we aim to complete the picture on the use of adapters in IR. First, we study adapters for SPLADE, a sparse retriever, for which adapters not only retain the efficiency and effectiveness otherwise achieved by finetuning, but are memory-efficient and orders of magnitude lighter to train. We observe that Adapters-SPLADE not only optimizes just 2% of training parameters, but outperforms fully fine-tuned counterpart and existing parameter-efficient dense IR models on IR benchmark datasets. Secondly, we address domain adaptation of neural retrieval thanks to adapters on cross-domain BEIR datasets and TripClick. Finally, we also consider knowledge sharing between rerankers and first stage rankers. Overall, our study complete the examination of adapters for neural IR. (The code can be found at: https://github.com/naver/splade/tree/adapter-splade.)
引用
收藏
页码:16 / 31
页数:16
相关论文
共 50 条
  • [41] Parameter-efficient online knowledge distillation for pretrained language models
    Wang, Yukun
    Wang, Jin
    Zhang, Xuejie
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 265
  • [42] Parameter-Efficient Abstractive Question Answering over Tables or Text
    Pal, Vaishali
    Kanoulas, Evangelos
    de Rijke, Maarten
    PROCEEDINGS OF THE SECOND DIALDOC WORKSHOP ON DOCUMENT-GROUNDED DIALOGUE AND CONVERSATIONAL QUESTION ANSWERING (DIALDOC 2022), 2022, : 41 - 53
  • [43] Prefix-Propagation: Parameter-Efficient Tuning for Long Sequences
    Li, Jonathan
    Aitken, Will
    Bhambhoria, Rohan
    Zhu, Xiaodan
    61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, 2023, : 1408 - 1419
  • [44] Parameter-Efficient Learning for Text-to-Speech Accent Adaptation
    Yang, Li-Jen
    Yang, Chao-Han Huck
    Chien, Jen-Tzung
    INTERSPEECH 2023, 2023, : 4354 - 4358
  • [45] Plant leaf disease identification by parameter-efficient transformer with adapter
    Xu, Xingshi
    Yang, Guangyuan
    Wang, Yunfei
    Shang, Yuying
    Hua, Zhixin
    Wang, Zheng
    Song, Huaibo
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 138
  • [46] Parameter-efficient feature-based transfer for paraphrase identification
    Liu, Xiaodong
    Rzepka, Rafal
    Araki, Kenji
    NATURAL LANGUAGE ENGINEERING, 2023, 29 (04) : 1066 - 1096
  • [47] LoRA-Whisper: Parameter-Efficient and Extensible Multilingual ASR
    Song, Zheshu
    Zhuoi, Jianheng
    Yang, Yifan
    Ma, Ziyang
    Zhang, Shixiong
    Chen, Xie
    INTERSPEECH 2024, 2024, : 3934 - 3938
  • [48] Parameter-Efficient Adaptation of Foundation Models for Damaged Building Assessment
    Zhao, Fei
    Zhang, Chengcui
    2024 IEEE 7TH INTERNATIONAL CONFERENCE ON MULTIMEDIA INFORMATION PROCESSING AND RETRIEVAL, MIPR 2024, 2024, : 417 - 422
  • [49] AnomalyAdapters: Parameter-Efficient Multi-Anomaly Task Detection
    Unal, Ugur
    Dag, Hasan
    IEEE ACCESS, 2022, 10 : 5635 - 5646
  • [50] Refocus the Attention for Parameter-Efficient Thermal Infrared Object Tracking
    Lai, Simiao
    Liu, Chang
    Wang, Dong
    Lu, Huchuan
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,