Small Language Models Improve Giants by Rewriting Their Outputs

被引:0
|
作者
Vernikos, Giorgos [1 ,2 ,4 ]
Brazinskas, Arthur [3 ]
Adamek, Jakub [3 ]
Mallinson, Jonathan [3 ]
Severyn, Aliaksei [3 ]
Malmi, Eric [3 ]
机构
[1] Ecole Polytech Fed Lausanne, Lausanne, Switzerland
[2] HEIG VD HES SO, Yverdon, Switzerland
[3] Google Res, Mountain View, CA USA
[4] Google, Mountain View, CA USA
基金
瑞士国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Despite the impressive performance of large language models (LLMs), they often lag behind specialized models in various tasks. LLMs only use a fraction of the existing training data for in-context learning, while task-specific models harness the full dataset for fine-tuning. In this work, we tackle the problem of leveraging training data to improve the performance of LLMs without fine-tuning. Our approach directly targets LLM predictions without requiring access to their weights. We create a pool of candidates from the LLM through few-shot prompting and we employ a compact model, the LM-corrector (LMCOR), specifically trained to merge these candidates to produce an enhanced output. Our experiments on four natural language generation tasks demonstrate that even a small LMCOR model (250M) substantially improves the few-shot performance of LLMs (62B), matching and even outperforming standard fine-tuning. Furthermore, we illustrate the robustness of LMCOR against different prompts, thereby minimizing the need for extensive prompt engineering. Finally, we show that LMCOR can be seamlessly integrated with different LLMs at inference, serving as a plug-and-play module to improve their performance.
引用
收藏
页码:2703 / 2718
页数:16
相关论文
共 50 条
  • [1] Smoothing County-Level Sampling Variances to Improve Small Area Models' Outputs
    Chen, Lu
    Sartore, Luca
    Benecha, Habtamu
    Bejleri, Valbona
    Nandram, Balgobin
    STATS, 2022, 5 (03): : 898 - 915
  • [2] Small Language Models Fine-tuned to Coordinate Larger Language Models improve Complex Reasoning
    Juneja, Gurusha
    Dutta, Subhabrata
    Chakrabarti, Soumen
    Manchhanda, Sunny
    Chakraborty, Tanmoy
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 3675 - 3691
  • [3] Rewriting Conversational Utterances with Instructed Large Language Models
    Galimzhanova, Elnara
    Muntean, Cristina Ioana
    Nardini, Franco Maria
    Perego, Raffaele
    Rocchietti, Guido
    2023 IEEE INTERNATIONAL CONFERENCE ON WEB INTELLIGENCE AND INTELLIGENT AGENT TECHNOLOGY, WI-IAT, 2023, : 56 - 63
  • [4] Enhancing Interactive Image Retrieval With Query Rewriting Using Large Language Models and Vision Language Models
    Zhu, Hongyi
    Huang, Jia-Hong
    Rudinac, Stevan
    Kanoulas, Evangelos
    PROCEEDINGS OF THE 4TH ANNUAL ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2024, 2024, : 978 - 987
  • [5] Query Rewriting for Retrieval-Augmented Large Language Models
    Ma, Xinbei
    Gong, Yeyun
    He, Pengcheng
    Zhao, Hai
    Duan, Nan
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 5303 - 5315
  • [6] Post Hoc Explanations of Language Models Can Improve Language Models
    Krishna, Satyapriya
    Ma, Jiaqi
    Slack, Dylan
    Ghandeharioun, Asma
    Singh, Sameer
    Lakkaraju, Himabindu
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [7] How Emerging Giants Are Rewriting the Rules of M&A
    Kumar, Nirmalya
    HARVARD BUSINESS REVIEW, 2009, 87 (05) : 115 - +
  • [8] The Rise of Small Language Models
    Zhang, Qin
    Liu, Ziqi
    Pan, Shirui
    IEEE INTELLIGENT SYSTEMS, 2025, 40 (01) : 30 - 37
  • [9] REWRITING AND APPROPRIATING LANGUAGE
    Barbeiro, Luis
    ICERI2015: 8TH INTERNATIONAL CONFERENCE OF EDUCATION, RESEARCH AND INNOVATION, 2015, : 1059 - 1069
  • [10] Mitigating Insecure Outputs in Large Language Models(LLMs): A Practical Educational Module
    Barek, Md Abdul
    Rahman, Md Mostafizur
    Akter, Mst Shapna
    Riad, A. B. M. Kamrul Islam
    Rahman, Md Abdur
    Shahriar, Hossain
    Rahman, Akond
    Wu, Fan
    2024 IEEE 48TH ANNUAL COMPUTERS, SOFTWARE, AND APPLICATIONS CONFERENCE, COMPSAC 2024, 2024, : 2424 - 2429