Small Language Model Can Self-Correct

被引:0
|
作者
Han, Haixia [1 ]
Liang, Jiaqing [2 ]
Shi, Jie [3 ]
He, Qianyu [3 ]
Xiao, Yanghua [1 ,3 ]
机构
[1] East China Normal Univ, Shanghai Inst AI Educ & Sch Comp Sci & Technol, Shanghai, Peoples R China
[2] Fudan Univ, Sch Data Sci, Shanghai, Peoples R China
[3] Fudan Univ, Shanghai Key Lab Data Sci, Sch Comp Sci, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Generative Language Models (LMs) such as ChatGPT have exhibited remarkable performance across various downstream tasks. Nevertheless, one of their most prominent drawbacks is generating inaccurate or false information with a confident tone. Previous studies have devised sophisticated pipelines and prompts to induce large LMs to exhibit the capability for self-correction. However, large LMs are explicitly prompted to verify and modify their answers separately rather than completing all steps spontaneously like humans. Moreover, these complex prompts are extremely challenging for small LMs to follow. In this paper, we introduce the Intrinsic Self-Correction (ISC) in generative language models, aiming to correct the initial output of LMs in a self-triggered manner, even for those small LMs with 6 billion parameters. Specifically, we devise a pipeline for constructing self-correction data and propose Partial Answer Masking (PAM), aiming to endow the model with the capability for intrinsic self-correction through fine-tuning. We conduct experiments using LMs with parameters sizes ranging from 6 billion to 13 billion in two tasks, including commonsense reasoning and factual knowledge reasoning. Our experiments demonstrate that the outputs generated using ISC outperform those generated without self-correction. We believe that the output quality of even small LMs can be further improved by empowering them with the ability to intrinsic self-correct.
引用
收藏
页码:18162 / 18170
页数:9
相关论文
共 50 条
  • [1] Small Language Models Need Strong Verifiers to Self-Correct Reasoning
    Zhang, Yunxiang
    Khalifa, Muhammad
    Logeswaran, Lajanugen
    Kim, Jaekyeom
    Lee, Moontae
    Lee, Honglak
    Wang, Lu
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 15637 - 15653
  • [2] Can the behavioral sciences self-correct? A social epistemic study
    Romero, Felipe
    STUDIES IN HISTORY AND PHILOSOPHY OF SCIENCE, 2016, 60 : 55 - 69
  • [3] To err is human; to self-correct is to learn
    Forbers, S
    Poparad, MA
    McBride, M
    READING TEACHER, 2004, 57 (06): : 566 - 572
  • [4] An Improved Self-correct Algorithm for Pavement Texture
    He Huayang
    Dou Guangwu
    Zhang Jinning
    LIDAR IMAGING DETECTION AND TARGET RECOGNITION 2017, 2017, 10605
  • [5] Teaching Early Readers to Self-Monitor and Self-Correct
    Pratt, Sharon M.
    Urbanowski, Melena
    READING TEACHER, 2016, 69 (05): : 559 - 567
  • [6] When to Self-Correct Spelling Words: A Systematic Replication
    Sheila R. Alber
    Suzanne E. Walshe
    Journal of Behavioral Education, 2004, 13 (1) : 51 - 66
  • [7] An Example of Psychological Science's Failure to Self-Correct
    Kelley, Lance P.
    Blashfield, Roger K.
    REVIEW OF GENERAL PSYCHOLOGY, 2009, 13 (02) : 122 - 129
  • [8] When to self-correct?: A comparison of two procedures on spelling performance
    Morton W.L.
    Heward W.L.
    Alber S.R.
    Journal of Behavioral Education, 1998, 8 (3) : 321 - 335
  • [9] NEW SOFTWARE TO HELP EFL STUDENTS SELF-CORRECT THEIR WRITING
    Lawley, Jim
    LANGUAGE LEARNING & TECHNOLOGY, 2015, 19 (01): : 23 - 33
  • [10] Silicon synapses self-correct for both mismatch and design inhomogeneities
    Bamford, S. A.
    Murray, A. F.
    Willshaw, D. J.
    ELECTRONICS LETTERS, 2012, 48 (07) : 360 - U88