Smaller Language Models are capable of selecting Instruction-Tuning Training Data for Larger Language Models

被引:0
|
作者
Mekala, Dheeraj [1 ]
Nguyen, Alex [1 ]
Shang, Jingbo [1 ,2 ]
机构
[1] Univ Calif San Diego, Dept Comp Sci & Engn, La Jolla, CA 92093 USA
[2] Univ Calif San Diego, Halicioglu Data Sci Inst, La Jolla, CA 92093 USA
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Instruction-tuning language models has become a crucial step in aligning them for general use. Typically, this process involves extensive training on large datasets, incurring high training costs. In this paper, we introduce a novel training data selection based on the learning percentage of the samples. We assert that current language models possess the capability to autonomously select high-quality training data, leading to comparable or improved performance compared to training on the entire dataset. Our experiments span different-sized models, revealing that this characteristic holds for models ranging from 1B (small) to 13B (large) in size. Moreover, we demonstrate an interesting finding that the data hardness transfers across model sizes, and a smaller 350M model can effectively curate high-quality training data with hard samples for a larger 13B model, resulting in an equally or superior instructiontuned model compared to training on the complete dataset. Utilizing open-sourced OPT and Llama-2 models up to 13B in size, two publicly available instruction-tuning training datasets and evaluated by both automatic metrics & humans, our paper introduces a novel approach to training data selection, showcasing a more efficient alternative.
引用
收藏
页码:10456 / 10470
页数:15
相关论文
共 50 条
  • [21] Advancing entity recognition in biomedicine via instruction tuning of large language models
    Keloth, Vipina K.
    Hu, Yan
    Xie, Qianqian
    Peng, Xueqing
    Wang, Yan
    Zheng, Andrew
    Selek, Melih
    Raja, Kalpana
    Wei, Chih Hsuan
    Jin, Qiao
    Lu, Zhiyong
    Chen, Qingyu
    Xu, Hua
    BIOINFORMATICS, 2024, 40 (04)
  • [22] Reflective Instruction Tuning: Mitigating Hallucinations in Large Vision-Language Models
    Zhang, Jinrui
    Wang, Teng
    Zhang, Haigang
    Lu, Ping
    Zheng, Feng
    COMPUTER VISION - ECCV 2024, PT XXXVII, 2025, 15095 : 196 - 213
  • [23] Instruction Tuning Large Language Models for Multimodal Relation Extraction Using LoRA
    Li, Zou
    Pang, Ning
    Zhao, Xiang
    WEB INFORMATION SYSTEMS AND APPLICATIONS, WISA 2024, 2024, 14883 : 364 - 376
  • [24] WaveCoder: Widespread And Versatile Enhancement For Code Large Language Models By Instruction Tuning
    Yu, Zhaojian
    Zhang, Xin
    Shang, Ning
    Huang, Yangyu
    Xu, Can
    Zhao, Yishujie
    Hu, Wenxiang
    Yin, Qiufeng
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 5140 - 5153
  • [25] Distilling Reasoning Capabilities into Smaller Language Models
    Shridhar, Kumar
    Stolfo, Alessandro
    Sachan, Mrinmaya
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 7059 - 7073
  • [26] Are Large Language Models Capable of Causal Reasoning for Sensing Data Analysis?
    Hu, Zhizhang
    Zhang, Yue
    Rossi, Ryan
    Yu, Tong
    Kim, Sungchul
    Pan, Shijia
    PROCEEDINGS OF THE 2024 WORKSHOP ON EDGE AND MOBILE FOUNDATION MODELS, EDGEFM 2024, 2024, : 24 - 29
  • [27] Extracting Training Data from Large Language Models
    Carlini, Nicholas
    Tramer, Florian
    Wallace, Eric
    Jagielski, Matthew
    Herbert-Voss, Ariel
    Lee, Katherine
    Roberts, Adam
    Brown, Tom
    Song, Dawn
    Erlingsson, Ulfar
    Oprea, Alina
    Raffel, Colin
    PROCEEDINGS OF THE 30TH USENIX SECURITY SYMPOSIUM, 2021, : 2633 - 2650
  • [28] Deduplicating Training Data Makes Language Models Better
    Lee, Katherine
    Ippolito, Daphne
    Nystrom, Andrew
    Zhang, Chiyuan
    Eck, Douglas
    Callison-Burch, Chris
    Carlini, Nicholas
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 8424 - 8445
  • [29] LAMM: Language-Assisted Multi-Modal Instruction-Tuning Dataset, Framework, and Benchmark
    Yin, Zhenfei
    Wang, Jiong
    Cao, Jianjian
    Shi, Zhelun
    Liu, Dingning
    Li, Mukai
    Huang, Xiaoshui
    Wang, Zhiyong
    Sheng, Lu
    Bai, Lei
    Shao, Jing
    Ouyang, Wanli
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [30] Small Language Models Fine-tuned to Coordinate Larger Language Models improve Complex Reasoning
    Juneja, Gurusha
    Dutta, Subhabrata
    Chakrabarti, Soumen
    Manchhanda, Sunny
    Chakraborty, Tanmoy
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 3675 - 3691