Smaller Language Models are capable of selecting Instruction-Tuning Training Data for Larger Language Models

被引:0
|
作者
Mekala, Dheeraj [1 ]
Nguyen, Alex [1 ]
Shang, Jingbo [1 ,2 ]
机构
[1] Univ Calif San Diego, Dept Comp Sci & Engn, La Jolla, CA 92093 USA
[2] Univ Calif San Diego, Halicioglu Data Sci Inst, La Jolla, CA 92093 USA
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Instruction-tuning language models has become a crucial step in aligning them for general use. Typically, this process involves extensive training on large datasets, incurring high training costs. In this paper, we introduce a novel training data selection based on the learning percentage of the samples. We assert that current language models possess the capability to autonomously select high-quality training data, leading to comparable or improved performance compared to training on the entire dataset. Our experiments span different-sized models, revealing that this characteristic holds for models ranging from 1B (small) to 13B (large) in size. Moreover, we demonstrate an interesting finding that the data hardness transfers across model sizes, and a smaller 350M model can effectively curate high-quality training data with hard samples for a larger 13B model, resulting in an equally or superior instructiontuned model compared to training on the complete dataset. Utilizing open-sourced OPT and Llama-2 models up to 13B in size, two publicly available instruction-tuning training datasets and evaluated by both automatic metrics & humans, our paper introduces a novel approach to training data selection, showcasing a more efficient alternative.
引用
收藏
页码:10456 / 10470
页数:15
相关论文
共 50 条
  • [31] Smaller But Better: Unifying Layout Generation with Smaller Large Language Models
    Zhang, Peirong
    Zhang, Jiaxin
    Cao, Jiahuan
    Li, Hongliang
    Jin, Lianwen
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2025,
  • [32] InstructBLIP: Towards General-purpose Vision-Language Models with Instruction Tuning
    Dai, Wenliang
    Li, Junnan
    Li, Dongxu
    Tiong, Anthony Meng Huat
    Zhao, Junqi
    Wang, Weisheng
    Li, Boyang
    Fung, Pascale
    Hoi, Steven
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [33] Generating Training Data with Language Models: Towards Zero-Shot Language Understanding
    Meng, Yu
    Huang, Jiaxin
    Zhang, Yu
    Han, Jiawei
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [34] Scaling Instruction-Finetuned Language Models
    Chung, Hyung Won
    Hou, Le
    Longpre, Shayne
    Zoph, Barret
    Tai, Yi
    Fedus, William
    Li, Yunxuan
    Wang, Xuezhi
    Dehghani, Mostafa
    Brahma, Siddhartha
    Webson, Albert
    Gu, Shixiang Shane
    Dai, Zhuyun
    Suzgun, Mirac
    Chen, Xinyun
    Chowdhery, Aakanksha
    Castro-Ros, Alex
    Pellat, Marie
    Robinson, Kevin
    Valter, Dasha
    Narang, Sharan
    Mishra, Gaurav
    Yu, Adams
    Zhao, Vincent
    Huang, Yanping
    Dai, Andrew
    Yu, Hongkun
    Petrov, Slav
    Chi, Ed H.
    Dean, Jeff
    Devlin, Jacob
    Roberts, Adam
    Zhou, Denny
    Le, Quoc, V
    Wei, Jason
    JOURNAL OF MACHINE LEARNING RESEARCH, 2024, 25
  • [35] Deduplicating Training Data Mitigates Privacy Risks in Language Models
    Kandpal, Nikhil
    Wallace, Eric
    Raffel, Colin
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022, : 10697 - 10707
  • [36] NEWTON: Are Large Language Models Capable of Physical Reasoning?
    Wang, Yi Ru
    Du, Jiafei
    Fox, Dieter
    Srinivasa, Siddhartha
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 9743 - 9758
  • [37] Staged Training for Transformer Language Models
    Shen, Sheng
    Walsh, Pete
    Keutzer, Kurt
    Dodge, Jesse
    Peters, Matthew
    Beltagy, Iz
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [38] ACOUSTICALLY DISCRIMINATIVE TRAINING FOR LANGUAGE MODELS
    Kurata, Gakuto
    Itoh, Nobuyasu
    Nishimura, Masafumi
    2009 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS 1- 8, PROCEEDINGS, 2009, : 4717 - 4720
  • [39] A Comparative Analysis of Instruction Fine-Tuning Large Language Models for Financial Text Classification
    Fatemi, Sorouralsadat
    Hu, Yuheng
    Mousavi, Maryam
    ACM TRANSACTIONS ON MANAGEMENT INFORMATION SYSTEMS, 2025, 16 (01)
  • [40] Discovering Photoswitchable Molecules for Drug Delivery with Large Language Models and Chemist Instruction Training
    Hu, Junjie
    Wu, Peng
    Li, Yulin
    Li, Qi
    Wang, Shiyi
    Liu, Yang
    Qian, Kun
    Yang, Guang
    PHARMACEUTICALS, 2024, 17 (10)