Smaller Language Models are capable of selecting Instruction-Tuning Training Data for Larger Language Models

被引:0
|
作者
Mekala, Dheeraj [1 ]
Nguyen, Alex [1 ]
Shang, Jingbo [1 ,2 ]
机构
[1] Univ Calif San Diego, Dept Comp Sci & Engn, La Jolla, CA 92093 USA
[2] Univ Calif San Diego, Halicioglu Data Sci Inst, La Jolla, CA 92093 USA
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Instruction-tuning language models has become a crucial step in aligning them for general use. Typically, this process involves extensive training on large datasets, incurring high training costs. In this paper, we introduce a novel training data selection based on the learning percentage of the samples. We assert that current language models possess the capability to autonomously select high-quality training data, leading to comparable or improved performance compared to training on the entire dataset. Our experiments span different-sized models, revealing that this characteristic holds for models ranging from 1B (small) to 13B (large) in size. Moreover, we demonstrate an interesting finding that the data hardness transfers across model sizes, and a smaller 350M model can effectively curate high-quality training data with hard samples for a larger 13B model, resulting in an equally or superior instructiontuned model compared to training on the complete dataset. Utilizing open-sourced OPT and Llama-2 models up to 13B in size, two publicly available instruction-tuning training datasets and evaluated by both automatic metrics & humans, our paper introduces a novel approach to training data selection, showcasing a more efficient alternative.
引用
收藏
页码:10456 / 10470
页数:15
相关论文
共 50 条
  • [1] An Empirical Study of Instruction-tuning Large Language Models in Chinese
    Si, Qingyi
    Wang, Tong
    Lin, Zheng
    Zhang, Xu
    Cao, Yanan
    Wang, Weiping
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 4086 - 4107
  • [2] JMedLoRA:Medical Domain Adaptation on Japanese Large Language Models using Instruction-tuning
    Sukeda, Issey
    Suzuki, Masahiro
    Kodera, Satoshi
    Sakaji, Hiroki
    arXiv, 2023,
  • [3] Exploring the Benefits of Training Expert Language Models over Instruction Tuning
    Jang, Joel
    Kim, Seungone
    Ye, Seonghyeon
    Kim, Doyoung
    Logeswaran, Lajanugen
    Lee, Moontae
    Lee, Kyungjae
    Seo, Minjoon
    Proceedings of Machine Learning Research, 2023, 202 : 14702 - 14729
  • [4] Exploring the Benefits of Training Expert Language Models over Instruction Tuning
    Jang, Joel
    Kim, Seungone
    Ye, Seonghyeon
    Kim, Doyoung
    Logeswaran, Lajanugen
    Lee, Moontae
    Lee, Kyungjae
    Seo, Minjoon
    arXiv, 2023,
  • [5] EcomGPT: Instruction-Tuning Large Language Models with Chain-of-Task Tasks for E-commerce
    Li, Yangning
    Ma, Shirong
    Wang, Xiaobin
    Huang, Shen
    Jiang, Chengyue
    Zheng, Hai-Tao
    Xie, Pengjun
    Huang, Fei
    Jiang, Yong
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 18582 - 18590
  • [6] Empowering Legal Citation Recommendation via Efficient Instruction-Tuning of Pre-trained Language Models
    Wang, Jie
    Bansal, Kanha
    Arapakis, Ioannis
    Ge, Xuri
    Jose, Joemon M.
    ADVANCES IN INFORMATION RETRIEVAL, ECIR 2024, PT I, 2024, 14608 : 310 - 324
  • [7] Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes
    Hsieh, Cheng-Yu
    Li, Chun-Liang
    Yeh, Chih-Kuan
    Nakhost, Hootan
    Fujii, Yasuhisa
    Ratner, Alexander
    Krishna, Ranjay
    Lee, Chen-Yu
    Pfister, Tomas
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 8003 - 8017
  • [8] OCTOPACK: INSTRUCTION TUNING CODE LARGE LANGUAGE MODELS
    Muennighoff, Niklas
    Liu, Qian
    Zebaze, Armel
    Zheng, Qinkai
    Hui, Binyuan
    Zhuo, Terry Yue
    Singh, Swayam
    Tang, Xiangru
    von Werra, Leandro
    Longpre, Shayne
    arXiv, 2023,
  • [9] GraphGPT: Graph Instruction Tuning for Large Language Models
    Tang, Jiabin
    Yang, Yuhao
    Wei, Wei
    Shi, Lei
    Su, Lixin
    Cheng, Suqi
    Yin, Dawei
    Huang, Chao
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 491 - 500
  • [10] BioInstruct: instruction tuning of large language models for biomedical natural language processing
    Tran, Hieu
    Yang, Zhichao
    Yao, Zonghai
    Yu, Hong
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024, 31 (09) : 1821 - 1832