EcomGPT: Instruction-Tuning Large Language Models with Chain-of-Task Tasks for E-commerce

被引:0
|
作者
Li, Yangning [1 ,4 ]
Ma, Shirong [1 ]
Wang, Xiaobin [3 ]
Huang, Shen [3 ]
Jiang, Chengyue [2 ]
Zheng, Hai-Tao [1 ,4 ]
Xie, Pengjun [3 ]
Huang, Fei [3 ]
Jiang, Yong [3 ]
机构
[1] Tsinghua Univ, SIGS, Beijing, Peoples R China
[2] ShanghaiTech Univ, Shanghai, Peoples R China
[3] Alibaba Grp, DAMO Acad, Hangzhou, Peoples R China
[4] PengCheng Lab, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, instruction-following Large Language Models (LLMs) , represented by ChatGPT, have exhibited exceptional performance in general Natural Language Processing (NLP) tasks. However, the unique characteristics of E-commerce data pose significant challenges to general LLMs. An LLM tailored specifically for E-commerce scenarios, possessing robust cross-dataset/task generalization capabilities, is a pressing necessity. To solve this issue, in this work, we proposed the first E-commerce instruction dataset EcomInstruct, with a total of 2.5 million instruction data. EcomInstruct scales up the data size and task diversity by constructing atomic tasks with E-commerce basic data types, such as product information, user reviews. Atomic tasks are defined as intermediate tasks implicitly involved in solving a final task, which we also call Chain-of-Task tasks. We developed EcomGPT with different parameter scales by training the backbone model BLOOMZ with the EcomInstruct. Benefiting from the fundamental semantic understanding capabilities acquired from the Chain-of-Task tasks, EcomGPT exhibits excellent zero-shot generalization capabilities. Extensive experiments and human evaluations demonstrate that EcomGPT outperforms ChatGPT in term of cross-dataset/task generalization on E-commerce tasks. The EcomGPT will be public at https://github.com/Alibaba-NLP/EcomGPT.
引用
收藏
页码:18582 / 18590
页数:9
相关论文
共 48 条
  • [1] An Empirical Study of Instruction-tuning Large Language Models in Chinese
    Si, Qingyi
    Wang, Tong
    Lin, Zheng
    Zhang, Xu
    Cao, Yanan
    Wang, Weiping
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 4086 - 4107
  • [2] JMedLoRA:Medical Domain Adaptation on Japanese Large Language Models using Instruction-tuning
    Sukeda, Issey
    Suzuki, Masahiro
    Kodera, Satoshi
    Sakaji, Hiroki
    arXiv, 2023,
  • [3] Smaller Language Models are capable of selecting Instruction-Tuning Training Data for Larger Language Models
    Mekala, Dheeraj
    Nguyen, Alex
    Shang, Jingbo
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 10456 - 10470
  • [4] LiLiuM: eBay's Large Language Models for e-commerce
    Herold, Christian
    Kozielski, Michael
    Ekimov, Leonid
    Petrushkov, Pavel
    Vandenbussche, Pierre-Yves
    Khadivi, Shahram
    arXiv,
  • [5] Use of a large language model with instruction-tuning for reliable clinical frailty scoring
    Kee, Xiang Lee Jamie
    Sng, Gerald Gui Ren
    Lim, Daniel Yan Zheng
    Tung, Joshua Yi Min
    Abdullah, Hairil Rizal
    Chowdury, Anupama Roy
    JOURNAL OF THE AMERICAN GERIATRICS SOCIETY, 2024, 72 (12) : 3849 - 3854
  • [6] Relation labeling in product knowledge graphs with large language models for e-commerce
    Chen, Jiao
    Ma, Luyi
    Li, Xiaohan
    Xu, Jianpeng
    Cho, Jason H. D.
    Nag, Kaushiki
    Korpeoglu, Evren
    Kumar, Sushant
    Achan, Kannan
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (12) : 5725 - 5743
  • [7] OCTOPACK: INSTRUCTION TUNING CODE LARGE LANGUAGE MODELS
    Muennighoff, Niklas
    Liu, Qian
    Zebaze, Armel
    Zheng, Qinkai
    Hui, Binyuan
    Zhuo, Terry Yue
    Singh, Swayam
    Tang, Xiangru
    von Werra, Leandro
    Longpre, Shayne
    arXiv, 2023,
  • [8] GraphGPT: Graph Instruction Tuning for Large Language Models
    Tang, Jiabin
    Yang, Yuhao
    Wei, Wei
    Shi, Lei
    Su, Lixin
    Cheng, Suqi
    Yin, Dawei
    Huang, Chao
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 491 - 500
  • [9] Empowering Legal Citation Recommendation via Efficient Instruction-Tuning of Pre-trained Language Models
    Wang, Jie
    Bansal, Kanha
    Arapakis, Ioannis
    Ge, Xuri
    Jose, Joemon M.
    ADVANCES IN INFORMATION RETRIEVAL, ECIR 2024, PT I, 2024, 14608 : 310 - 324
  • [10] E-Commerce Supply Chain Models under Altruistic Preference
    Wang, Yuyan
    Yu, Zhaoqing
    Shen, Liang
    Dong, Wenquan
    MATHEMATICS, 2021, 9 (06)