EcomGPT: Instruction-Tuning Large Language Models with Chain-of-Task Tasks for E-commerce

被引:0
|
作者
Li, Yangning [1 ,4 ]
Ma, Shirong [1 ]
Wang, Xiaobin [3 ]
Huang, Shen [3 ]
Jiang, Chengyue [2 ]
Zheng, Hai-Tao [1 ,4 ]
Xie, Pengjun [3 ]
Huang, Fei [3 ]
Jiang, Yong [3 ]
机构
[1] Tsinghua Univ, SIGS, Beijing, Peoples R China
[2] ShanghaiTech Univ, Shanghai, Peoples R China
[3] Alibaba Grp, DAMO Acad, Hangzhou, Peoples R China
[4] PengCheng Lab, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, instruction-following Large Language Models (LLMs) , represented by ChatGPT, have exhibited exceptional performance in general Natural Language Processing (NLP) tasks. However, the unique characteristics of E-commerce data pose significant challenges to general LLMs. An LLM tailored specifically for E-commerce scenarios, possessing robust cross-dataset/task generalization capabilities, is a pressing necessity. To solve this issue, in this work, we proposed the first E-commerce instruction dataset EcomInstruct, with a total of 2.5 million instruction data. EcomInstruct scales up the data size and task diversity by constructing atomic tasks with E-commerce basic data types, such as product information, user reviews. Atomic tasks are defined as intermediate tasks implicitly involved in solving a final task, which we also call Chain-of-Task tasks. We developed EcomGPT with different parameter scales by training the backbone model BLOOMZ with the EcomInstruct. Benefiting from the fundamental semantic understanding capabilities acquired from the Chain-of-Task tasks, EcomGPT exhibits excellent zero-shot generalization capabilities. Extensive experiments and human evaluations demonstrate that EcomGPT outperforms ChatGPT in term of cross-dataset/task generalization on E-commerce tasks. The EcomGPT will be public at https://github.com/Alibaba-NLP/EcomGPT.
引用
收藏
页码:18582 / 18590
页数:9
相关论文
共 48 条
  • [41] Conversational Recommender System and Large Language Model Are Made for Each Other in E-commerce Pre-sales Dialogue
    Liu, Yuanxing
    Zhang, Wei-Nan
    Chen, Yifan
    Zhang, Yuchi
    Bai, Haopeng
    Feng, Fan
    Cui, Hengbin
    Lie, Yongbin
    Che, Wanxiang
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 9587 - 9605
  • [42] Research on the Large-scale E-commerce Platform Development Mode Based on Oracle Database and Java']Java Programming Language
    Wang, Meiyan
    2015 3RD INTERNATIONAL CONFERENCE ON ADVANCES IN SOCIAL SCIENCE, HUMANITIES, AND MANAGEMENT, ASSHM 2015, 2015, : 1082 - 1091
  • [43] Sample Size Considerations for Fine-Tuning Large Language Models for Named Entity Recognition Tasks: Methodological Study
    Majdik, Zoltan P.
    Graham, S. Scott
    Edward, Jade C. Shiva
    Rodriguez, Sabrina N.
    Karnes, Martha S.
    Jensen, Jared T.
    Barbour, Joshua B.
    Rousseau, Justin F.
    JMIR AI, 2024, 3
  • [44] DeCoT: Debiasing Chain-of-Thought for Knowledge-Intensive Tasks in Large Language Models via Causal Intervention
    Wu, Junda
    Yu, Tong
    Chen, Xiang
    Wang, Haoliang
    Rossi, Ryan A.
    Kim, Sungchul
    Rao, Anup
    McAuley, Julian
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 14073 - 14087
  • [45] John-Arthur at SemEval-2023 Task 4: Fine-Tuning Large Language Models for Arguments Classification
    Balikas, Georgios
    17TH INTERNATIONAL WORKSHOP ON SEMANTIC EVALUATION, SEMEVAL-2023, 2023, : 1428 - 1432
  • [46] Efficient fine-tuning of small-parameter large language models for biomedical bilingual multi-task applications
    Li, Yinghong
    Yan, Yudong
    Tong, Zhuohao
    Wang, Yu
    Yang, Yinqi
    Bai, Mingze
    Pu, Dan
    Xie, Jiazheng
    Liu, Chuan
    Li, Bo
    Liu, Mingwei
    Shu, Kunxian
    APPLIED SOFT COMPUTING, 2025, 175
  • [47] CoF-CoT: Enhancing Large Language Models with Coarse-to-Fine Chain-of-Thought Prompting for Multi-domain NLU Tasks
    Nguyen, Hoang H.
    Liu, Ye
    Zhang, Chenwei
    Zhang, Tao
    Yu, Philip S.
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 12109 - 12119
  • [48] NCL-UoR at SemEval-2024 Task 8: Fine-tuning Large Language Models for Multigenerator, Multidomain, and Multilingual Machine-Generated Text Detection
    Xiong, Feng
    Markchom, Thanet
    Zheng, Ziwei
    Jung, Subin
    Ojha, Varun
    Liang, Huizhi
    PROCEEDINGS OF THE 18TH INTERNATIONAL WORKSHOP ON SEMANTIC EVALUATION, SEMEVAL-2024, 2024, : 163 - 169