Large language model for patent concept generation

被引:0
|
作者
Ren, Runtao [1 ]
Ma, Jian [1 ]
Luo, Jianxi [2 ]
机构
[1] City Univ Hong Kong, Dept Informat Syst, Kowloon Tong, Hong Kong, Peoples R China
[2] City Univ Hong Kong, Dept Syst Engn, Kowloon Tong, Hong Kong, Peoples R China
关键词
Generative AI; Large language model; Finetuning; Patent;
D O I
10.1016/j.aei.2025.103301
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In traditional innovation practices, concept and IP generation are often iteratively integrated. Both processes demand an intricate understanding of advanced technical domain knowledge. Existing large language models (LLMs), while possessing massive pre-trained knowledge, often fall short in the innovative concept generation due to a lack of specialized knowledge necessary for the generation. To bridge this critical gap, we propose a novel knowledge finetuning (KFT) framework to endow LLM-based AI with the ability to autonomously mine, understand, and apply domain-specific knowledge and concepts for invention generation, i.e., concept and patent generation together. Our proposed PatentGPT integrates knowledge injection pre-training (KPT), domainspecific supervised finetuning (SFT), and reinforcement learning from human feedback (RLHF). Extensive evaluation shows that PatentGPT significantly outperforms the state-of-the-art models on patent-related benchmark tests. Our method not only provides new insights into data-driven innovation but also paves a new path to fine-tune LLMs for applications in the context of technology. We also discuss the managerial and policy implications of AI-generating inventions in the future.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] Large language model-based code generation for the control of construction assembly robots: A hierarchical generation approach
    Luo, Hanbin
    Wu, Jianxin
    Liu, Jiajing
    Antwi-Afari, Maxwell Fordjour
    DEVELOPMENTS IN THE BUILT ENVIRONMENT, 2024, 19
  • [42] Towards Concept-Aware Large Language Models
    Shani, Chen
    Vreeken, Jilles
    Shahaf, Dafna
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 13158 - 13170
  • [43] The Large Language Model GreekLegalRoBERTa
    Saketos, Vasileios
    Pantazi, Despina-Athanasia
    Koubarakis, Manolis
    PROCEEDINGS OF THE 13TH HELLENIC CONFERENCE ON ARTIFICIAL INTELLIGENCE, SETN 2024, 2024,
  • [44] Large language model in electrocatalysis
    Zhang, Chengyi
    Wang, Xingyu
    Wang, Ziyun
    CHINESE JOURNAL OF CATALYSIS, 2024, 59 : 7 - 14
  • [45] Game Generation via Large Language Models
    Hu, Chengpeng
    Zhao, Yunlong
    Liu, Jialin
    2024 IEEE CONFERENCE ON GAMES, COG 2024, 2024,
  • [46] Level Generation Through Large Language Models
    Todd, Graham
    Earle, Sam
    Nasir, Muhammad Umair
    Green, Michael Cerny
    Togelius, Julian
    PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON THE FOUNDATIONS OF DIGITAL GAMES, FDG 2023, 2023,
  • [47] On the Capacity of Citation Generation by Large Language Models
    Qian, Haosheng
    Fan, Yixing
    Zhang, Ruqing
    Guo, Jiafeng
    INFORMATION RETRIEVAL, CCIR 2024, 2025, 15418 : 109 - 123
  • [48] Acquisitions in a patent contest model with large and small firms
    Kleer R.
    Journal of Industry, Competition and Trade, 2009, 9 (4) : 307 - 328
  • [49] Hierarchical Concept-Driven Language Model
    Wang, Yashen
    Zhang, Huanhuan
    Liu, Zhirun
    Zhou, Qiang
    ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2021, 15 (06)
  • [50] Patent classification by fine-tuning BERT language model
    Lee, Jieh-Sheng
    Hsiang, Jieh
    WORLD PATENT INFORMATION, 2020, 61