Accelerating materials language processing with large language models

被引:0
|
作者
Jaewoong Choi
Byungju Lee
机构
[1] Korea Institute of Science and Technology,Computational Science Research Center
来源
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Materials language processing (MLP) can facilitate materials science research by automating the extraction of structured data from research papers. Despite the existence of deep learning models for MLP tasks, there are ongoing practical issues associated with complex model architectures, extensive fine-tuning, and substantial human-labelled datasets. Here, we introduce the use of large language models, such as generative pretrained transformer (GPT), to replace the complex architectures of prior MLP models with strategic designs of prompt engineering. We find that in-context learning of GPT models with few or zero-shots can provide high performance text classification, named entity recognition and extractive question answering with limited datasets, demonstrated for various classes of materials. These generative models can also help identify incorrect annotated data. Our GPT-based approach can assist material scientists in solving knowledge-intensive MLP tasks, even if they lack relevant expertise, by offering MLP guidelines applicable to any materials science domain. In addition, the outcomes of GPT models are expected to reduce the workload of researchers, such as manual labelling, by producing an initial labelling set and verifying human-annotations.
引用
收藏
相关论文
共 50 条
  • [1] Accelerating materials language processing with large language models
    Choi, Jaewoong
    Lee, Byungju
    [J]. COMMUNICATIONS MATERIALS, 2024, 5 (01)
  • [2] Accelerating Pharmacovigilance using Large Language Models
    Prakash, Mukkamala Venkata Sai
    Parab, Ganesh
    Veeramalla, Meghana
    Reddy, Siddartha
    Varun, V.
    Gopalakrishnan, Saisubramaniam
    Pagidipally, Vishal
    Vaddina, Vishal
    [J]. PROCEEDINGS OF THE 17TH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, WSDM 2024, 2024, : 1182 - 1183
  • [3] Natural language processing in the era of large language models
    Zubiaga, Arkaitz
    [J]. FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2024, 6
  • [4] Robustness of GPT Large Language Models on Natural Language Processing Tasks
    Xuanting, Chen
    Junjie, Ye
    Can, Zu
    Nuo, Xu
    Tao, Gui
    Qi, Zhang
    [J]. Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2024, 61 (05): : 1128 - 1142
  • [5] Unlocking the Capabilities of Large Language Models for Accelerating Drug Development
    Anderson, Wes
    Braun, Ian
    Bhatnagar, Roopal
    Romero, Klaus
    Walls, Ramona
    Schito, Marco
    Podichetty, Jagdeep T.
    [J]. CLINICAL PHARMACOLOGY & THERAPEUTICS, 2024, 116 (01) : 38 - 41
  • [6] BioInstruct: instruction tuning of large language models for biomedical natural language processing
    Tran, Hieu
    Yang, Zhichao
    Yao, Zonghai
    Yu, Hong
    [J]. JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024, 31 (09) : 1821 - 1832
  • [7] Research and Exploration on Chinese Natural Language Processing in Era of Large Language Models
    大模型时代下的汉语自然语言处理研究与探索
    [J]. Xi, Xuefeng (xfxi@mail.usts.edu.cn), 2025, 61 (01) : 80 - 97
  • [8] Accelerating Contextualization in AI Large Language Models Using Vector Databases
    Bin Tareaf, Raad
    AbuJarour, Mohammed
    Engelman, Tom
    Liermann, Philipp
    Klotz, Jesse
    [J]. 38TH INTERNATIONAL CONFERENCE ON INFORMATION NETWORKING, ICOIN 2024, 2024, : 316 - 321
  • [9] Exploring large language models for microstructure evolution in materials
    Satpute, Prathamesh
    Tiwari, Saurabh
    Gupta, Maneet
    Ghosh, Supriyo
    [J]. MATERIALS TODAY COMMUNICATIONS, 2024, 40
  • [10] Materials science in the era of large language models: a perspective
    Lei, Ge
    Docherty, Ronan
    Cooper, Samuel J.
    [J]. DIGITAL DISCOVERY, 2024, 3 (07):