Accelerating materials language processing with large language models

被引:0
|
作者
Jaewoong Choi
Byungju Lee
机构
[1] Korea Institute of Science and Technology,Computational Science Research Center
来源
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Materials language processing (MLP) can facilitate materials science research by automating the extraction of structured data from research papers. Despite the existence of deep learning models for MLP tasks, there are ongoing practical issues associated with complex model architectures, extensive fine-tuning, and substantial human-labelled datasets. Here, we introduce the use of large language models, such as generative pretrained transformer (GPT), to replace the complex architectures of prior MLP models with strategic designs of prompt engineering. We find that in-context learning of GPT models with few or zero-shots can provide high performance text classification, named entity recognition and extractive question answering with limited datasets, demonstrated for various classes of materials. These generative models can also help identify incorrect annotated data. Our GPT-based approach can assist material scientists in solving knowledge-intensive MLP tasks, even if they lack relevant expertise, by offering MLP guidelines applicable to any materials science domain. In addition, the outcomes of GPT models are expected to reduce the workload of researchers, such as manual labelling, by producing an initial labelling set and verifying human-annotations.
引用
收藏
相关论文
共 50 条
  • [31] Utilizing natural language processing and large language models in the diagnosis and prediction of infectious diseases: A systematic review
    Omar, Mahmud
    Brin, Dana
    Glicksberg, Benjamin
    Klang, Eyal
    [J]. AMERICAN JOURNAL OF INFECTION CONTROL, 2024, 52 (09) : 992 - 1001
  • [32] Fine Tuning of large language Models for Arabic Language
    Tamer, Ahmed
    Hassan, Al-Amir
    Ali, Asmaa
    Salah, Nada
    Medhat, Walaa
    [J]. 2023 20TH ACS/IEEE INTERNATIONAL CONFERENCE ON COMPUTER SYSTEMS AND APPLICATIONS, AICCSA, 2023,
  • [33] Understanding Telecom Language Through Large Language Models
    Bariah, Lina
    Zou, Hang
    Zhao, Qiyang
    Mouhouche, Belkacem
    Bader, Faouzi
    Debbah, Merouane
    [J]. IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 6542 - 6547
  • [34] Prompting Is Programming: A Query Language for Large Language Models
    Beurer-Kellner, Luca
    Fischer, Marc
    Vechev, Martin
    [J]. PROCEEDINGS OF THE ACM ON PROGRAMMING LANGUAGES-PACMPL, 2023, 7 (PLDI):
  • [35] Structuring Natural Language Requirements with Large Language Models
    Norheim, Johannes J.
    Rebentisch, Eric
    [J]. 32ND INTERNATIONAL REQUIREMENTS ENGINEERING CONFERENCE WORKSHOPS, REW 2024, 2024, : 68 - 71
  • [36] Accelerating Natural Language Processing for Applications in Pharmaceutical Research
    Torfs, Bert
    [J]. WMSCI 2008: 12TH WORLD MULTI-CONFERENCE ON SYSTEMICS, CYBERNETICS AND INFORMATICS, VOL I, PROCEEDINGS, 2008, : 145 - 149
  • [37] The role of large language models in medical image processing: a narrative review
    Tian, Dianzhe
    Jiang, Shitao
    Zhang, Lei
    Lu, Xin
    Xu, Yiyao
    [J]. QUANTITATIVE IMAGING IN MEDICINE AND SURGERY, 2024, 14 (01) : 1108 - 1121
  • [38] Imitation and Large Language Models
    Boisseau, Éloïse
    [J]. Minds and Machines, 2024, 34 (04)
  • [39] Autoformalization with Large Language Models
    Wu, Yuhuai
    Jiang, Albert Q.
    Li, Wenda
    Rabe, Markus N.
    Staats, Charles
    Jamnik, Mateja
    Szegedy, Christian
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [40] The Smallness of Large Language Models
    Denning, Peter J.
    [J]. COMMUNICATIONS OF THE ACM, 2023, 66 (09) : 24 - 27