Accelerating materials language processing with large language models

被引:0
|
作者
Jaewoong Choi
Byungju Lee
机构
[1] Korea Institute of Science and Technology,Computational Science Research Center
来源
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Materials language processing (MLP) can facilitate materials science research by automating the extraction of structured data from research papers. Despite the existence of deep learning models for MLP tasks, there are ongoing practical issues associated with complex model architectures, extensive fine-tuning, and substantial human-labelled datasets. Here, we introduce the use of large language models, such as generative pretrained transformer (GPT), to replace the complex architectures of prior MLP models with strategic designs of prompt engineering. We find that in-context learning of GPT models with few or zero-shots can provide high performance text classification, named entity recognition and extractive question answering with limited datasets, demonstrated for various classes of materials. These generative models can also help identify incorrect annotated data. Our GPT-based approach can assist material scientists in solving knowledge-intensive MLP tasks, even if they lack relevant expertise, by offering MLP guidelines applicable to any materials science domain. In addition, the outcomes of GPT models are expected to reduce the workload of researchers, such as manual labelling, by producing an initial labelling set and verifying human-annotations.
引用
收藏
相关论文
共 50 条
  • [21] Large Language Models in der WissenschaftLarge language models in science
    Karl-Friedrich Kowalewski
    Severin Rodler
    [J]. Die Urologie, 2024, 63 (9) : 860 - 866
  • [22] MaScQA: investigating materials science knowledge of large language models
    Zaki, Mohd
    Jayadeva
    Mausam
    Krishnan, N. M. Anoop
    [J]. DIGITAL DISCOVERY, 2024, 3 (02): : 313 - 327
  • [23] THE ROLE OF MODELS OF LANGUAGE PROCESSING IN REHABILITATION OF LANGUAGE IMPAIRMENTS
    HILLIS, AE
    [J]. APHASIOLOGY, 1993, 7 (01) : 5 - 26
  • [24] COMPUTATIONAL MODELS OF LANGUAGE PROCESSING
    STABLER, EP
    [J]. BEHAVIORAL AND BRAIN SCIENCES, 1986, 9 (03) : 550 - 551
  • [25] Neurocomputational Models of Language Processing
    Hale, John T.
    Campanelli, Luca
    Li, Jixing
    Bhattasali, Shohini
    Pallier, Christophe
    Brennan, Jonathan R.
    [J]. ANNUAL REVIEW OF LINGUISTICS, 2022, 8 : 427 - 446
  • [26] Models and modularity in language processing
    Dell, GS
    [J]. PROCEEDINGS OF THE EIGHTEENTH ANNUAL CONFERENCE OF THE COGNITIVE SCIENCE SOCIETY, 1996, : 47 - 47
  • [27] Accelerated evidence synthesis in orthopaedics—the roles of natural language processing, expert annotation and large language models
    Bálint Zsidai
    Janina Kaarre
    Ann-Sophie Hilkert
    Eric Narup
    Eric Hamrin Senorski
    Alberto Grassi
    Olufemi R. Ayeni
    Volker Musahl
    Christophe Ley
    Elmar Herbst
    Michael T. Hirschmann
    Sebastian Kopf
    Romain Seil
    Thomas Tischer
    Kristian Samuelsson
    Robert Feldt
    [J]. Journal of Experimental Orthopaedics, 10
  • [28] Comparison of large language models and traditional natural language processing techniques in predicting arteriovenous fistula failure
    Lama, Suman
    Zhang, Hanjie
    Monaghan, Caitlin
    Bellocchio, Francesco
    Chaudhuri, Sheetal
    Neri, Luca
    Usvyat, Len
    [J]. NEPHROLOGY DIALYSIS TRANSPLANTATION, 2024, 39 : I1303 - I1304
  • [29] Recent Advances in Natural Language Processing via Large Pre-trained Language Models: A Survey
    Min, Bonan
    Ross, Hayley
    Sulem, Elior
    Ben Veyseh, Amir Pouran
    Nguyen, Thien Huu
    Sainz, Oscar
    Agirre, Eneko
    Heintz, Ilana
    Roth, Dan
    [J]. ACM COMPUTING SURVEYS, 2024, 56 (02)
  • [30] Evaluation of large language models in natural language processing of PET/CT free-text reports
    Bradshaw, Tyler
    Cho, Steve
    [J]. JOURNAL OF NUCLEAR MEDICINE, 2021, 62