Unveiling the power of language models in chemical research question answering

被引:0
|
作者
Xiuying Chen [1 ]
Tairan Wang [2 ]
Taicheng Guo [2 ]
Kehan Guo [3 ]
Juexiao Zhou [3 ]
Haoyang Li [2 ]
Zirui Song [2 ]
Xin Gao [1 ]
Xiangliang Zhang [2 ]
机构
[1] Mohamed bin Zayed University of Artificial Intelligence,
[2] King Abdullah University of Science and Technology,undefined
[3] University of Notre Dame,undefined
关键词
D O I
10.1038/s42004-024-01394-x
中图分类号
学科分类号
摘要
While the abilities of language models are thoroughly evaluated in areas like general domains and biomedicine, academic chemistry remains less explored. Chemical QA tools also play a crucial role in both education and research by effectively translating complex chemical information into an understandable format. Addressing this gap, we introduce ScholarChemQA, a large-scale QA dataset constructed from chemical papers. Specifically, the questions are from paper titles with a question mark, and the multi-choice answers are reasoned out based on the corresponding abstracts. This dataset reflects typical real-world challenges, including an imbalanced data distribution and a substantial amount of unlabeled data that can be potentially useful. Correspondingly, we introduce a ChemMatch model, specifically designed to effectively answer chemical questions by fully leveraging our collected data. Experiments show that Large Language Models (LLMs) still have significant room for improvement in the field of chemistry. Moreover, ChemMatch significantly outperforms recent similar-scale baselines: https://github.com/iriscxy/chemmatch.
引用
收藏
相关论文
共 50 条
  • [41] Zero-Shot Video Question Answering via Frozen Bidirectional Language Models
    Yang, Antoine
    Miech, Antoine
    Sivic, Josef
    Laptev, Ivan
    Schmid, Cordelia
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [42] Language independent passage retrieval for question answering
    Gómez-Soriano, JM
    Montes-y-Gómez, M
    Sanchis-Arnal, E
    Villaseñor-Pineda, L
    Rosso, P
    MICAI 2005: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2005, 3789 : 816 - 823
  • [43] A Qualitative Evaluation of Language Models on Automatic Question-Answering for COVID-19
    Oniani, David
    Wang, Yanshan
    ACM-BCB 2020 - 11TH ACM CONFERENCE ON BIOINFORMATICS, COMPUTATIONAL BIOLOGY, AND HEALTH INFORMATICS, 2020,
  • [44] Question-Answering in a Low-resourced Language: Benchmark Dataset and Models for Tigrinya
    Gaim, Fitsum
    Yang, Wonsuk
    Park, Hancheol
    Park, Jong C.
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 11857 - 11870
  • [45] End-to-End Quantum-Like Language Models with Application to Question Answering
    Zhang, Peng
    Niu, Jiabin
    Su, Zhan
    Wang, Benyou
    Ma, Liqun
    Song, Dawei
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 5666 - 5673
  • [46] A question-answering framework for automated abstract screening using large language models
    Akinseloyin, Opeoluwa
    Jiang, Xiaorui
    Palade, Vasile
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024, 31 (09)
  • [47] Research on question retrieval method for community question answering
    Yong Sun
    Junfang Song
    Xiangyu Song
    Jiazheng Hou
    Multimedia Tools and Applications, 2023, 82 : 24309 - 24325
  • [48] Finnish as source language in bilingual question answering
    Aunimo, L
    Kuuskoski, R
    Makkonen, J
    MULTILINGUAL INFORMATION ACCESS FOR TEXT, SPEECH AND IMAGES, 2005, 3491 : 482 - 493
  • [49] Natural Language Question Answering in Open Domains
    Tufis, Dan
    COMPUTER SCIENCE JOURNAL OF MOLDOVA, 2011, 19 (02) : 146 - 164
  • [50] ANNA: Enhanced Language Representation for Question Answering
    Jun, Changwook
    Jang, Hansol
    Sim, Myoseop
    Kim, Hyun
    Choi, Jooyoung
    Min, Kyungkoo
    Bae, Kyunghoon
    PROCEEDINGS OF THE 7TH WORKSHOP ON REPRESENTATION LEARNING FOR NLP, 2022, : 121 - 132