Unveiling the power of language models in chemical research question answering

被引:0
|
作者
Xiuying Chen [1 ]
Tairan Wang [2 ]
Taicheng Guo [2 ]
Kehan Guo [3 ]
Juexiao Zhou [3 ]
Haoyang Li [2 ]
Zirui Song [2 ]
Xin Gao [1 ]
Xiangliang Zhang [2 ]
机构
[1] Mohamed bin Zayed University of Artificial Intelligence,
[2] King Abdullah University of Science and Technology,undefined
[3] University of Notre Dame,undefined
关键词
D O I
10.1038/s42004-024-01394-x
中图分类号
学科分类号
摘要
While the abilities of language models are thoroughly evaluated in areas like general domains and biomedicine, academic chemistry remains less explored. Chemical QA tools also play a crucial role in both education and research by effectively translating complex chemical information into an understandable format. Addressing this gap, we introduce ScholarChemQA, a large-scale QA dataset constructed from chemical papers. Specifically, the questions are from paper titles with a question mark, and the multi-choice answers are reasoned out based on the corresponding abstracts. This dataset reflects typical real-world challenges, including an imbalanced data distribution and a substantial amount of unlabeled data that can be potentially useful. Correspondingly, we introduce a ChemMatch model, specifically designed to effectively answer chemical questions by fully leveraging our collected data. Experiments show that Large Language Models (LLMs) still have significant room for improvement in the field of chemistry. Moreover, ChemMatch significantly outperforms recent similar-scale baselines: https://github.com/iriscxy/chemmatch.
引用
收藏
相关论文
共 50 条
  • [1] Finetuning Language Models for Multimodal Question Answering
    Zhang, Xin
    Xie, Wen
    Dai, Ziqi
    Rao, Jun
    Wen, Haokun
    Luo, Xuan
    Zhang, Meishan
    Zhang, Min
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 9420 - 9424
  • [2] Reasoning with large language models for medical question answering
    Lucas, Mary M.
    Yang, Justin
    Pomeroy, Jon K.
    Yang, Christopher C.
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024, 31 (09)
  • [3] Enhancing Biomedical Question Answering with Large Language Models
    Yang, Hua
    Li, Shilong
    Goncalves, Teresa
    INFORMATION, 2024, 15 (08)
  • [4] Review of Research Progress on Question-Answering Techniques Based on Large Language Models
    Wen, Sen
    Qian, Li
    Hu, Maodi
    Chang, Zhijun
    Data Analysis and Knowledge Discovery, 2024, 8 (06) : 16 - 29
  • [5] Efficient Question Answering Based on Language Models and Knowledge Graphs
    Li, Fengying
    Huang, Hongfei
    Dong, Rongsheng
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT IV, 2023, 14257 : 340 - 351
  • [6] Language processing and learning models for community question answering in Arabic
    Romeo, Salvatore
    Da San Martino, Giovanni
    Belinkov, Yonatan
    Barron-Cedeno, Alberto
    Eldesouki, Mohamed
    Darwish, Kareem
    Mubarak, Hamdy
    Glass, James
    Moschitti, Alessandro
    INFORMATION PROCESSING & MANAGEMENT, 2019, 56 (02) : 274 - 290
  • [7] An astronomical question answering dataset for evaluating large language models
    Jie Li
    Fuyong Zhao
    Panfeng Chen
    Jiafu Xie
    Xiangrui Zhang
    Hui Li
    Mei Chen
    Yanhao Wang
    Ming Zhu
    Scientific Data, 12 (1)
  • [8] A General Approach to Website Question Answering with Large Language Models
    Ding, Yilang
    Nie, Jiawei
    Wu, Di
    Liu, Chang
    SOUTHEASTCON 2024, 2024, : 894 - 896
  • [9] Tree -of-Reasoning Question Decomposition for Complex Question Answering with Large Language Models
    Zhang, Kun
    Zeng, Jiali
    Meng, Fandong
    Wang, Yuanzhuo
    Sun, Shiqi
    Bai, Long
    Shen, Huawei
    Zhou, Jie
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 19560 - 19568
  • [10] How Can We Know When Language Models Know? On the Calibration of Language Models for Question Answering
    Jiang, Zhengbao
    Araki, Jun
    Ding, Haibo
    Neubig, Graham
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2021, 9 (09) : 962 - 977