Large Language Models are Versatile Decomposers: Decomposing Evidence and Questions for Table-based Reasoning

被引:11
|
作者
Ye, Yunhu [1 ,4 ]
Hui, Binyuan [2 ]
Yang, Min [3 ]
Li, Binhua [2 ]
Huang, Fei [2 ]
Li, Yongbin [2 ]
机构
[1] Univ Sci & Technol China, Hefei, Peoples R China
[2] DAMO Acad, Alibaba Grp, Hangzhou, Peoples R China
[3] Chinese Acad Sci, SIAT, Shenzhen, Peoples R China
[4] Chinese Acad Sci, Shenzhen Inst Adv Technol SIAT, Shenzhen, Peoples R China
关键词
Table-based reasoning; Large language models; Pre-trained language models;
D O I
10.1145/3539618.3591708
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Table-based reasoning has shown remarkable progress in a wide range of table-based tasks. It is a challenging task, which requires reasoning over both free-form natural language (NL) questions and (semi-)structured tabular data. However, previous table-based reasoning solutions usually suffer from significant performance degradation on "huge" evidence (tables). In addition, most existing methods struggle to reason over complex questions since the essential information is scattered in different places. To alleviate the above challenges, we exploit large language models (LLMs) as decomposers for effective table-based reasoning, which (i) decompose huge evidence (a huge table) into sub-evidence (a small table) to mitigate the interference of useless information for table reasoning, and (ii) decompose a complex question into simpler sub-questions for text reasoning. First, we use a powerful LLM to decompose the evidence involved in the current question into the sub-evidence that retains the relevant information and excludes the remaining irrelevant information from the "huge" evidence. Second, we propose a novel "parsing-execution-filling" strategy to decompose a complex question into simper step-by-step sub-questions by generating intermediate SQL queries as a bridge to produce numerical and logical sub-questions with a powerful LLM. Finally, we leverage the decomposed sub-evidence and sub-questions to get the final answer with a few in-context prompting examples. Extensive experiments on three benchmark datasets (TabFact, WikiTableQuestion, and FetaQA) demonstrate that our method achieves significantly better results than competitive baselines for table-based reasoning. Notably, our method outperforms human performance for the first time on the TabFact dataset. In addition to impressive overall performance, our method also has the advantage of interpretability, where the returned results are to some extent tractable with the generated sub-evidence and sub-questions. For reproducibility, we release our source code and data at: https://github.com/AlibabaResearch/DAMO-ConvAI.
引用
收藏
页码:174 / 184
页数:11
相关论文
共 50 条
  • [21] Emergent analogical reasoning in large language models
    Webb, Taylor
    Holyoak, Keith J.
    Lu, Hongjing
    NATURE HUMAN BEHAVIOUR, 2023, 7 (09) : 1526 - 1541
  • [22] DialCoT Meets PPO: Decomposing and Exploring Reasoning Paths in Smaller Language Models
    Han, Chengcheng
    Du, Xiaowei
    Zhang, Che
    Lian, Yixin
    Li, Xiang
    Gao, Ming
    Wang, Baoyuan
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 8055 - 8068
  • [23] Large Language Models are Complex Table Parsers
    Zhao, Bowen
    Ji, Changkai
    Zhang, Yuejie
    He, Wen
    Wang, Yingwen
    Wang, Qing
    Feng, Rui
    Zhang, Xiaobo
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 14786 - 14802
  • [24] MindMap: Constructing Evidence Chains for Multi-Step Reasoning in Large Language Models
    Wu, Yangyu
    Han, Xu
    Song, Wei
    Cheng, Miaomiao
    Li, Fei
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 19270 - 19278
  • [25] Table-based FET model assembled from small-signal models
    Wei, CJ
    Tkachenko, YA
    Bartle, D
    1998 IEEE RADIO AND WIRELESS CONFERENCE PROCEEDINGS - RAWCON 98, 1998, : 355 - 358
  • [26] Large Language Models for Mathematical Reasoning: Progresses and Challenges
    Ahn, Janice
    Verma, Rishu
    Lou, Renze
    Zhang, Rui
    Yin, Wenpeng
    PROCEEDINGS OF THE 18TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: STUDENT RESEARCH WORKSHOP, 2024, : 225 - 237
  • [27] The use of large language models as scaffolds for proleptic reasoning
    Olya Kudina
    Brian Ballsun-Stanton
    Mark Alfano
    Asian Journal of Philosophy, 4 (1):
  • [28] The Impact of Reasoning Step Length on Large Language Models
    Jin, Mingyu
    Yu, Qinkai
    Dong, Shu
    Zhao, Haiyan
    Hua, Wenyue
    Meng, Yanda
    Zhang, Yongfeng
    Du, Mengnan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 1830 - 1842
  • [29] TRAM: Benchmarking Temporal Reasoning for Large Language Models
    Wang, Yuqing
    Zhao, Yun
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 6389 - 6415
  • [30] EconNLI: Evaluating Large Language Models on Economics Reasoning
    Guo, Yue
    Yang, Yi
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 982 - 994