Unifying Structure Reasoning and Language Pre-Training for Complex Reasoning Tasks

被引:2
|
作者
Wang, Siyuan [1 ]
Wei, Zhongyu [1 ,2 ]
Xu, Jiarong [3 ]
Li, Taishan [4 ]
Fan, Zhihao [1 ]
机构
[1] Fudan Univ, Sch Data Sci, Shanghai 200433, Peoples R China
[2] Fudan Univ, Res Inst Intelligent & Complex Syst, Shanghai 200433, Peoples R China
[3] Fudan Univ, Sch Management, Shanghai 200433, Peoples R China
[4] ShanghaiTech Univ, Sch Informat Sci & Technol, Shanghai 201210, Peoples R China
基金
中国国家自然科学基金;
关键词
Cognition; Task analysis; Semantics; Films; Speech processing; Context modeling; Data models; Structure reasoning skill; language model pre-training; complex reasoning;
D O I
10.1109/TASLP.2023.3325973
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Recent pre-trained language models (PLMs) equipped with foundation reasoning skills have shown remarkable performance on downstream complex tasks. However, the significant structure reasoning skill has been rarely studied, which involves modeling implicit structure information within the text and performing explicit logical reasoning over them to deduce the conclusion. This paper proposes a unified learning framework that combines explicit structure reasoning and language pre-training to endow PLMs with the structure reasoning skill. It first identifies several elementary structures within contexts to construct structured queries and performs step-by-step reasoning along the queries to identify the answer entity. The fusion of textual semantics and structure reasoning is achieved by using contextual representations learned by PLMs to initialize the representation space of structures, and performing stepwise reasoning on this semantic representation space. Experimental results on four datasets demonstrate that the proposed model achieves significant improvements in complex reasoning tasks involving diverse structures, and shows transferability to downstream tasks with limited training data and effectiveness for complex reasoning of KGs modality.
引用
收藏
页码:1586 / 1595
页数:10
相关论文
共 50 条
  • [1] Unifying Structure Reasoning and Language Pre-Training for Complex Reasoning Tasks
    Wang, Siyuan
    Wei, Zhongyu
    Xu, Jiarong
    Li, Taishan
    Fan, Zhihao
    IEEE/ACM Transactions on Audio Speech and Language Processing, 2024, 32 : 1586 - 1595
  • [2] Pre-training Language Models for Comparative Reasoning
    Yu, Mengxia
    Zhang, Zhihan
    Yu, Wenhao
    Jiang, Meng
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 12421 - 12433
  • [3] LogiGAN: Learning Logical Reasoning via Adversarial Pre-training
    Pi, Xinyu
    Zhong, Wanjun
    Gao, Yan
    Duan, Nan
    Lou, Jian-Guang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [4] To Boost Zero-Shot Generalization for Embodied Reasoning With Vision-Language Pre-Training
    Su, Ke
    Zhang, Xingxing
    Zhang, Siyang
    Zhu, Jun
    Zhang, Bo
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 5370 - 5381
  • [5] IDOL: Indicator-oriented Logic Pre-training for Logical Reasoning
    Xu, Zihang
    Yang, Ziqing
    Cui, Yiming
    Wang, Shijin
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 8099 - 8111
  • [6] Pre-training Is (Almost) All You Need: An Application to Commonsense Reasoning
    Tamborrino, Alexandre
    Pellicano, Nicola
    Pannier, Baptiste
    Voitot, Pascal
    Naudin, Louise
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 3878 - 3887
  • [7] Superpixel semantics representation and pre-training for vision-language tasks
    Zhang, Siyu
    Chen, Yeming
    Sun, Yaoru
    Wang, Fang
    Yang, Jun
    Bai, Lizhi
    Gao, Shangce
    NEUROCOMPUTING, 2025, 615
  • [8] Cross-modality interaction reasoning for enhancing vision-language pre-training in image-text retrieval
    Yao, Tao
    Peng, Shouyong
    Wang, Lili
    Li, Ying
    Sun, Yujuan
    APPLIED INTELLIGENCE, 2024, 54 (23) : 12230 - 12245
  • [9] Subgoal Search For Complex Reasoning Tasks
    Czechowski, Konrad
    Odrzygozdz, Tomasz
    Zbysinski, Marek
    Zawalski, Michal
    Olejnik, Krzysztof
    Wu, Yuhuai
    Kucinski, Lukasz
    Milos, Piotr
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [10] Table Pre-training: A Survey on Model Architectures, Pre-training Objectives, and Downstream Tasks
    Dong, Haoyu
    Cheng, Zhoujun
    He, Xinyi
    Zhou, Mengyu
    Zhou, Anda
    Zhou, Fan
    Liu, Ao
    Han, Shi
    Zhang, Dongmei
    PROCEEDINGS OF THE THIRTY-FIRST INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2022, 2022, : 5426 - 5435