Unifying Structure Reasoning and Language Pre-Training for Complex Reasoning Tasks

被引:0
|
作者
Wang, Siyuan [1 ]
Wei, Zhongyu [1 ,2 ]
Xu, Jiarong [3 ]
Li, Taishan [4 ]
Fan, Zhihao [1 ]
机构
[1] Fudan University, School of Data Science, Shanghai,200433, China
[2] Fudan University, Research Institute of Intelligent and Complex Systems, Shanghai,200433, China
[3] Fudan University, School of Management, Shanghai,200433, China
[4] ShanghaiTech University, School of Information Science and Technology, Shanghai,201210, China
关键词
Computational linguistics - Job analysis - Modeling languages - Personnel training - Query processing - Speech processing;
D O I
暂无
中图分类号
学科分类号
摘要
Recent pre-trained language models (PLMs) equipped with foundation reasoning skills have shown remarkable performance on downstream complex tasks. However, the significant structure reasoning skill has been rarely studied, which involves modeling implicit structure information within the text and performing explicit logical reasoning over them to deduce the conclusion. This paper proposes a unified learning framework that combines explicit structure reasoning and language pre-training to endow PLMs with the structure reasoning skill. It first identifies several elementary structures within contexts to construct structured queries and performs step-by-step reasoning along the queries to identify the answer entity. The fusion of textual semantics and structure reasoning is achieved by using contextual representations learned by PLMs to initialize the representation space of structures, and performing stepwise reasoning on this semantic representation space. Experimental results on four datasets demonstrate that the proposed model achieves significant improvements in complex reasoning tasks involving diverse structures, and shows transferability to downstream tasks with limited training data and effectiveness for complex reasoning of KGs modality. © 2014 IEEE.
引用
收藏
页码:1586 / 1595
相关论文
共 50 条
  • [1] Unifying Structure Reasoning and Language Pre-Training for Complex Reasoning Tasks
    Wang, Siyuan
    Wei, Zhongyu
    Xu, Jiarong
    Li, Taishan
    Fan, Zhihao
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 1586 - 1595
  • [2] LogiGAN: Learning Logical Reasoning via Adversarial Pre-training
    Pi, Xinyu
    Zhong, Wanjun
    Gao, Yan
    Duan, Nan
    Lou, Jian-Guang
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [3] To Boost Zero-Shot Generalization for Embodied Reasoning with Vision-Language Pre-Training
    Su, Ke
    Zhang, Xingxing
    Zhang, Siyang
    Zhu, Jun
    Zhang, Bo
    [J]. IEEE Transactions on Image Processing, 2024, 33 : 5370 - 5381
  • [4] Pre-training Is (Almost) All You Need: An Application to Commonsense Reasoning
    Tamborrino, Alexandre
    Pellicano, Nicola
    Pannier, Baptiste
    Voitot, Pascal
    Naudin, Louise
    [J]. 58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 3878 - 3887
  • [5] Cross-modality interaction reasoning for enhancing vision-language pre-training in image-text retrieval
    Yao, Tao
    Peng, Shouyong
    Wang, Lili
    Li, Ying
    Sun, Yujuan
    [J]. APPLIED INTELLIGENCE, 2024, 54 (23) : 12230 - 12245
  • [6] Subgoal Search For Complex Reasoning Tasks
    Czechowski, Konrad
    Odrzygozdz, Tomasz
    Zbysinski, Marek
    Zawalski, Michal
    Olejnik, Krzysztof
    Wu, Yuhuai
    Kucinski, Lukasz
    Milos, Piotr
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [7] Towards Unifying Medical Vision-and-Language Pre-training via Soft Prompts
    Chen, Zhihong
    Diao, Shizhe
    Wang, Benyou
    Li, Guanbin
    Wan, Xiang
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 23346 - 23356
  • [8] Improving the Identification of Abusive Language Through Careful Design of Pre-training Tasks
    Jarquin-Vasquez, Horacio
    Jair Escalante, Hugo
    Montes-y-Gomez, Manuel
    [J]. PATTERN RECOGNITION, MCPR 2023, 2023, 13902 : 283 - 292
  • [9] ALERT: Adapting Language Models to Reasoning Tasks
    Yu, Ping
    Wang, Tianlu
    Golovneva, Olga
    AlKhamissi, Badr
    Verma, Siddharth
    Jin, Zhijing
    Ghosh, Gargi
    Diab, Mona
    Celikyilmaz, Asli
    [J]. PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 1055 - 1081
  • [10] A Graph Reasoning Network for Multi-turn Response Selection via Customized Pre-training
    Liu, Yongkang
    Feng, Shi
    Wang, Daling
    Song, Kaisong
    Ren, Feiliang
    Zhang, Yifei
    [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 13433 - 13442