Unifying Structure Reasoning and Language Pre-Training for Complex Reasoning Tasks

被引:2
|
作者
Wang, Siyuan [1 ]
Wei, Zhongyu [1 ,2 ]
Xu, Jiarong [3 ]
Li, Taishan [4 ]
Fan, Zhihao [1 ]
机构
[1] Fudan Univ, Sch Data Sci, Shanghai 200433, Peoples R China
[2] Fudan Univ, Res Inst Intelligent & Complex Syst, Shanghai 200433, Peoples R China
[3] Fudan Univ, Sch Management, Shanghai 200433, Peoples R China
[4] ShanghaiTech Univ, Sch Informat Sci & Technol, Shanghai 201210, Peoples R China
基金
中国国家自然科学基金;
关键词
Cognition; Task analysis; Semantics; Films; Speech processing; Context modeling; Data models; Structure reasoning skill; language model pre-training; complex reasoning;
D O I
10.1109/TASLP.2023.3325973
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Recent pre-trained language models (PLMs) equipped with foundation reasoning skills have shown remarkable performance on downstream complex tasks. However, the significant structure reasoning skill has been rarely studied, which involves modeling implicit structure information within the text and performing explicit logical reasoning over them to deduce the conclusion. This paper proposes a unified learning framework that combines explicit structure reasoning and language pre-training to endow PLMs with the structure reasoning skill. It first identifies several elementary structures within contexts to construct structured queries and performs step-by-step reasoning along the queries to identify the answer entity. The fusion of textual semantics and structure reasoning is achieved by using contextual representations learned by PLMs to initialize the representation space of structures, and performing stepwise reasoning on this semantic representation space. Experimental results on four datasets demonstrate that the proposed model achieves significant improvements in complex reasoning tasks involving diverse structures, and shows transferability to downstream tasks with limited training data and effectiveness for complex reasoning of KGs modality.
引用
收藏
页码:1586 / 1595
页数:10
相关论文
共 50 条
  • [41] eMLM: A New Pre-training Objective for Emotion Related Tasks
    Sosea, Tiberiu
    Caragea, Cornelia
    ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 2, 2021, : 286 - 293
  • [42] Enhancing abusive language detection: A domain-adapted approach leveraging BERT pre-training tasks
    Jarquin-Vasquez, Horacio
    Escalante, Hugo Jair
    Montes-y-Gomez, Manuel
    PATTERN RECOGNITION LETTERS, 2024, 186 : 361 - 368
  • [43] Training Complements for Belief Reasoning in Developmental Language Disorder
    Durrleman, Stephanie
    Delage, Helene
    JOURNAL OF SPEECH LANGUAGE AND HEARING RESEARCH, 2020, 63 (06): : 1861 - 1877
  • [44] Training analogical reasoning skills in children with language disorders
    Masterson, JJ
    Perrey, CD
    AMERICAN JOURNAL OF SPEECH-LANGUAGE PATHOLOGY, 1999, 8 (01) : 53 - 61
  • [45] Targeted training for numerical reasoning with large language models
    Li, Xiao
    Liu, Sichen
    Zhu, Yin
    Cheng, Gong
    KNOWLEDGE AND INFORMATION SYSTEMS, 2025, 67 (01) : 197 - 221
  • [46] Contrastive Language-knowledge Graph Pre-training
    Yuan, Xiaowei
    Liu, Kang
    Wang, Yequan
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2024, 23 (04)
  • [47] VLP: A Survey on Vision-language Pre-training
    Chen, Fei-Long
    Zhang, Du-Zhen
    Han, Ming-Lun
    Chen, Xiu-Yi
    Shi, Jing
    Xu, Shuang
    Xu, Bo
    MACHINE INTELLIGENCE RESEARCH, 2023, 20 (01) : 38 - 56
  • [48] Improving the Sample Efficiency of Pre-training Language Models
    Berend, Gabor
    ERCIM NEWS, 2024, (136): : 38 - 40
  • [49] Continual pre-training mitigates forgetting in language and vision
    Cossu, Andrea
    Carta, Antonio
    Passaro, Lucia
    Lomonaco, Vincenzo
    Tuytelaars, Tinne
    Bacciu, Davide
    NEURAL NETWORKS, 2024, 179
  • [50] Visual Alignment Pre-training for Sign Language Translation
    Jiao, Peiqi
    Min, Yuecong
    Chen, Xilin
    COMPUTER VISION - ECCV 2024, PT XLII, 2025, 15100 : 349 - 367