Chinese Semantic Role Labeling Based on Dynamic Syntax Pruning

被引:0
|
作者
Fei H. [1 ]
Ji D.-H. [1 ]
Ren Y.-F. [2 ]
机构
[1] School of Cyber Science and Engineering, Wuhan University, Wuhan
[2] School of Interpreting and Translation Studies, Guangdong University of Foreign Studies, Guangzhou
来源
基金
中国国家自然科学基金;
关键词
Deep learning; Natural language processing; Neural network; Semantic role labeling; Syntax pruning;
D O I
10.11897/SP.J.1016.2022.01746
中图分类号
学科分类号
摘要
Semantic role labeling(SRL), as the shallow semantic parsing task, which has received extensive research attention in recent years and plays a core role in the natural language processing(NLP) community. The SRL task aims to identify the corresponding argument roles for the predicates of a given sentence, which can facilitate the downstream NLP tasks, such as information extraction, question answer system and reading comprehension, etc. A great number of methods have been proposed for the task, and the existing studies can be divided into two main categories: machine learning based methods with hand-crafted discrete features and deep learning methods with automatic distributed features. The early studies largely separate SRL into two individual subtasks, i.e., predicate disambiguation and argument role labeling. More recently, great efforts have been paid for constructing various end-to-end SRL architectures, solving two pipeline steps in one shot via one unified model. Recent studies also show that integrating external syntactic features, such as syntactic dependency trees, are important for the SRL task highly. So designing a novel neural model, which can capture syntactic features effectively, has become a heated research topic. Recently, He et al.(2018) find that only a part of syntactic structure information can offer valuable information for the SRL task, which calls for pruning the syntactic structure features. However, the existing work adopts the offline syntactic pruning strategy, which can inevitably lead to either the loss of key syntactic information or the weakening of pruning effectiveness. Extracting syntactic features, as an important step of the SRL task, will largely affect the final performance of the task. However, the existing neural network methods fail to effectively model syntactic features. For example, the existing studies adopt the offline syntactic pruning strategy with fixed human labor, which inevitably leads to the loss of key syntactic information or the weakening of pruning effectiveness. To address the above issues, we propose an end-to-end neural network model for the Chinese SRL task based on dynamic syntactic pruning mechanism. Specifically, we propose two novel methods: recursive neural network model with dynamic syntactic pruning(Recur-DSP) and syntax-label graph convolutional network with dynamic syntactic pruning(SGCN-DSP). Recur-DSP uses a recursive neural network model to encode and fuse syntactic structure knowledge, and applies the Gumbel-Softmax function to realize dynamic syntactic pruning. SGCN-DSP exploits a graph convolutional neural network model that can simultaneously encode syntactic arcs and labels, based on which we introduce the corresponding dynamic syntactic pruning strategy. Experimental results on multiple benchmark datasets show the effectiveness of the proposed methods. Our proposed methods outperform the current best method by a large margin, giving the state-of-the-art performances for the Chinese SRL task. Specifically, our proposed model SGCN-DSP achieves 86.9% F1 score in argument role labeling and 89.1% F1 score in predicate identification based on the CoNLL09 dataset. By integrating the current pre-trained language model BERT(Bidirectional Encoder Representation from Transformers, BERT), the task performance can be further improved. The proposed SGCN-DSP gives 90.4% F1 score in argument role labeling, and 90.8% F1 scores in predicate identification, respectively. © 2022, Science Press. All right reserved.
引用
收藏
页码:1746 / 1764
页数:18
相关论文
共 57 条
  • [41] Marcus M P, Santorini B, Marcinkiewicz M A., Building a large annotated corpus of English: The Penn Treebank, Computational Linguistics, 19, 2, pp. 313-330, (1993)
  • [42] Kipper K, Dang H T, Palmer M S., Class-based construction of a verb lexicon, Proceedings of the Seventeenth National Conference on Artificial Intelligence and Twelfth Conference on Innovative Applications of Artificial Intelligence, pp. 691-696, (2000)
  • [43] Fei H, Ren Y, Ji D., High-order refining for end-to-end Chinese semantic role labeling, Proceedings of the 1st Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 10th International Joint Conference on Natural Language Processing, pp. 100-105, (2020)
  • [44] Sun W, Sui Z, Wang M, Et al., Chinese semantic role labeling with shallow parsing, Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing, pp. 1475-1483, (2009)
  • [45] Wang Z, Jiang T, Chang B, Et al., Chinese semantic role labeling with bidirectional recurrent neural networks, Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp. 1626-1631, (2015)
  • [46] Xia Q, Sha L, Chang B, Et al., A progressive learning approach to Chinese SRL using heterogeneous data, Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, pp. 2069-2077, (2017)
  • [47] Dozat T, Manning C D., Deep biaffine attention for neural dependency parsing, Proceedings of the International Conference on Learning Representations, pp. 1-10, (2017)
  • [48] Tai K S, Socher R, Manning C D., Improved semantic representations from tree-structured long short-term memory networks, Proceedings of the Annual Meeting of the Association for Computational Linguistics, pp. 1556-1566, (2015)
  • [49] Gumbel E J., Statistical theory of extreme values and some practical applications: A series of lectures: Volume 33, (1954)
  • [50] Zhao H, Chen W, Kit C., Semantic dependency parsing of NomBank and PropBank: An efficient integrated approach via a large-scale feature selection, Proceedings of the Conference on Empirical Methods in Natural Language Processing, pp. 30-39, (2009)