Multi-task learning with contextual hierarchical attention for Korean coreference resolution

被引:1
|
作者
Park, Cheoneum [1 ]
机构
[1] AIRS Co, Hyundai Motor Grp, Seoul, South Korea
关键词
coreference resolution; hierarchical model; head-final language; multi-task learning; pointer network;
D O I
10.4218/etrij.2021-0293
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Coreference resolution is a task in discourse analysis that links several headwords used in any document object. We suggest pointer networks-based coreference resolution for Korean using multi-task learning (MTL) with an attention mechanism for a hierarchical structure. As Korean is a head-final language, the head can easily be found. Our model learns the distribution by referring to the same entity position and utilizes a pointer network to conduct coreference resolution depending on the input headword. As the input is a document, the input sequence is very long. Thus, the core idea is to learn the word- and sentence-level distributions in parallel with MTL, while using a shared representation to address the long sequence problem. The suggested technique is used to generate word representations for Korean based on contextual information using pre-trained language models for Korean. In the same experimental conditions, our model performed roughly 1.8% better on CoNLL F1 than previous research without hierarchical structure.
引用
收藏
页码:93 / 104
页数:12
相关论文
共 50 条
  • [31] Enhanced task attention with adversarial learning for dynamic multi-task CNN
    Fang, Yuchun
    Xiao, Shiwei
    Zhou, Menglu
    Cai, Sirui
    Zhang, Zhaoxiang
    [J]. PATTERN RECOGNITION, 2022, 128
  • [32] Multi-Faceted Hierarchical Multi-Task Learning for Recommender Systems
    Liu, Junning
    Li, Xinjian
    An, Bo
    Xia, Zijie
    Wang, Xu
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 3332 - 3341
  • [33] Unsupervised learning of contextual role knowledge for coreference resolution
    Bean, D
    Riloff, E
    [J]. HLT-NAACL 2004: HUMAN LANGUAGE TECHNOLOGY CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, PROCEEDINGS OF THE MAIN CONFERENCE, 2004, : 297 - 304
  • [34] Multi-task gradient descent for multi-task learning
    Bai, Lu
    Ong, Yew-Soon
    He, Tiantian
    Gupta, Abhishek
    [J]. MEMETIC COMPUTING, 2020, 12 (04) : 355 - 369
  • [35] Multi-task gradient descent for multi-task learning
    Lu Bai
    Yew-Soon Ong
    Tiantian He
    Abhishek Gupta
    [J]. Memetic Computing, 2020, 12 : 355 - 369
  • [36] Multi-task learning with Attention : Constructing auxiliary tasks for learning to learn
    Li, Benying
    Dong, Aimei
    [J]. 2021 IEEE 33RD INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2021), 2021, : 145 - 152
  • [37] Multi-task Learning Method for Hierarchical Time Series Forecasting
    Yang, Maoxin
    Hu, Qinghua
    Wang, Yun
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2019: TEXT AND TIME SERIES, PT IV, 2019, 11730 : 474 - 485
  • [38] Multi-Task Decouple Learning With Hierarchical Attentive Point Process
    Wu, Weichang
    Zhang, Xiaolu
    Zhao, Shiwan
    Fu, Chilin
    Zhou, Jun
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (04) : 1741 - 1757
  • [39] Contextual Anomaly Detection in Solder Paste Inspection with Multi-Task Learning
    Zheng, Zimu
    Pu, Jie
    Liu, Linghui
    Wang, Dan
    Mei, Xiangming
    Zhang, Sen
    Dai, Quanyu
    [J]. ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2020, 11 (06)
  • [40] A Novel Orthogonality Loss for Deep Hierarchical Multi-Task Learning
    He, Guiqing
    Huo, Yincheng
    He, Mingyao
    Zhang, Haixi
    Fan, Jianping
    [J]. IEEE ACCESS, 2020, 8 : 67735 - 67744