A Study on Hierarchical Text Classification as a Seq2seq Task

被引:0
|
作者
Torba, Fatos [1 ,2 ]
Gravier, Christophe [2 ]
Laclau, Charlotte [3 ]
Kammoun, Abderrhammen [1 ]
Subercaze, Julien [1 ]
机构
[1] AItenders, St Etienne, France
[2] CNRS, Lab Hubert Curien, UMR 5516, St Etienne, France
[3] Inst Polytech Paris, Telecom Paris, Paris, France
关键词
Hierarchical text classification; generative model; reproducibility;
D O I
10.1007/978-3-031-56063-7_20
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the progress of generative neural models, Hierarchical Text Classification (HTC) can be cast as a generative task. In this case, given an input text, the model generates the sequence of predicted class labels taken from a label tree of arbitrary width and depth. Treating HTC as a generative task introduces multiple modeling choices. These choices vary from choosing the order for visiting the class tree and therefore defining the order of generating tokens, choosing either to constrain the decoding to labels that respect the previous level predictions, up to choosing the pre-trained Language Model itself. Each HTC model therefore differs from the others from an architectural standpoint, but also from the modeling choices that were made. Prior contributions lack transparent modeling choices and open implementations, hindering the assessment of whether model performance stems from architectural or modeling decisions. For these reasons, we propose with this paper an analysis of the impact of different modeling choices along with common model errors and successes for this task. This analysis is based on an open framework coming along this paper that can facilitate the development of future contributions in the field by providing datasets, metrics, error analysis toolkit and the capability to readily test various modeling choices for one given model.
引用
收藏
页码:287 / 296
页数:10
相关论文
共 50 条
  • [31] Tool Wear Monitoring System Using Seq2Seq
    Jeon, Wang-Su
    Rhee, Sang-Yong
    MACHINES, 2024, 12 (03)
  • [32] Keyphrase Generation Based on Deep Seq2seq Model
    Zhang, Yong
    Xiao, Weidong
    IEEE ACCESS, 2018, 6 : 46047 - 46057
  • [33] Guesswork for Inference in Machine Translation with Seq2seq Model
    Liu, Lilian
    Malak, Derya
    Medard, Muriel
    2019 IEEE INFORMATION THEORY WORKSHOP (ITW), 2019, : 60 - 64
  • [34] Adversarial Oracular Seq2seq Learning for Sequential Recommendation
    Zhao, Pengyu
    Shui, Tianxiao
    Zhang, Yuanxing
    Xiao, Kecheng
    Bian, Kaigui
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 1905 - 1911
  • [35] A Method for Estimating Process Maliciousness with Seq2Seq Model
    Tobiyama, Shun
    Yamaguchi, Yukiko
    Hasegawa, Hirokazu
    Shimada, Hajime
    Akiyama, Mitsuaki
    Yagi, Takeshi
    2018 32ND INTERNATIONAL CONFERENCE ON INFORMATION NETWORKING (ICOIN), 2018, : 255 - 260
  • [36] Learning Transductions and Alignments with RNN Seq2seq Models
    Wang, Zhengxiang
    INTERNATIONAL CONFERENCE ON GRAMMATICAL INFERENCE, VOL 217, 2023, 217 : 223 - 249
  • [37] Seq2Seq Deep Learning Models for Microtext Normalization
    Satapathy, Ranjan
    Li, Yang
    Cavallari, Sandro
    Cambria, Erik
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [38] A Study on the Application of Seq2seq in English Flipped Classroom Teaching in Colleges and Universities
    Liu Q.
    Applied Mathematics and Nonlinear Sciences, 2024, 9 (01)
  • [39] SEQ2SEQ++: A Multitasking-Based Seq2seq Model to Generate Meaningful and Relevant Answers
    Palasundram, Kulothunkan
    Sharef, Nurfadhlina Mohd
    Kasmiran, Khairul Azhar
    Azman, Azreen
    IEEE ACCESS, 2021, 9 (09): : 164949 - 164975
  • [40] How Do Seq2Seq Models Perform on End-to-End Data-to-Text Generation?
    Yin, Xunjian
    Wan, Xiaojun
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 7701 - 7710