A Study on Hierarchical Text Classification as a Seq2seq Task

被引:0
|
作者
Torba, Fatos [1 ,2 ]
Gravier, Christophe [2 ]
Laclau, Charlotte [3 ]
Kammoun, Abderrhammen [1 ]
Subercaze, Julien [1 ]
机构
[1] AItenders, St Etienne, France
[2] CNRS, Lab Hubert Curien, UMR 5516, St Etienne, France
[3] Inst Polytech Paris, Telecom Paris, Paris, France
关键词
Hierarchical text classification; generative model; reproducibility;
D O I
10.1007/978-3-031-56063-7_20
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the progress of generative neural models, Hierarchical Text Classification (HTC) can be cast as a generative task. In this case, given an input text, the model generates the sequence of predicted class labels taken from a label tree of arbitrary width and depth. Treating HTC as a generative task introduces multiple modeling choices. These choices vary from choosing the order for visiting the class tree and therefore defining the order of generating tokens, choosing either to constrain the decoding to labels that respect the previous level predictions, up to choosing the pre-trained Language Model itself. Each HTC model therefore differs from the others from an architectural standpoint, but also from the modeling choices that were made. Prior contributions lack transparent modeling choices and open implementations, hindering the assessment of whether model performance stems from architectural or modeling decisions. For these reasons, we propose with this paper an analysis of the impact of different modeling choices along with common model errors and successes for this task. This analysis is based on an open framework coming along this paper that can facilitate the development of future contributions in the field by providing datasets, metrics, error analysis toolkit and the capability to readily test various modeling choices for one given model.
引用
收藏
页码:287 / 296
页数:10
相关论文
共 50 条
  • [1] A Hierarchical Attention Seq2seq Model with CopyNet for Text Summarization
    Zhang, Yong
    Wang, Yuheng
    Liao, Jinzhi
    Xiao, Weidong
    2018 INTERNATIONAL CONFERENCE ON ROBOTS & INTELLIGENT SYSTEM (ICRIS 2018), 2018, : 316 - 320
  • [2] Multi-label Text Classification Based on Improved Seq2Seq
    Chen, Xiaolong
    Cheng, Jieren
    Rong, Zhixin
    Xu, Wenghang
    Hua, Shuai
    Tang, Zhu
    PROCEEDINGS OF THE 13TH INTERNATIONAL CONFERENCE ON COMPUTER ENGINEERING AND NETWORKS, VOL II, CENET 2023, 2024, 1126 : 439 - 446
  • [3] A Chinese text corrector based on seq2seq model
    Gu, Sunyan
    Lang, Fei
    2017 INTERNATIONAL CONFERENCE ON CYBER-ENABLED DISTRIBUTED COMPUTING AND KNOWLEDGE DISCOVERY (CYBERC), 2017, : 322 - 325
  • [4] Seq2Seq models for recommending short text conversations
    Torres, Johnny
    Vaca, Carmen
    Teran, Luis
    Abad, Cristina L.
    EXPERT SYSTEMS WITH APPLICATIONS, 2020, 150
  • [5] Abstract Text Summarization with a Convolutional Seq2seq Model
    Zhang, Yong
    Li, Dan
    Wang, Yuheng
    Fang, Yang
    Xiao, Weidong
    APPLIED SCIENCES-BASEL, 2019, 9 (08):
  • [6] Seq2Seq dynamic planning network for progressive text generation
    Wu, Di
    Cheng, Peng
    Zheng, Yuying
    COMPUTER SPEECH AND LANGUAGE, 2025, 89
  • [7] History-based attention in Seq2Seq model for multi-label text classification
    Xiao, Yaoqiang
    Li, Yi
    Yuan, Jin
    Guo, Songrui
    Xiao, Yi
    Li, Zhiyong
    KNOWLEDGE-BASED SYSTEMS, 2021, 224
  • [8] A Hierarchical Attention Based Seq2Seq Model for Chinese Lyrics Generation
    Fan, Haoshen
    Wang, Jie
    Zhuang, Bojin
    Wang, Shaojun
    Xiao, Jing
    PRICAI 2019: TRENDS IN ARTIFICIAL INTELLIGENCE, PT III, 2019, 11672 : 279 - 288
  • [9] Open-Domain Table-to-Text Generation based on Seq2seq
    Cao, Juan
    Gong, Junpeng
    Zhang, Pengzhou
    2018 INTERNATIONAL CONFERENCE ON ALGORITHMS, COMPUTING AND ARTIFICIAL INTELLIGENCE (ACAI 2018), 2018,
  • [10] Enhanced Seq2Seq Autoencoder via Contrastive Learning for Abstractive Text Summarization
    Zheng, Chujie
    Zhang, Kunpeng
    Wang, Harry Jiannan
    Fan, Ling
    Wang, Zhe
    2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2021, : 1764 - 1771