A Study on Hierarchical Text Classification as a Seq2seq Task

被引:0
|
作者
Torba, Fatos [1 ,2 ]
Gravier, Christophe [2 ]
Laclau, Charlotte [3 ]
Kammoun, Abderrhammen [1 ]
Subercaze, Julien [1 ]
机构
[1] AItenders, St Etienne, France
[2] CNRS, Lab Hubert Curien, UMR 5516, St Etienne, France
[3] Inst Polytech Paris, Telecom Paris, Paris, France
关键词
Hierarchical text classification; generative model; reproducibility;
D O I
10.1007/978-3-031-56063-7_20
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the progress of generative neural models, Hierarchical Text Classification (HTC) can be cast as a generative task. In this case, given an input text, the model generates the sequence of predicted class labels taken from a label tree of arbitrary width and depth. Treating HTC as a generative task introduces multiple modeling choices. These choices vary from choosing the order for visiting the class tree and therefore defining the order of generating tokens, choosing either to constrain the decoding to labels that respect the previous level predictions, up to choosing the pre-trained Language Model itself. Each HTC model therefore differs from the others from an architectural standpoint, but also from the modeling choices that were made. Prior contributions lack transparent modeling choices and open implementations, hindering the assessment of whether model performance stems from architectural or modeling decisions. For these reasons, we propose with this paper an analysis of the impact of different modeling choices along with common model errors and successes for this task. This analysis is based on an open framework coming along this paper that can facilitate the development of future contributions in the field by providing datasets, metrics, error analysis toolkit and the capability to readily test various modeling choices for one given model.
引用
收藏
页码:287 / 296
页数:10
相关论文
共 50 条
  • [21] Abstractive social media text summarization using selective reinforced Seq2Seq attention model
    Liang, Zeyu
    Du, Junping
    Li, Chaoyang
    NEUROCOMPUTING, 2020, 410 : 432 - 440
  • [22] Work modes recognition and boundary identification of MFR pulse sequences with a hierarchical seq2seq LSTM
    Li, Yunjie
    Zhu, Mengtao
    Ma, Yihao
    Yang, Jian
    IET RADAR SONAR AND NAVIGATION, 2020, 14 (09): : 1343 - 1353
  • [23] Residual Seq2Seq model for Building energy management
    Kim, Marie
    Kim, Nae-soo
    Song, YuJin
    Pyo, Cheol Sig
    2019 10TH INTERNATIONAL CONFERENCE ON INFORMATION AND COMMUNICATION TECHNOLOGY CONVERGENCE (ICTC): ICT CONVERGENCE LEADING THE AUTONOMOUS FUTURE, 2019, : 1126 - 1128
  • [24] Automatic Generation of Pseudocode with Attention Seq2seq Model
    Xu, Shaofeng
    Xiong, Yun
    2018 25TH ASIA-PACIFIC SOFTWARE ENGINEERING CONFERENCE (APSEC 2018), 2018, : 711 - 712
  • [25] Map Matching Based on Seq2Seq with Topology Information
    Bai, Yulong
    Li, Guolian
    Lu, Tianxiu
    Wu, Yadong
    Zhang, Weihan
    Feng, Yidan
    APPLIED SCIENCES-BASEL, 2023, 13 (23):
  • [26] Seq2Seq模型的短期水位预测
    刘艳
    张婷
    康爱卿
    李建柱
    雷晓辉
    水利水电科技进展, 2022, 42 (03) : 57 - 63
  • [27] Smoothing and Shrinking the Sparse Seq2Seq Search Space
    Peters, Ben
    Martins, Andre F. T.
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 2642 - 2654
  • [28] Sliding Window Seq2seq Modeling for Engagement Estimation
    Yu, Jun
    Lu, Keda
    Jing, Mohan
    Liang, Ziqi
    Zhang, Bingyuan
    Sun, Jianqing
    Liang, Jiaen
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 9496 - 9500
  • [29] SparQL Query Prediction Based on Seq2Seq Model
    Yang D.-H.
    Zou K.-F.
    Wang H.-Z.
    Wang J.-B.
    Ruan Jian Xue Bao/Journal of Software, 2021, 32 (03): : 805 - 817
  • [30] Untargeted Code Authorship Evasion with Seq2Seq Transformation
    Choi, Soohyeon
    Jang, Rhongho
    Nyang, DaeHun
    Mohaisen, David
    arXiv, 2023,