TY - GEN
T1 - A Study on Hierarchical Text Classification as a Seq2seq Task
AU - Torba, Fatos
AU - Gravier, Christophe
AU - Laclau, Charlotte
AU - Kammoun, Abderrhammen
AU - Subercaze, Julien
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024.
PY - 2024/1/1
Y1 - 2024/1/1
N2 - With the progress of generative neural models, Hierarchical Text Classification (HTC) can be cast as a generative task. In this case, given an input text, the model generates the sequence of predicted class labels taken from a label tree of arbitrary width and depth. Treating HTC as a generative task introduces multiple modeling choices. These choices vary from choosing the order for visiting the class tree and therefore defining the order of generating tokens, choosing either to constrain the decoding to labels that respect the previous level predictions, up to choosing the pre-trained Language Model itself. Each HTC model therefore differs from the others from an architectural standpoint, but also from the modeling choices that were made. Prior contributions lack transparent modeling choices and open implementations, hindering the assessment of whether model performance stems from architectural or modeling decisions. For these reasons, we propose with this paper an analysis of the impact of different modeling choices along with common model errors and successes for this task. This analysis is based on an open framework coming along this paper that can facilitate the development of future contributions in the field by providing datasets, metrics, error analysis toolkit and the capability to readily test various modeling choices for one given model.
AB - With the progress of generative neural models, Hierarchical Text Classification (HTC) can be cast as a generative task. In this case, given an input text, the model generates the sequence of predicted class labels taken from a label tree of arbitrary width and depth. Treating HTC as a generative task introduces multiple modeling choices. These choices vary from choosing the order for visiting the class tree and therefore defining the order of generating tokens, choosing either to constrain the decoding to labels that respect the previous level predictions, up to choosing the pre-trained Language Model itself. Each HTC model therefore differs from the others from an architectural standpoint, but also from the modeling choices that were made. Prior contributions lack transparent modeling choices and open implementations, hindering the assessment of whether model performance stems from architectural or modeling decisions. For these reasons, we propose with this paper an analysis of the impact of different modeling choices along with common model errors and successes for this task. This analysis is based on an open framework coming along this paper that can facilitate the development of future contributions in the field by providing datasets, metrics, error analysis toolkit and the capability to readily test various modeling choices for one given model.
KW - Hierarchical text classification
KW - generative model
KW - reproducibility
U2 - 10.1007/978-3-031-56063-7_20
DO - 10.1007/978-3-031-56063-7_20
M3 - Conference contribution
AN - SCOPUS:85189373735
SN - 9783031560620
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 287
EP - 296
BT - Advances in Information Retrieval - 46th European Conference on Information Retrieval, ECIR 2024, Proceedings
A2 - Goharian, Nazli
A2 - Tonellotto, Nicola
A2 - He, Yulan
A2 - Lipani, Aldo
A2 - McDonald, Graham
A2 - Macdonald, Craig
A2 - Ounis, Iadh
PB - Springer Science and Business Media Deutschland GmbH
T2 - 46th European Conference on Information Retrieval, ECIR 2024
Y2 - 24 March 2024 through 28 March 2024
ER -