TY - GEN
T1 - Confidence-based weighted loss for multi-label classification with missing labels
AU - Ibrahim, Karim M.
AU - Epure, Elena V.
AU - Peeters, Geoffroy
AU - Richard, Gaël
N1 - Publisher Copyright:
© 2020 ACM.
PY - 2020/6/11
Y1 - 2020/6/11
N2 - The problem of multi-label classification with missing labels (MLML) is a common challenge that is prevalent in several domains, e.g. image annotation and auto-tagging. In multi-label classification, each instance may belong to multiple class labels simultaneously. Due to the nature of the dataset collection and labelling procedure, it is common to have incomplete annotations in the dataset, i.e. not all samples are labelled with all the corresponding labels. However, the incomplete data labelling hinders the training of classification models. MLML has received much attention from the research community. However, in cases where a pre-trained model is fine-tuned on an MLML dataset, there has been no straightforward approach to tackle the missing labels, specifically when there is no information about which are the missing ones. In this paper, we propose a weighted loss function to account for the confidence in each label/sample pair that can easily be incorporated to fine-tune a pre-trained model on an incomplete dataset. Our experiment results show that using the proposed loss function improves the performance of the model as the ratio of missing labels increases.
AB - The problem of multi-label classification with missing labels (MLML) is a common challenge that is prevalent in several domains, e.g. image annotation and auto-tagging. In multi-label classification, each instance may belong to multiple class labels simultaneously. Due to the nature of the dataset collection and labelling procedure, it is common to have incomplete annotations in the dataset, i.e. not all samples are labelled with all the corresponding labels. However, the incomplete data labelling hinders the training of classification models. MLML has received much attention from the research community. However, in cases where a pre-trained model is fine-tuned on an MLML dataset, there has been no straightforward approach to tackle the missing labels, specifically when there is no information about which are the missing ones. In this paper, we propose a weighted loss function to account for the confidence in each label/sample pair that can easily be incorporated to fine-tune a pre-trained model on an incomplete dataset. Our experiment results show that using the proposed loss function improves the performance of the model as the ratio of missing labels increases.
KW - Missing labels
KW - Multi-label classification
KW - Neural networks
U2 - 10.1145/3372278.3390728
DO - 10.1145/3372278.3390728
M3 - Conference contribution
AN - SCOPUS:85086898910
T3 - ICMR 2020 - Proceedings of the 2020 International Conference on Multimedia Retrieval
SP - 291
EP - 295
BT - ICMR 2020 - Proceedings of the 2020 International Conference on Multimedia Retrieval
PB - Association for Computing Machinery
T2 - 10th ACM International Conference on Multimedia Retrieval, ICMR 2020
Y2 - 8 June 2020 through 11 June 2020
ER -