Passer à la navigation principale Passer à la recherche Passer au contenu principal

CLIP-QDA: An Explainable Concept Bottleneck Model

Résultats de recherche: Contribution à un journalArticleRevue par des pairs

Résumé

In this paper, we introduce an explainable algorithm designed from a multi-modal foundation model, that performs fast and explainable image classification. Drawing inspiration from CLIP-based Concept Bottleneck Models (CBMs), our method creates a latent space where each neuron is linked to a specific word. Observing that this latent space can be modeled with simple distributions, we use a Mixture of Gaussians (MoG) formalism to enhance the interpretability of this latent space. Then, we introduce CLIP-QDA, a classifier that only uses statistical values to infer labels from the concepts. In addition, this formalism allows for both sample-wise and dataset-wise explanations. These explanations come from the inner design of our architecture, our work is part of a new family of greybox models, combining performances of opaque foundation models and the interpretability of transparent models. Our empirical findings show that in instances where the MoG assumption holds, CLIP-QDA achieves similar accuracy with state-of-the-art CBMs. Our explanations compete with existing XAI methods while being faster to compute.

langue originaleAnglais
journalTransactions on Machine Learning Research
Volume2024
étatPublié - 1 janv. 2024

Empreinte digitale

Examiner les sujets de recherche de « CLIP-QDA: An Explainable Concept Bottleneck Model ». Ensemble, ils forment une empreinte digitale unique.

Contient cette citation