CLIP-QDA: An Explainable Concept Bottleneck Model

Research output: Contribution to journalArticlepeer-review

Abstract

In this paper, we introduce an explainable algorithm designed from a multi-modal foundation model, that performs fast and explainable image classification. Drawing inspiration from CLIP-based Concept Bottleneck Models (CBMs), our method creates a latent space where each neuron is linked to a specific word. Observing that this latent space can be modeled with simple distributions, we use a Mixture of Gaussians (MoG) formalism to enhance the interpretability of this latent space. Then, we introduce CLIP-QDA, a classifier that only uses statistical values to infer labels from the concepts. In addition, this formalism allows for both sample-wise and dataset-wise explanations. These explanations come from the inner design of our architecture, our work is part of a new family of greybox models, combining performances of opaque foundation models and the interpretability of transparent models. Our empirical findings show that in instances where the MoG assumption holds, CLIP-QDA achieves similar accuracy with state-of-the-art CBMs. Our explanations compete with existing XAI methods while being faster to compute.

Original languageEnglish
JournalTransactions on Machine Learning Research
Volume2024
Publication statusPublished - 1 Jan 2024

Fingerprint

Dive into the research topics of 'CLIP-QDA: An Explainable Concept Bottleneck Model'. Together they form a unique fingerprint.

Cite this