One Wave To Explain Them All: A Unifying Perspective On Feature Attribution

Research output: Contribution to journalConference articlepeer-review

Abstract

Feature attribution methods aim to improve the transparency of deep neural networks by identifying the input features that influence a model’s decision. Pixel-based heatmaps have become the standard for attributing features to high-dimensional inputs, such as images, audio representations, and volumes. While intuitive and convenient, these pixel-based attributions fail to capture the underlying structure of the data. Moreover, the choice of domain for computing attributions has often been overlooked. This work demonstrates that the wavelet domain allows for informative and meaningful attributions. It handles any input dimension and offers a unified approach to feature attribution. Our method, the Wavelet Attribution Method (WAM), leverages the spatial and scale-localized properties of wavelet coefficients to provide explanations that capture both the where and what of a model’s decision-making process. We show that WAM quantitatively matches or outperforms existing gradient-based methods across multiple modalities, including audio, images, and volumes. Additionally, we discuss how WAM bridges attribution with broader aspects of model robustness and transparency. Project page: https: //gabrielkasmi.github.io/wam/.

Original languageEnglish
Pages (from-to)29265-29293
Number of pages29
JournalProceedings of Machine Learning Research
Volume267
Publication statusPublished - 1 Jan 2025
Externally publishedYes
Event42nd International Conference on Machine Learning, ICML 2025 - Vancouver, Canada
Duration: 13 Jul 202519 Jul 2025

Fingerprint

Dive into the research topics of 'One Wave To Explain Them All: A Unifying Perspective On Feature Attribution'. Together they form a unique fingerprint.

Cite this