Wasserstein discriminant analysis

Research output: Contribution to journalArticlepeer-review

Abstract

Wasserstein discriminant analysis (WDA) is a new supervised linear dimensionality reduction algorithm. Following the blueprint of classical Fisher Discriminant Analysis, WDA selects the projection matrix that maximizes the ratio of the dispersion of projected points pertaining to different classes and the dispersion of projected points belonging to a same class. To quantify dispersion, WDA uses regularized Wasserstein distances. Thanks to the underlying principles of optimal transport, WDA is able to capture both global (at distribution scale) and local (at samples’ scale) interactions between classes. In addition, we show that WDA leverages a mechanism that induces neighborhood preservation. Regularized Wasserstein distances can be computed using the Sinkhorn matrix scaling algorithm; the optimization problem of WDA can be tackled using automatic differentiation of Sinkhorn’s fixed-point iterations. Numerical experiments show promising results both in terms of prediction and visualization on toy examples and real datasets such as MNIST and on deep features obtained from a subset of the Caltech dataset.

Original languageEnglish
Pages (from-to)1923-1945
Number of pages23
JournalMachine Learning
Volume107
Issue number12
DOIs
Publication statusPublished - 1 Dec 2018
Externally publishedYes

Keywords

  • Linear discriminant analysis
  • Optimal transport
  • Wasserstein distance

Fingerprint

Dive into the research topics of 'Wasserstein discriminant analysis'. Together they form a unique fingerprint.

Cite this