Abstract
We consider the binary supervised classification problem with the Gaussian functional model introduced in (Math. Methods Statist. 22 (2013) 213-225). Taking advantage of the Gaussian structure, we design a natural plug-in classifier and derive a family of upper bounds on its worst-case excess risk over Sobolev spaces. These bounds are parametrized by a separation distance quantifying the difficulty of the problem, and are proved to be optimal (up to logarithmic factors) through matching minimax lower bounds. Using the recent works of (In Advances in Neural Information Processing Systems (2014) 3437-3445 Curran Associates) and (Ann. Statist. 44 (2016) 982-1009), we also derive a logarithmic lower bound showing that the popular k-nearest neighbors classifier is far from optimality in this specific functional setting.
| Original language | English |
|---|---|
| Pages (from-to) | 1797-1831 |
| Number of pages | 35 |
| Journal | Bernoulli |
| Volume | 26 |
| Issue number | 3 |
| DOIs | |
| Publication status | Published - 1 Aug 2020 |
| Externally published | Yes |
Keywords
- Functional data
- Supervised classification
Fingerprint
Dive into the research topics of 'Optimal functional supervised classification with separation condition'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver