Abstract
Assessing the probability of occurrence of extreme events is a crucial issue in various fields like finance, insurance, telecommunication or environmental sciences. In a multivariate framework, the tail dependence is characterized by the so-called stable tail dependence function (STDF). Learning this structure is the keystone of multivariate extremes. Although extensive studies have proved consistency and asymptotic normality for the empirical version of the STDF, non-asymptotic bounds are still missing. The main purpose of this paper is to fill this gap. Taking advantage of adapted VC-type concentration inequalities, upper bounds are derived with expected rate of convergence in O(k-1/2). The concentration tools involved in this analysis rely on a more general study of maximal deviations in low probability regions, and thus directly apply to the classification of extreme data.
| Original language | English |
|---|---|
| Journal | Journal of Machine Learning Research |
| Volume | 40 |
| Issue number | 2015 |
| Publication status | Published - 1 Jan 2015 |
| Externally published | Yes |
| Event | 28th Conference on Learning Theory, COLT 2015 - Paris, France Duration: 2 Jul 2015 → 6 Jul 2015 |
Keywords
- Concentration inequalities
- Extreme data classification
- Multivariate extremes
- Stable tail dependence function
- VC theory