Abstract
We consider combinatorial semi-bandits over a set X g (0,1)d where rewards are uncorrelated across items. For this problem, the algorithm ESCB yields the smallest known regret bound R(T) = O ( d (łn m)2 (łn T) øver "min) after T rounds, where m = maxx g X 1Tx. However, ESCB has computational complexity O(|X|), which is typically exponential in d, and cannot be used in large dimensions. We propose the first algorithm that is both computationally and statistically efficient for this problem with regret R(T) = O (d (łn m)2 (łn T)øver "min) and computational asymptotic complexity O( T-1 poly (d)), where T is a function which vanishes arbitrarily slowly. Our approach involves carefully designing AESCB, an approximate version of ESCB with the same regret guarantees. We show that, whenever budgeted linear maximization over X can be solved up to a given approximation ratio, AESCB is implementable in polynomial time O (T-1 poly (d)) by repeatedly maximizing a linear function over X subject to a linear budget constraint, and showing how to solve these maximization problems efficiently. Additional algorithms, proofs and numerical experiments are given in the complete version of this work.
| Original language | English |
|---|---|
| Pages (from-to) | 5-6 |
| Number of pages | 2 |
| Journal | Performance Evaluation Review |
| Volume | 49 |
| Issue number | 1 |
| DOIs | |
| Publication status | Published - 1 Jun 2021 |
| Externally published | Yes |
Keywords
- bandits
- combinatorial bandits
- combinatorial optimization
Fingerprint
Dive into the research topics of 'Statistically Efficient, Polynomial-Time Algorithms for Combinatorial Semi-Bandits'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver