Multi-Armed Bandits
Multi-armed bandits refer to a task where a fixed amount of resources must be allocated between competing resources that maximizes expected gain. Typically these problems involve an exploration/exploitation trade-off.
( Image credit: Microsoft Research )
Papers
Showing 126–150 of 1262 papers
Benchmark Results
| # | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| 1 | NeuralLinear FullPosterior-MR | Cumulative regret | 1.92 | — | Unverified |
| 2 | Linear FullPosterior-MR | Cumulative regret | 1.82 | — | Unverified |