SOTAVerified

Multi-Armed Bandits

Multi-armed bandits refer to a task where a fixed amount of resources must be allocated between competing resources that maximizes expected gain. Typically these problems involve an exploration/exploitation trade-off.

( Image credit: Microsoft Research )

Papers

Showing 11911200 of 1262 papers

TitleStatusHype
BISTRO: An Efficient Relaxation-Based Method for Contextual Bandits0
Bandits meet Computer Architecture: Designing a Smartly-allocated Cache0
Personalized Course Sequence Recommendations0
On Top-k Selection in Multi-Armed Bandits and Hidden Bipartite Graphs0
Algorithms for Differentially Private Multi-Armed Bandits0
Regret Analysis of the Finite-Horizon Gittins Index Strategy for Multi-Armed Bandits0
Context-Aware Bandits0
A Survey of Online Experiment Design with the Stochastic Multi-Armed BanditCode0
Multi-armed Bandits with Application to 5G Small Cells0
Sequential Design for Ranking Response Surfaces0
Show:102550
← PrevPage 120 of 127Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1NeuralLinear FullPosterior-MRCumulative regret1.92Unverified
2Linear FullPosterior-MRCumulative regret1.82Unverified