SOTAVerified

Multi-Armed Bandits

Multi-armed bandits refer to a task where a fixed amount of resources must be allocated between competing resources that maximizes expected gain. Typically these problems involve an exploration/exploitation trade-off.

( Image credit: Microsoft Research )

Papers

Showing 11311140 of 1262 papers

TitleStatusHype
Skyline Identification in Multi-Armed Bandits0
Small-loss bounds for online learning with partial information0
Multi-Player Bandits Revisited0
Sparsity, variance and curvature in multi-armed bandits0
Medoids in almost linear time via multi-armed banditsCode0
Multi-Armed Bandits with Metric Movement Costs0
Combinatorial Multi-armed Bandits for Real-Time Strategy Games0
An Analysis of the Value of Information when Exploring Stochastic, Discrete Multi-Armed Bandits0
Trend Detection based Regret Minimization for Bandit Problems0
Optimal Learning for Sequential Decision Making for Expensive Cost Functions with Stochastic Binary Feedbacks0
Show:102550
← PrevPage 114 of 127Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1NeuralLinear FullPosterior-MRCumulative regret1.92Unverified
2Linear FullPosterior-MRCumulative regret1.82Unverified