SOTAVerified

Multi-Armed Bandits

Multi-armed bandits refer to a task where a fixed amount of resources must be allocated between competing resources that maximizes expected gain. Typically these problems involve an exploration/exploitation trade-off.

( Image credit: Microsoft Research )

Papers

Showing 971980 of 1262 papers

TitleStatusHype
Social Learning in Multi Agent Multi Armed Bandits0
Decision Automation for Electric Power Network Recovery0
An Optimal Algorithm for Multiplayer Multi-Armed Bandits0
NeuralUCB: Contextual Bandits with Neural Network-Based Exploration0
Neural Linear Bandits: Overcoming Catastrophic Forgetting through Likelihood MatchingCode0
Learning Effective Exploration Strategies For Contextual Bandits0
Practical Calculation of Gittins Indices for Multi-armed BanditsCode0
AutoML for Contextual Bandits0
Smooth Contextual Bandits: Bridging the Parametric and Non-differentiable Regret RegimesCode0
Censored Semi-Bandits: A Framework for Resource Allocation with Censored FeedbackCode0
Show:102550
← PrevPage 98 of 127Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1NeuralLinear FullPosterior-MRCumulative regret1.92Unverified
2Linear FullPosterior-MRCumulative regret1.82Unverified