SOTAVerified

Multi-Armed Bandits

Multi-armed bandits refer to a task where a fixed amount of resources must be allocated between competing resources that maximizes expected gain. Typically these problems involve an exploration/exploitation trade-off.

( Image credit: Microsoft Research )

Papers

Showing 241250 of 1262 papers

TitleStatusHype
Combinatorial Pure Exploration of Multi-Armed Bandits0
Combinatorial Pure Exploration with Full-bandit Feedback and Beyond: Solving Combinatorial Optimization under Uncertainty with Limited Observation0
Combinatorial Semi-Bandits with Knapsacks0
Combining Difficulty Ranking with Multi-Armed Bandits to Sequence Educational Content0
A Survey of Learning in Multiagent Environments: Dealing with Non-Stationarity0
Combining Online Learning and Offline Learning for Contextual Bandits with Deficient Support0
Adversarial Bandits with Knapsacks0
Communication Efficient Distributed Learning for Kernelized Contextual Bandits0
Comparative Performance of Collaborative Bandit Algorithms: Effect of Sparsity and Exploration Intensity0
A framework for optimizing COVID-19 testing policy using a Multi Armed Bandit approach0
Show:102550
← PrevPage 25 of 127Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1NeuralLinear FullPosterior-MRCumulative regret1.92Unverified
2Linear FullPosterior-MRCumulative regret1.82Unverified