SOTAVerified

Multi-Armed Bandits

Multi-armed bandits refer to a task where a fixed amount of resources must be allocated between competing resources that maximizes expected gain. Typically these problems involve an exploration/exploitation trade-off.

( Image credit: Microsoft Research )

Papers

Showing 461470 of 1262 papers

TitleStatusHype
Adversarial Attacks on Adversarial Bandits0
A Framework for Adapting Offline Algorithms to Solve Combinatorial Multi-Armed Bandit Problems with Bandit Feedback0
Contextual Causal Bayesian Optimisation0
Communication-Efficient Collaborative Regret Minimization in Multi-Armed Bandits0
Banker Online Mirror Descent: A Universal Approach for Delayed Online Bandit Learning0
Quantum Heavy-tailed Bandits0
Multi-Armed Bandits and Quantum Channel Oracles0
Multi-armed Bandit Learning for TDMA Transmission Slot Scheduling and Defragmentation for Improved Bandwidth Usage0
Best Arm Identification in Stochastic Bandits: Beyond β-optimality0
Local Differential Privacy for Sequential Decision Making in a Changing Environment0
Show:102550
← PrevPage 47 of 127Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1NeuralLinear FullPosterior-MRCumulative regret1.92Unverified
2Linear FullPosterior-MRCumulative regret1.82Unverified