SOTAVerified

Multi-Armed Bandits

Multi-armed bandits refer to a task where a fixed amount of resources must be allocated between competing resources that maximizes expected gain. Typically these problems involve an exploration/exploitation trade-off.

( Image credit: Microsoft Research )

Papers

Showing 421430 of 1262 papers

TitleStatusHype
Improving Fairness in Adaptive Social Exergames via Shapley Bandits0
Practical Contextual Bandits with Feedback Graphs0
Infinite Action Contextual Bandits with Reusable Data ExhaustCode0
Bandit Social Learning: Exploration under Myopic Behavior0
Genetic multi-armed bandits: a reinforcement learning approach for discrete optimization via simulation0
Adversarial Rewards in Universal Learning for Contextual Bandits0
Piecewise-Stationary Multi-Objective Multi-Armed Bandit with Application to Joint Communications and SensingCode0
Leveraging User-Triggered Supervision in Contextual Bandits0
On Private and Robust Bandits0
Multiplier Bootstrap-based Exploration0
Show:102550
← PrevPage 43 of 127Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1NeuralLinear FullPosterior-MRCumulative regret1.92Unverified
2Linear FullPosterior-MRCumulative regret1.82Unverified