SOTAVerified

Multi-Armed Bandits

Multi-armed bandits refer to a task where a fixed amount of resources must be allocated between competing resources that maximizes expected gain. Typically these problems involve an exploration/exploitation trade-off.

( Image credit: Microsoft Research )

Papers

Showing 911920 of 1262 papers

TitleStatusHype
Decentralized Multi-player Multi-armed Bandits with No Collision Information0
Designing Truthful Contextual Multi-Armed Bandits based Sponsored Search Auctions0
Structured Linear Contextual Bandits: A Sharp and Geometric Smoothed Analysis0
Bandit Learning with Delayed Impact of Actions0
The Unreasonable Effectiveness of Greedy Algorithms in Multi-Armed Bandit with Many ArmsCode0
Survey Bandits with Regret Guarantees0
Online Learning in Contextual Bandits using Gated Linear Networks0
Residual Bootstrap Exploration for Bandit Algorithms0
On conditional versus marginal bias in multi-armed bandits0
Adaptive Estimator Selection for Off-Policy EvaluationCode0
Show:102550
← PrevPage 92 of 127Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1NeuralLinear FullPosterior-MRCumulative regret1.92Unverified
2Linear FullPosterior-MRCumulative regret1.82Unverified