SOTAVerified

Multi-Armed Bandits

Multi-armed bandits refer to a task where a fixed amount of resources must be allocated between competing resources that maximizes expected gain. Typically these problems involve an exploration/exploitation trade-off.

( Image credit: Microsoft Research )

Papers

Showing 861870 of 1262 papers

TitleStatusHype
Multiple-Play Stochastic Bandits with Shareable Finite-Capacity Arms0
Multiplier Bootstrap-based Exploration0
MultiScale Contextual Bandits for Long Term Objectives0
Multi-Statistic Approximate Bayesian Computation with Multi-Armed Bandits0
Multi-Task Learning for Contextual Bandits0
Multi-User MABs with User Dependent Rewards for Uncoordinated Spectrum Access0
Multi-User Multi-Armed Bandits for Uncoordinated Spectrum Access0
Navigating the Rashomon Effect: How Personalization Can Help Adjust Interpretable Machine Learning Models to Individual Users0
Nearest Neighbor Search Under Uncertainty0
Nearly Minimax-Optimal Regret for Linearly Parameterized Bandits0
Show:102550
← PrevPage 87 of 127Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1NeuralLinear FullPosterior-MRCumulative regret1.92Unverified
2Linear FullPosterior-MRCumulative regret1.82Unverified