SOTAVerified

Multi-Armed Bandits

Multi-armed bandits refer to a task where a fixed amount of resources must be allocated between competing resources that maximizes expected gain. Typically these problems involve an exploration/exploitation trade-off.

( Image credit: Microsoft Research )

Papers

Showing 10911100 of 1262 papers

TitleStatusHype
Contextual bandits with surrogate losses: Margin bounds and efficient algorithms0
Greybox fuzzing as a contextual bandits problem0
Finding the bandit in a graph: Sequential search-and-stop0
Mitigating Bias in Adaptive Data Gathering via Differential Privacy0
A General Framework for Bandit Problems Beyond Cumulative Objectives0
The Externalities of Exploration and How Data Diversity Helps Exploitation0
Myopic Bayesian Design of Experiments via Posterior Sampling and Probabilistic ProgrammingCode0
Learning Contextual Bandits in a Non-stationary EnvironmentCode0
Multi-Statistic Approximate Bayesian Computation with Multi-Armed Bandits0
Bandit-Based Monte Carlo Optimization for Nearest NeighborsCode0
Show:102550
← PrevPage 110 of 127Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1NeuralLinear FullPosterior-MRCumulative regret1.92Unverified
2Linear FullPosterior-MRCumulative regret1.82Unverified