SOTAVerified

Multi-Armed Bandits

Multi-armed bandits refer to a task where a fixed amount of resources must be allocated between competing resources that maximizes expected gain. Typically these problems involve an exploration/exploitation trade-off.

( Image credit: Microsoft Research )

Papers

Showing 881890 of 1262 papers

TitleStatusHype
Contextual Bandits with Side-Observations0
Concurrent Decentralized Channel Allocation and Access Point Selection using Multi-Armed Bandits in multi BSS WLANs0
(Locally) Differentially Private Combinatorial Semi-Bandits0
Locally Differentially Private (Contextual) Bandits LearningCode0
To update or not to update? Delayed Nonparametric Bandits with Randomized Allocation0
Greedy Algorithm almost Dominates in Smoothed Contextual Bandits0
Unified Models of Human Behavioral Agents in Bandits, Contextual Bandits and RLCode1
Neural Network Retraining for Model Serving0
Learning to Rank in the Position Based Model with Bandit Feedback0
Thompson Sampling for Linearly Constrained BanditsCode0
Show:102550
← PrevPage 89 of 127Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1NeuralLinear FullPosterior-MRCumulative regret1.92Unverified
2Linear FullPosterior-MRCumulative regret1.82Unverified