SOTAVerified

Multi-Armed Bandits

Multi-armed bandits refer to a task where a fixed amount of resources must be allocated between competing resources that maximizes expected gain. Typically these problems involve an exploration/exploitation trade-off.

( Image credit: Microsoft Research )

Papers

Showing 641650 of 1262 papers

TitleStatusHype
Neural Contextual Bandits for Personalized Recommendation0
Neural Contextual Bandits Under Delayed Feedback Constraints0
Reward-Biased Maximum Likelihood Estimation for Neural Contextual Bandits0
Neural Contextual Bandits with Deep Representation and Shallow Exploration0
Neural Network Retraining for Model Serving0
Neural Risk-sensitive Satisficing in Contextual Bandits0
NeuralUCB: Contextual Bandits with Neural Network-Based Exploration0
No DBA? No regret! Multi-armed bandits for index tuning of analytical and HTAP workloads with provable guarantees0
Nonlinear Sequential Accepts and Rejects for Identification of Top Arms in Stochastic Bandits0
Nonparametric Contextual Bandits in an Unknown Metric Space0
Show:102550
← PrevPage 65 of 127Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1NeuralLinear FullPosterior-MRCumulative regret1.92Unverified
2Linear FullPosterior-MRCumulative regret1.82Unverified