SOTAVerified

Multi-Armed Bandits

Multi-armed bandits refer to a task where a fixed amount of resources must be allocated between competing resources that maximizes expected gain. Typically these problems involve an exploration/exploitation trade-off.

( Image credit: Microsoft Research )

Papers

Showing 621630 of 1262 papers

TitleStatusHype
Scalable Decision-Focused Learning in Restless Multi-Armed Bandits with Application to Maternal and Child Health0
Context Uncertainty in Contextual Bandits with Applications to Recommender Systems0
Evaluating Deep Vs. Wide & Deep Learners As Contextual Bandits For Personalized Email Promo RecommendationsCode0
Neural Collaborative Filtering Bandits via Meta Learning0
Optimal Regret Is Achievable with Bounded Approximate Inference Error: An Enhanced Bayesian Upper Confidence Bound FrameworkCode0
Coordinated Attacks against Contextual Bandits: Fundamental Limits and Defense Mechanisms0
Networked Restless Multi-Armed Bandits for Mobile Interventions0
Top-K Ranking Deep Contextual Bandits for Information Selection Systems0
Adaptive Best-of-Both-Worlds Algorithm for Heavy-Tailed Multi-Armed Bandits0
Learning Neural Contextual Bandits Through Perturbed Rewards0
Show:102550
← PrevPage 63 of 127Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1NeuralLinear FullPosterior-MRCumulative regret1.92Unverified
2Linear FullPosterior-MRCumulative regret1.82Unverified