SOTAVerified

Multi-Armed Bandits

Multi-armed bandits refer to a task where a fixed amount of resources must be allocated between competing resources that maximizes expected gain. Typically these problems involve an exploration/exploitation trade-off.

( Image credit: Microsoft Research )

Papers

Showing 601610 of 1262 papers

TitleStatusHype
Multi-armed Bandits for Link Configuration in Millimeter-wave Networks0
Efficient Algorithms for Learning to Control Bandits with Unobserved Contexts0
Adaptive Experimentation with Delayed Binary FeedbackCode0
Scalable Decision-Focused Learning in Restless Multi-Armed Bandits with Application to Maternal and Child Health0
Context Uncertainty in Contextual Bandits with Applications to Recommender Systems0
Optimal Regret Is Achievable with Bounded Approximate Inference Error: An Enhanced Bayesian Upper Confidence Bound FrameworkCode0
Evaluating Deep Vs. Wide & Deep Learners As Contextual Bandits For Personalized Email Promo RecommendationsCode0
Neural Collaborative Filtering Bandits via Meta Learning0
Coordinated Attacks against Contextual Bandits: Fundamental Limits and Defense Mechanisms0
Networked Restless Multi-Armed Bandits for Mobile Interventions0
Show:102550
← PrevPage 61 of 127Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1NeuralLinear FullPosterior-MRCumulative regret1.92Unverified
2Linear FullPosterior-MRCumulative regret1.82Unverified