SOTAVerified

Multi-Armed Bandits

Multi-armed bandits refer to a task where a fixed amount of resources must be allocated between competing resources that maximizes expected gain. Typically these problems involve an exploration/exploitation trade-off.

( Image credit: Microsoft Research )

Papers

Showing 631640 of 1262 papers

TitleStatusHype
Nearly-tight Approximation Guarantees for the Improving Multi-Armed Bandits Problem0
Nearly Tight Bounds for Cross-Learning Contextual Bandits with Graphical Feedback0
Nearly Tight Bounds for Exploration in Streaming Multi-armed Bandits with Known Optimality Gap0
Near Optimal Best Arm Identification for Clustered Bandits0
Near-Optimal Private Learning in Linear Contextual Bandits0
Networked Restless Multi-Armed Bandits for Mobile Interventions0
Networked Stochastic Multi-Armed Bandits with Combinatorial Strategies0
Neural Bandit with Arm Group Graph0
Neural Collaborative Filtering Bandits via Meta Learning0
Neural Contextual Bandits Based Dynamic Sensor Selection for Low-Power Body-Area Networks0
Show:102550
← PrevPage 64 of 127Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1NeuralLinear FullPosterior-MRCumulative regret1.92Unverified
2Linear FullPosterior-MRCumulative regret1.82Unverified