SOTAVerified

Multi-Armed Bandits

Multi-armed bandits refer to a task where a fixed amount of resources must be allocated between competing resources that maximizes expected gain. Typically these problems involve an exploration/exploitation trade-off.

( Image credit: Microsoft Research )

Papers

Showing 421430 of 1262 papers

TitleStatusHype
Efficient Contextual Bandits with Uninformed Feedback Graphs0
Cost-Efficient Distributed Learning via Combinatorial Multi-Armed Bandits0
Delegating via Quitting Games0
Delay-Adaptive Learning in Generalized Linear Contextual Bandits0
An Analysis of the Value of Information when Exploring Stochastic, Discrete Multi-Armed Bandits0
Efficient Generalized Low-Rank Tensor Contextual Bandits0
Efficient Implementation of LinearUCB through Algorithmic Improvements and Vector Computing Acceleration for Embedded Learning Systems0
Deep Upper Confidence Bound Algorithm for Contextual Bandit Ranking of Information Selection0
Deep Contextual Bandits for Fast Initial Access in mmWave Based User-Centric Ultra-Dense Networks0
Deep Neural Linear Bandits: Overcoming Catastrophic Forgetting through Likelihood Matching0
Show:102550
← PrevPage 43 of 127Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1NeuralLinear FullPosterior-MRCumulative regret1.92Unverified
2Linear FullPosterior-MRCumulative regret1.82Unverified