SOTAVerified

Multi-Armed Bandits

Multi-armed bandits refer to a task where a fixed amount of resources must be allocated between competing resources that maximizes expected gain. Typically these problems involve an exploration/exploitation trade-off.

( Image credit: Microsoft Research )

Papers

Showing 921930 of 1262 papers

TitleStatusHype
Stochastic Bandits for Egalitarian Assignment0
Stochastic Bandits with Linear Constraints0
Stochastic Bandits with Vector Losses: Minimizing ^-Norm of Relative Losses0
Stochastic Contextual Bandits with Graph-based Contexts0
Stochastic contextual bandits with graph feedback: from independence number to MAS number0
Stochastic Contextual Bandits with Known Reward Functions0
Stochastic Contextual Bandits with Long Horizon Rewards0
Stochastic differential equations for limiting description of UCB rule for Gaussian multi-armed bandits0
Stochastic Graph Bandit Learning with Side-Observations0
Stochastic Linear Contextual Bandits with Diverse Contexts0
Show:102550
← PrevPage 93 of 127Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1NeuralLinear FullPosterior-MRCumulative regret1.92Unverified
2Linear FullPosterior-MRCumulative regret1.82Unverified