SOTAVerified

Thompson Sampling

Thompson sampling, named after William R. Thompson, is a heuristic for choosing actions that addresses the exploration-exploitation dilemma in the multi-armed bandit problem. It consists of choosing the action that maximizes the expected reward with respect to a randomly drawn belief.

Papers

Showing 151175 of 655 papers

TitleStatusHype
Thompson Sampling for Stochastic Bandits with Noisy Contexts: An Information-Theoretic Regret Analysis0
Model-Free Approximate Bayesian Learning for Large-Scale Conversion Funnel Optimization0
Decentralized Multi-Agent Active Search and Tracking when Targets Outnumber Agents0
Improving sample efficiency of high dimensional Bayesian optimization with MCMC0
Zero-Inflated Bandits0
Finite-Time Frequentist Regret Bounds of Multi-Agent Thompson Sampling on Sparse HypergraphsCode0
Best Arm Identification in Batched Multi-armed Bandit Problems0
Bayesian Analysis of Combinatorial Gaussian Process Bandits0
RoME: A Robust Mixed-Effects Bandit Algorithm for Optimizing Mobile Health InterventionsCode0
Sample-based Dynamic Hierarchical Transformer with Layer and Head Flexibility via Contextual Bandit0
The Sliding Regret in Stochastic Bandits: Discriminating Index and Randomized Policies0
Thompson sampling for zero-inflated count outcomes with an application to the Drink Less mobile health study0
Probabilistic Inference in Reinforcement Learning Done Right0
A Distributed Neural Linear Thompson Sampling Framework to Achieve URLLC in Industrial IoT0
Adaptive Interventions with User-Defined Goals for Health Behavior ChangeCode0
Exploration via linearly perturbed loss minimisation0
Posterior Sampling-Based Bayesian Optimization with Tighter Bayesian Regret Bounds0
Batch Bayesian Optimization for Replicable Experimental Design0
Dual-Directed Algorithm Design for Efficient Pure Exploration0
Improved Bayesian Regret Bounds for Thompson Sampling in Reinforcement Learning0
Little Exploration is All You Need0
Making RL with Preference-based Feedback Efficient via Randomization0
Parallel Bayesian Optimization Using Satisficing Thompson Sampling for Time-Sensitive Black-Box Optimization0
Using Adaptive Bandit Experiments to Increase and Investigate Engagement in Mental HealthCode0
Optimal Exploration is no harder than Thompson Sampling0
Show:102550
← PrevPage 7 of 27Next →

No leaderboard results yet.