PAC Reinforcement Learning with Rich Observations Feb 8, 2016 Decision Making Multi-Armed Bandits
— Unverified 0Pairwise Elimination with Instance-Dependent Guarantees for Bandits with Cost Subsidy Jan 17, 2025 Multi-Armed Bandits
— Unverified 0Parallel Contextual Bandits in Wireless Handover Optimization Jan 21, 2019 Multi-Armed Bandits Thompson Sampling
— Unverified 0Parallelizing Contextual Bandits May 21, 2021 Decision Making Decision Making Under Uncertainty
— Unverified 0Parameterized Exploration Jul 13, 2019 Multi-Armed Bandits
— Unverified 0Partial Bandit and Semi-Bandit: Making the Most Out of Scarce Users' Feedback Sep 16, 2020 Multi-Armed Bandits Recommendation Systems
— Unverified 0Partially Observable Contextual Bandits with Linear Payoffs Sep 17, 2024 Decision Making Multi-Armed Bandits
— Unverified 0Personalization Paradox in Behavior Change Apps: Lessons from a Social Comparison-Based Personalized App for Physical Activity Jan 25, 2021 Multi-Armed Bandits
— Unverified 0Personalized Course Sequence Recommendations Dec 30, 2015 Multi-Armed Bandits
— Unverified 0Perturbed-History Exploration in Stochastic Multi-Armed Bandits Feb 26, 2019 Multi-Armed Bandits
— Unverified 0Pessimism for Offline Linear Contextual Bandits using _p Confidence Sets May 21, 2022 Multi-Armed Bandits
— Unverified 0PG-TS: Improved Thompson Sampling for Logistic Contextual Bandits May 18, 2018 Multi-Armed Bandits Thompson Sampling
— Unverified 0Phasic Diversity Optimization for Population-Based Reinforcement Learning Mar 17, 2024 Diversity MuJoCo
— Unverified 0Non-Stationary Off-Policy Optimization Jun 15, 2020 Multi-Armed Bandits
— Unverified 0Player Modeling via Multi-Armed Bandits Feb 10, 2021 Multi-Armed Bandits
— Unverified 0Policy Gradients for Contextual Recommendations Feb 12, 2018 Decision Making Multi-Armed Bandits
— Unverified 0Practical Algorithms for Best-K Identification in Multi-Armed Bandits May 19, 2017 Multi-Armed Bandits
— Unverified 0Practical Contextual Bandits with Regression Oracles Mar 3, 2018 General Classification Multi-Armed Bandits
— Unverified 0Preference-based Online Learning with Dueling Bandits: A Survey Jul 30, 2018 Multi-Armed Bandits Survey
— Unverified 0Preference-centric Bandits: Optimality of Mixtures and Regret-efficient Algorithms Apr 29, 2025 Multi-Armed Bandits Navigate
— Unverified 0Privacy Amplification via Shuffling for Linear Contextual Bandits Dec 11, 2021 Multi-Armed Bandits
— Unverified 0Privacy-Preserving Communication-Efficient Federated Multi-Armed Bandits Nov 2, 2021 Decision Making Multi-Armed Bandits
— Unverified 0Privacy-Preserving Multi-Party Contextual Bandits Oct 11, 2019 Multi-Armed Bandits Privacy Preserving
— Unverified 0Problem Dependent Reinforcement Learning Bounds Which Can Identify Bandit Structure in MDPs Nov 3, 2019 Multi-Armed Bandits reinforcement-learning
— Unverified 0Productization Challenges of Contextual Multi-Armed Bandits Jul 10, 2019 Multi-Armed Bandits
— Unverified 0Proportional Response: Contextual Bandits for Simple and Cumulative Regret Minimization Jul 5, 2023 Multi-Armed Bandits
— Unverified 0Provable Benefits of Policy Learning from Human Preferences in Contextual Bandit Problems Jul 24, 2023 Decision Making Multi-Armed Bandits
— Unverified 0Provable General Function Class Representation Learning in Multitask Bandits and MDPs May 31, 2022 Multi-Armed Bandits Reinforcement Learning (RL)
— Unverified 0Provably and Practically Efficient Neural Contextual Bandits May 31, 2022 Multi-Armed Bandits
— Unverified 0Provably Efficient High-Dimensional Bandit Learning with Batched Feedbacks Nov 22, 2023 Multi-Armed Bandits
— Unverified 0Transfer Learning with Partially Observable Offline Data via Causal Bounds Aug 7, 2023 Multi-Armed Bandits Transfer Learning
— Unverified 0Provably Efficient Reinforcement Learning for Adversarial Restless Multi-Armed Bandits with Unknown Transitions and Bandit Feedback May 2, 2024 Multi-Armed Bandits Sequential Decision Making
— Unverified 0Provably Efficient RLHF Pipeline: A Unified View from Contextual Bandits Feb 11, 2025 Computational Efficiency Multi-Armed Bandits
— Unverified 0Provably Optimal Algorithms for Generalized Linear Contextual Bandits Feb 28, 2017 Multi-Armed Bandits News Recommendation
— Unverified 0Pure Exploration in Asynchronous Federated Bandits Oct 17, 2023 Multi-Armed Bandits
— Unverified 0Pure exploration in multi-armed bandits with low rank structure using oblivious sampler Jun 28, 2023 Multi-Armed Bandits
— Unverified 0Combinatorial Pure Exploration of Causal Bandits Jun 16, 2022 Causal Inference Multi-Armed Bandits
— Unverified 0Pure Exploration under Mediators' Feedback Aug 29, 2023 Decision Making Multi-Armed Bandits
— Unverified 0QoS-Aware Multi-Armed Bandits Feb 28, 2017 Decision Making Multi-Armed Bandits
— Unverified 0Quantile Multi-Armed Bandits with 1-bit Feedback Feb 10, 2025 Multi-Armed Bandits
— Unverified 0Quantum contextual bandits and recommender systems for quantum data Jan 31, 2023 Multi-Armed Bandits Recommendation Systems
— Unverified 0Quantum Heavy-tailed Bandits Jan 23, 2023 Multi-Armed Bandits
— Unverified 0Quantum Multi-Armed Bandits and Stochastic Linear Bandits Enjoy Logarithmic Regrets May 30, 2022 Multi-Armed Bandits reinforcement-learning
— Unverified 0Query-Efficient Correlation Clustering with Noisy Oracle Feb 2, 2024 Clustering Multi-Armed Bandits
— Unverified 0Queue Scheduling with Adversarial Bandit Learning Mar 3, 2023 Multi-Armed Bandits Scheduling
— Unverified 0Quick-Draw Bandits: Quickly Optimizing in Nonstationary Environments with Extremely Many Arms May 30, 2025 Multi-Armed Bandits
— Unverified 0Raising Student Completion Rates with Adaptive Curriculum and Contextual Bandits Jul 28, 2022 Model-based Reinforcement Learning Multi-Armed Bandits
— Unverified 0Random Effect Bandits Jun 23, 2021 Multi-Armed Bandits Thompson Sampling
— Unverified 0Randomized Allocation with Nonparametric Estimation for Contextual Multi-Armed Bandits with Delayed Rewards Feb 3, 2019 Multi-Armed Bandits
— Unverified 0Randomized Greedy Learning for Non-monotone Stochastic Submodular Maximization Under Full-bandit Feedback Feb 2, 2023 Multi-Armed Bandits
— Unverified 0