SOTAVerified

Off-policy evaluation

Off-policy Evaluation (OPE), or offline evaluation in general, evaluates the performance of hypothetical policies leveraging only offline log data. It is particularly useful in applications where the online interaction involves high stakes and expensive setting such as precision medicine and recommender systems.

Papers

Showing 141150 of 265 papers

TitleStatusHype
Designing an Interpretable Interface for Contextual Bandits0
Development and Validation of Heparin Dosing Policies Using an Offline Reinforcement Learning Algorithm0
Discovering an Aid Policy to Minimize Student Evasion Using Offline Reinforcement Learning0
Distributional Shift-Aware Off-Policy Interval Estimation: A Unified Error Quantification Framework0
Double/Debiased Machine Learning for Dynamic Treatment Effects via g-Estimation0
Double Reinforcement Learning for Efficient and Robust Off-Policy Evaluation0
Doubly Robust Bias Reduction in Infinite Horizon Off-Policy Estimation0
Doubly robust off-policy evaluation with shrinkage0
Doubly-Robust Off-Policy Evaluation with Estimated Logging Policy0
Effective Off-Policy Evaluation and Learning in Contextual Combinatorial Bandits0
Show:102550
← PrevPage 15 of 27Next →

No leaderboard results yet.