SOTAVerified

Offline RL

Papers

Showing 5160 of 755 papers

TitleStatusHype
Q-value Regularized Transformer for Offline Reinforcement LearningCode1
Generating Code World Models with Large Language Models Guided by Monte Carlo Tree SearchCode1
Reinformer: Max-Return Sequence Modeling for Offline RLCode1
LTLDoG: Satisfying Temporally-Extended Symbolic Constraints for Safe Diffusion-based PlanningCode1
Pessimistic Value Iteration for Multi-Task Data Sharing in Offline Reinforcement LearningCode1
Stitching Sub-Trajectories with Conditional Diffusion Model for Goal-Conditioned Offline RLCode1
Entropy-regularized Diffusion Policy with Q-Ensembles for Offline Reinforcement LearningCode1
SEABO: A Simple Search-Based Method for Offline Imitation LearningCode1
Towards an Information Theoretic Framework of Context-Based Offline Meta-Reinforcement LearningCode1
ODICE: Revealing the Mystery of Distribution Correction Estimation via Orthogonal-gradient UpdateCode1
Show:102550
← PrevPage 6 of 76Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1KFCAverage Reward81.8Unverified
2ADMPOAverage Reward81Unverified
3Decision Transformer (DT)Average Reward73.5Unverified
#ModelMetricClaimedVerifiedStatus
1ParPID4RL Normalized Score151.4Unverified