SOTAVerified

D4RL

Papers

Showing 150 of 226 papers

TitleStatusHype
Flow Q-LearningCode3
CORL: Research-oriented Deep Offline Reinforcement Learning LibraryCode3
Skill Expansion and Composition in Parameter SpaceCode2
Datasets and Benchmarks for Offline Safe Reinforcement LearningCode2
Diffusion Policies as an Expressive Policy Class for Offline Reinforcement LearningCode2
Flowformer: Linearizing Transformers with Conservation FlowsCode2
Online Decision TransformerCode2
Rethinking Attention with PerformersCode2
D4RL: Datasets for Deep Data-Driven Reinforcement LearningCode2
Reformer: The Efficient TransformerCode2
Habitizing Diffusion Planning for Efficient and Effective Decision MakingCode1
Are Expressive Models Truly Necessary for Offline RL?Code1
M^3PC: Test-time Model Predictive Control for Pretrained Masked Trajectory ModelCode1
Aligning Diffusion Behaviors with Q-functions for Efficient Continuous ControlCode1
PlanDQ: Hierarchical Plan Orchestration via D-Conductor and Q-PerformerCode1
Strategically Conservative Q-LearningCode1
Diffusion Actor-Critic: Formulating Constrained Policy Iteration as Diffusion Noise Regression for Offline Reinforcement LearningCode1
In-Context Decision Transformer: Reinforcement Learning via Hierarchical Chain-of-ThoughtCode1
Diffusion Policies creating a Trust Region for Offline Reinforcement LearningCode1
Adaptive Advantage-Guided Policy Regularization for Offline Reinforcement LearningCode1
Q-value Regularized Transformer for Offline Reinforcement LearningCode1
Reinformer: Max-Return Sequence Modeling for Offline RLCode1
SEABO: A Simple Search-Based Method for Offline Imitation LearningCode1
Entropy-regularized Diffusion Policy with Q-Ensembles for Offline Reinforcement LearningCode1
Exploration and Anti-Exploration with Distributional Random Network DistillationCode1
Critic-Guided Decision Transformer for Offline Reinforcement LearningCode1
Train Once, Get a Family: State-Adaptive Balances for Offline-to-Online Reinforcement LearningCode1
CROP: Conservative Reward for Model-based Offline Policy OptimizationCode1
Offline Retraining for Online RL: Decoupled Policy Learning to Mitigate Exploration BiasCode1
Score Regularized Policy Optimization through Diffusion BehaviorCode1
Beyond Uniform Sampling: Offline Reinforcement Learning with Imbalanced DatasetsCode1
Reasoning with Latent Diffusion in Offline Reinforcement LearningCode1
Model-Bellman Inconsistency for Model-based Offline Reinforcement LearningCode1
Katakomba: Tools and Benchmarks for Data-Driven NetHackCode1
Curricular Subgoals for Inverse Reinforcement LearningCode1
Improving and Benchmarking Offline Reinforcement Learning AlgorithmsCode1
Efficient Diffusion Policies for Offline Reinforcement LearningCode1
Primal-Attention: Self-attention through Asymmetric Kernel SVD in Primal RepresentationCode1
When should we prefer Decision Transformers for Offline Reinforcement Learning?Code1
Revisiting the Minimalist Approach to Offline Reinforcement LearningCode1
Contrastive Energy Prediction for Exact Energy-Guided Diffusion Sampling in Offline Reinforcement LearningCode1
Offline RL with No OOD Actions: In-Sample Learning via Implicit Value RegularizationCode1
Optimal Transport for Offline Imitation LearningCode1
Behavior Proximal Policy OptimizationCode1
When Demonstrations Meet Generative World Models: A Maximum Likelihood Framework for Offline Inverse Reinforcement LearningCode1
Anti-Exploration by Random Network DistillationCode1
Extreme Q-Learning: MaxEnt RL without EntropyCode1
Adaptive Behavior Cloning Regularization for Stable Offline-to-Online Reinforcement LearningCode1
A Policy-Guided Imitation Approach for Offline Reinforcement LearningCode1
Semi-Supervised Offline Reinforcement Learning with Action-Free TrajectoriesCode1
Show:102550
← PrevPage 1 of 5Next →

No leaderboard results yet.