| Compositional Conservatism: A Transductive Approach in Offline Reinforcement Learning | Apr 6, 2024 | D4RLOffline RL | CodeCode Available | 0 | 5 |
| Learning on One Mode: Addressing Multi-Modality in Offline Reinforcement Learning | Dec 4, 2024 | D4RLImitation Learning | CodeCode Available | 0 | 5 |
| Directly Forecasting Belief for Reinforcement Learning with Delays | May 1, 2025 | D4RLMuJoCo | CodeCode Available | 0 | 5 |
| Learning from Sparse Offline Datasets via Conservative Density Estimation | Jan 16, 2024 | D4RLDensity Estimation | CodeCode Available | 0 | 5 |
| Learning to Trust Bellman Updates: Selective State-Adaptive Regularization for Offline RL | May 26, 2025 | D4RLOffline RL | CodeCode Available | 0 | 5 |
| CAWR: Corruption-Averse Advantage-Weighted Regression for Robust Policy Optimization | Jun 18, 2025 | D4RLOffline RL | CodeCode Available | 0 | 5 |
| Diffusion Models as Optimizers for Efficient Planning in Offline RL | Jul 23, 2024 | D4RLDecision Making | CodeCode Available | 0 | 5 |
| NetworkGym: Reinforcement Learning Environments for Multi-Access Traffic Management in Network Simulation | Oct 30, 2024 | D4RLManagement | CodeCode Available | 0 | 5 |
| Q-Distribution guided Q-learning for offline reinforcement learning: Uncertainty penalized Q-value via consistency model | Oct 27, 2024 | D4RLQ-Learning | CodeCode Available | 0 | 5 |
| Why So Pessimistic? Estimating Uncertainties for Offline RL through Ensembles, and Why Their Independence Matters | May 27, 2022 | D4RLOffline RL | CodeCode Available | 0 | 5 |