SOTAVerified

MuJoCo

Papers

Showing 101150 of 677 papers

TitleStatusHype
Simple random search provides a competitive approach to reinforcement learningCode1
SimSR: Simple Distance-based State Representation for Deep Reinforcement LearningCode1
Boosting Soft Actor-Critic: Emphasizing Recent Experience without Forgetting the PastCode1
An Equivalence between Loss Functions and Non-Uniform Sampling in Experience ReplayCode1
Generalized Decision Transformer for Offline Hindsight Information MatchingCode1
Improving Generalization in Meta-RL with Imaginary Tasks from Latent Dynamics MixtureCode1
Learning Invariant Representations for Reinforcement Learning without ReconstructionCode1
Towards Safe Reinforcement Learning via Constraining Conditional Value at RiskCode1
An Open-Source Multi-Goal Reinforcement Learning Environment for Robotic Manipulation with PybulletCode1
Trust Region Policy Optimisation in Multi-Agent Reinforcement LearningCode1
FM-TS: Flow Matching for Time Series GenerationCode1
VariBAD: A Very Good Method for Bayes-Adaptive Deep RL via Meta-LearningCode1
An Real-Sim-Real (RSR) Loop Framework for Generalizable Robotic Policy Transfer with Differentiable SimulationCode1
Wasserstein Distance guided Adversarial Imitation Learning with Reward Shape ExplorationCode1
Evolution Strategies as a Scalable Alternative to Reinforcement LearningCode1
FORK: A Forward-Looking Actor For Model-Free Reinforcement LearningCode1
Delay-Aware Model-Based Reinforcement Learning for Continuous ControlCode1
Zonal RL-RRT: Integrated RL-RRT Path Planning with Collision Probability and Zone ConnectivityCode1
Energy-Guided Diffusion Sampling for Offline-to-Online Reinforcement LearningCode1
Distributional Soft Actor-Critic: Off-Policy Reinforcement Learning for Addressing Value Estimation ErrorsCode1
Cross-Modal Domain Adaptation for Reinforcement LearningCode1
Cross-modal Domain Adaptation for Cost-Efficient Visual Reinforcement LearningCode1
Generalizable Episodic Memory for Deep Reinforcement LearningCode1
Learnings Options End-to-End for Continuous Action TasksCode1
Multi-Agent Trust Region LearningCode1
How to Learn a Useful Critic? Model-based Action-Gradient-Estimator Policy OptimizationCode1
S^2AC: Energy-Based Reinforcement Learning with Stein Soft Actor CriticCode1
Contrastive Variational Reinforcement Learning for Complex ObservationsCode1
Doubly Mild Generalization for Offline Reinforcement LearningCode1
EDGE: Explaining Deep Reinforcement Learning PoliciesCode1
A Deep Reinforcement Learning Approach to Marginalized Importance Sampling with the Successor RepresentationCode1
Disentangling Dynamics and Returns: Value Function Decomposition with Future Prediction0
Coagent Networks: Generalized and Scaled0
DisTop: Discovering a Topological representation to learn diverse and rewarding skills0
Closed Loop Interactive Embodied Reasoning for Robot Manipulation0
Addressing Distribution Shift in Online Reinforcement Learning with Offline Datasets0
Distributional Decision Transformer for Hindsight Information Matching0
CLARE: Conservative Model-Based Reward Learning for Offline Inverse Reinforcement Learning0
CKNet: A Convolutional Neural Network Based on Koopman Operator for Modeling Latent Dynamics from Pixels0
Action Redundancy in Reinforcement Learning0
CIM-PPO:Proximal Policy Optimization with Liu-Correntropy Induced Metric0
A Pontryagin Perspective on Reinforcement Learning0
DIDA: Denoised Imitation Learning based on Domain Adaptation0
Distributionally Robust Statistical Verification with Imprecise Neural Networks0
C-GAIL: Stabilizing Generative Adversarial Imitation Learning with Control Theory0
Modular Recurrence in Contextual MDPs for Universal Morphology Control0
CEIL: Generalized Contextual Imitation Learning0
CAT-SAC: Soft Actor-Critic with Curiosity-Aware Entropy Temperature0
Detecting and Mitigating Reward Hacking in Reinforcement Learning Systems: A Comprehensive Empirical Study0
CasIL: Cognizing and Imitating Skills via a Dual Cognition-Action Architecture0
Show:102550
← PrevPage 3 of 14Next →

No leaderboard results yet.