SOTAVerified

Safe Exploration

Safe Exploration is an approach to collect ground truth data by safely interacting with the environment.

Source: Chance-Constrained Trajectory Optimization for Safe Exploration and Learning of Nonlinear Systems

Papers

Showing 51100 of 135 papers

TitleStatusHype
Near-Optimal Multi-Agent Learning for Safe Coverage ControlCode1
Safe Exploration Method for Reinforcement Learning under Existence of DisturbanceCode0
Guiding Safe Exploration with Weakest Preconditions0
Safe Reinforcement Learning of Dynamic High-Dimensional Robotic Tasks: Navigation, Manipulation, Interaction0
Toward Safe and Accelerated Deep Reinforcement Learning for Next-Generation Wireless NetworksCode1
Safe Reinforcement Learning with Contrastive Risk Prediction0
Recursively Feasible Probabilistic Safe Online Learning with Control Barrier Functions0
Sample-efficient Safe Learning for Online Nonlinear Control with Control Barrier Functions0
Ablation Study of How Run Time Assurance Impacts the Training and Performance of Reinforcement Learning Agents0
A Safe Semi-supervised Graph Convolution Network0
Safe Exploration Incurs Nearly No Additional Sample Complexity for Reward-free RL0
Effects of Safety State Augmentation on Safe ExplorationCode0
Learning to Drive Using Sparse Imitation Reinforcement Learning0
Learn-to-Race Challenge 2022: Benchmarking Safe Learning and Cross-domain Generalisation in Autonomous Racing0
Exploration in Deep Reinforcement Learning: A Survey0
SCOPE: Safe Exploration for Dynamic Computer Systems Optimization0
SAAC: Safe Reinforcement Learning as an Adversarial Game of Actor-Critics0
Training and Evaluation of Deep Policies using Reinforcement Learning and Generative Models0
Safe Reinforcement Learning via Shielding under Partial Observability0
Safe Exploration for Efficient Policy Evaluation and Comparison0
CUP: A Conservative Update Policy Algorithm for Safe Reinforcement LearningCode0
Model-Based Offline Meta-Reinforcement Learning with Regularization0
Provably Efficient Primal-Dual Reinforcement Learning for CMDPs with Non-stationary Objectives and Constraints0
GoSafeOpt: Scalable Safe Exploration for Global Optimization of Dynamical SystemsCode0
MESA: Offline Meta-RL for Safe Adaptation and Fault Tolerance0
DOPE: Doubly Optimistic and Pessimistic Exploration for Safe Reinforcement LearningCode0
Safe Policy Optimization with Local Generalized Linear Function ApproximationsCode0
Infinite Time Horizon Safety of Bayesian Neural NetworksCode0
DESTA: A Framework for Safe Reinforcement Learning with Markov Games of Intervention0
Dual-Arm Adversarial Robot Learning0
Safe Exploration in Linear Equality Constraint0
MetaDrive: Composing Diverse Driving Scenarios for Generalizable Reinforcement LearningCode2
Safety-Critical Learning of Robot Control with Temporal Logic Specifications0
Bayesian Controller Fusion: Leveraging Control Priors in Deep Reinforcement Learning for Robotics0
Safe Exploration by Solving Early Terminated MDP0
Learning Policies with Zero or Bounded Constraint Violation for Constrained MDPs0
Feasible Actor-Critic: Constrained Reinforcement Learning for Ensuring Statewise SafetyCode1
Safe Exploration in Model-based Reinforcement Learning using Control Barrier Functions0
Safe Continuous Control with Constrained Model-Based Policy OptimizationCode0
Towards Safe Continuing Task Reinforcement Learning0
Safe model-based design of experiments using Gaussian processes0
Conservative Safety Critics for Exploration0
Autonomous UAV Exploration of Dynamic Environments via Incremental Sampling and Probabilistic RoadmapCode1
Align-RUDDER: Learning From Few Demonstrations by Reward RedistributionCode1
Neurosymbolic Reinforcement Learning with Formally Verified ExplorationCode1
Data-efficient visuomotor policy training using reinforcement learning and generative models0
Provably Safe PAC-MDP Exploration Using AnalogiesCode1
Verifiably Safe Exploration for End-to-End Reinforcement LearningCode1
Enforcing Almost-Sure Reachability in POMDPsCode0
Curiosity Killed or Incapacitated the Cat and the Asymptotically Optimal AgentCode0
Show:102550
← PrevPage 2 of 3Next →

No leaderboard results yet.