SOTAVerified

Efficient Exploration

Efficient Exploration is one of the main obstacles in scaling up modern deep reinforcement learning algorithms. The main challenge in Efficient Exploration is the balance between exploiting current estimates, and gaining information about poorly understood states and actions.

Source: Randomized Value Functions via Multiplicative Normalizing Flows

Papers

Showing 251300 of 514 papers

TitleStatusHype
Efficient Exploration Using Extra Safety Budget in Constrained Policy Optimization0
Efficient Informed Proposals for Discrete Distributions via Newton's Series Approximation0
Provably Efficient Exploration in Quantum Reinforcement Learning with Logarithmic Worst-Case Regret0
Efficient Exploration via Epistemic-Risk-Seeking Policy Optimization0
Ensemble Value Functions for Efficient Exploration in Multi-Agent Reinforcement Learning0
Computational Discovery of Microstructured Composites with Optimal Stiffness-Toughness Trade-Offs0
GFlowNets for AI-Driven Scientific Discovery0
ESC: Exploration with Soft Commonsense Constraints for Zero-shot Object Navigation0
Embodied Agents for Efficient Exploration and Smart Scene Description0
Exploration in Model-based Reinforcement Learning with Randomized Reward0
Strangeness-driven Exploration in Multi-Agent Reinforcement LearningCode0
SHIRO: Soft Hierarchical Reinforcement Learning0
Reinforcement Learning in Credit Scoring and Underwriting0
Efficient Exploration in Resource-Restricted Reinforcement Learning0
Learn to Explore: on Bootstrapping Interactive Data Exploration with Meta-learning0
CURO: Curriculum Learning for Relative Overgeneralization0
HEAT: Hardware-Efficient Automatic Tensor Decomposition for Transformer Compression0
Safe and Efficient Reinforcement Learning Using Disturbance-Observer-Based Control Barrier Functions0
CIM: Constrained Intrinsic Motivation for Sparse-Reward Continuous Control0
Efficient Exploration using Model-Based Quality-Diversity with Gradients0
Active Exploration based on Information Gain by Particle Filter for Efficient Spatial Concept Formation0
Exploring through Random Curiosity with General Value FunctionsCode0
Planning to the Information Horizon of BAMDPs via Epistemic State Abstraction0
Design of Convolutional Extreme Learning Machines for Vision-Based Navigation Around Small Bodies0
Deep Active Ensemble Sampling For Image Classification0
LECO: Learnable Episodic Count for Task-Specific Intrinsic RewardCode0
The Role of Coverage in Online Reinforcement Learning0
Deterministic Sequencing of Exploration and Exploitation for Reinforcement Learning0
Self-supervised Sequential Information Bottleneck for Robust Exploration in Deep Reinforcement Learning0
An Empirical Evaluation of Posterior Sampling for Constrained Reinforcement LearningCode0
Impact Makes a Sound and Sound Makes an Impact: Sound Guides Representations and ExplorationsCode0
Learning Action Translator for Meta Reinforcement Learning on Sparse-Reward Tasks0
SkexGen: Autoregressive Generation of CAD Construction Sequences with Disentangled Codebooks0
The split Gibbs sampler revisited: improvements to its algorithmic structure and augmented target distributionCode0
GAN-based Intrinsic Exploration For Sample Efficient Reinforcement Learning0
Curious Exploration via Structured World Models Yields Zero-Shot Object Manipulation0
Scalable Exploration for Neural Online Learning to Rank with Perturbed Feedback0
Sample-Efficient, Exploration-Based Policy Optimisation for Routing Problems0
On Preemption and Learning in Stochastic SchedulingCode0
Personalized Algorithmic Recourse with Preference ElicitationCode0
SFP: State-free Priors for Exploration in Off-Policy Reinforcement Learning0
Feature and Instance Joint Selection: A Reinforcement Learning Perspective0
Fire Burns, Sword Cuts: Commonsense Inductive Bias for Exploration in Text-based GamesCode0
On Machine Learning-Driven Surrogates for Sound Transmission Loss SimulationsCode0
A Variational Approach to Bayesian Phylogenetic InferenceCode0
Efficient Exploration via First-Person Behavior Cloning Assisted Rapidly-Exploring Random Trees0
TANDEM: Learning Joint Exploration and Decision Making with Tactile Sensors0
Collaborative Training of Heterogeneous Reinforcement Learning Agents in Environments with Sparse Rewards: What and When to Share?Code0
Learning Causal Overhypotheses through Exploration in Children and Computational Models0
A Unified Perspective on Value Backup and Exploration in Monte-Carlo Tree Search0
Show:102550
← PrevPage 6 of 11Next →

No leaderboard results yet.