SOTAVerified

Efficient Exploration

Efficient Exploration is one of the main obstacles in scaling up modern deep reinforcement learning algorithms. The main challenge in Efficient Exploration is the balance between exploiting current estimates, and gaining information about poorly understood states and actions.

Source: Randomized Value Functions via Multiplicative Normalizing Flows

Papers

Showing 201250 of 514 papers

TitleStatusHype
Go-Explore for Residential Energy Management0
Mutual Enhancement of Large Language and Reinforcement Learning Models through Bi-Directional Feedback Mechanisms: A Case Study0
Diffusion-ES: Gradient-free Planning with Diffusion for Autonomous and Instruction-guided Driving0
A Bayesian Framework of Deep Reinforcement Learning for Joint O-RAN/MEC Orchestration0
TransNAS-TSAD: Harnessing Transformers for Multi-Objective Neural Architecture Search in Time Series Anomaly DetectionCode0
Joint channel estimation and data detection in massive MIMO systems based on diffusion models0
Consensus-based adaptive sampling and approximation for high-dimensional energy landscapesCode0
Virtual Action Actor-Critic Framework for Exploration (Student Abstract)0
Regret Analysis of Learning-Based Linear Quadratic Gaussian Control with Additive Exploration0
Visual Analytics for Efficient Image Exploration and User-Guided Image Captioning0
Efficient Exploration in Continuous-time Model-based Reinforcement Learning0
Grow Your Limits: Continuous Improvement with Real-World RL for Robotic Locomotion0
Co-NavGPT: Multi-Robot Cooperative Visual Semantic Navigation Using Vision Language Models0
f-Policy Gradients: A General Framework for Goal Conditioned RL using f-Divergences0
Information Content Exploration0
Learning Optimal Power Flow Value Functions with Input-Convex Neural Networks0
Feature Interaction Aware Automated Data Representation TransformationCode0
DREAM: Decentralized Reinforcement Learning for Exploration and Efficient Energy Management in Multi-Robot Systems0
Provably Efficient Exploration in Constrained Reinforcement Learning:Posterior Sampling Is All You Need0
Curiosity as a Self-Supervised Method to Improve Exploration in De novo Drug DesignCode0
Learning Spatial and Temporal Hierarchies: Hierarchical Active Inference for navigation in Multi-Room Maze Environments0
Where2Explore: Few-shot Affordance Learning for Unseen Novel Categories of Articulated Objects0
Go Beyond Imagination: Maximizing Episodic Reachability with World ModelsCode0
Reinforcement learning informed evolutionary search for autonomous systems testing0
Bag of Policies for Distributional Deep Exploration0
Towards A Unified Agent with Foundation Models0
LISSNAS: Locality-based Iterative Search Space Shrinkage for Neural Architecture Search0
Approximate information for efficient exploration-exploitation strategies0
Maximum State Entropy Exploration using Predecessor and Successor Representations0
DISCO-10M: A Large-Scale Music Dataset0
Inferring Hierarchical Structure in Multi-Room Maze Environments0
Provably Efficient Representation Learning with Tractable Planning in Low-Rank POMDP0
A Simple Unified Uncertainty-Guided Framework for Offline-to-Online Reinforcement Learning0
PACER: A Fully Push-forward-based Distributional Reinforcement Learning Algorithm0
Magnitude Attention-based Dynamic Pruning0
Symmetric Replay Training: Enhancing Sample Efficiency in Deep Reinforcement Learning for Combinatorial OptimizationCode0
Large-Batch, Iteration-Efficient Neural Bayesian Design OptimizationCode0
Discovering Failure Modes of Text-guided Diffusion Models via Adversarial Search0
EXPODE: EXploiting POlicy Discrepancy for Efficient Exploration in Multi-agent Reinforcement LearningCode0
Successor-Predecessor Intrinsic Exploration0
Solving Diffusion ODEs with Optimal Boundary Conditions for Better Image Super-Resolution0
Shattering the Agent-Environment Interface for Fine-Tuning Inclusive Language Models0
Joint Falsification and Fidelity Settings Optimization for Validation of Safety-Critical Systems: A Theoretical Analysis0
Rescue Conversations from Dead-ends: Efficient Exploration for Task-oriented Dialogue Policy Optimization0
Conditionally Optimistic Exploration for Cooperative Deep Multi-Agent Reinforcement LearningCode0
Fast exploration and learning of latent graphs with aliased observations0
Exploration of the search space of Gaussian graphical models for paired data0
Policy Mirror Descent Inherently Explores Action Space0
Exploration via Epistemic Value Estimation0
Guarded Policy Optimization with Imperfect Online Demonstrations0
Show:102550
← PrevPage 5 of 11Next →

No leaderboard results yet.