SOTAVerified

Imitation Learning

Imitation Learning is a framework for learning a behavior policy from demonstrations. Usually, demonstrations are presented in the form of state-action trajectories, with each pair indicating the action to take at the state being visited. In order to learn the behavior policy, the demonstrated actions are usually utilized in two ways. The first, known as Behavior Cloning (BC), treats the action as the target label for each state, and then learns a generalized mapping from states to actions in a supervised manner. Another way, known as Inverse Reinforcement Learning (IRL), views the demonstrated actions as a sequence of decisions, and aims at finding a reward/cost function under which the demonstrated decisions are optimal.

Finally, a newer methodology, Inverse Q-Learning aims at directly learning Q-functions from expert data, implicitly representing rewards, under which the optimal policy can be given as a Boltzmann distribution similar to soft Q-learning

Source: Learning to Imitate

Papers

Showing 541550 of 2122 papers

TitleStatusHype
SENSOR: Imitate Third-Person Expert's Behaviors via Active Sensoring0
Imitation Game: A Model-based and Imitation Learning Deep Reinforcement Learning Hybrid0
Offline Imitation Learning from Multiple Baselines with Applications to Compiler Optimization0
Keypoint Action Tokens Enable In-Context Imitation Learning in Robotics0
RiEMann: Near Real-Time SE(3)-Equivariant Robot Manipulation without Point Cloud Segmentation0
Human-compatible driving partners through data-regularized self-play reinforcement learningCode1
LORD: Large Models based Opposite Reward Design for Autonomous Driving0
Uncertainty-Aware Deployment of Pre-trained Language-Conditioned Imitation Learning PoliciesCode0
LASIL: Learner-Aware Supervised Imitation Learning For Long-term Microscopic Traffic SimulationCode0
Imitating Cost-Constrained Behaviors in Reinforcement LearningCode0
Show:102550
← PrevPage 55 of 213Next →

No leaderboard results yet.