A Closer Look at Advantage-Filtered Behavioral Cloning in High-Noise Datasets
Jake Grigsby, Yanjun Qi
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/jakegrigsby/deep_controlOfficialIn paperpytorch★ 106
- github.com/jakegrigsby/super_sacOfficialIn paperpytorch★ 41
- github.com/jakegrigsby/cc-afbcOfficialIn paperpytorch★ 4
Abstract
Recent Offline Reinforcement Learning methods have succeeded in learning high-performance policies from fixed datasets of experience. A particularly effective approach learns to first identify and then mimic optimal decision-making strategies. Our work evaluates this method's ability to scale to vast datasets consisting almost entirely of sub-optimal noise. A thorough investigation on a custom benchmark helps identify several key challenges involved in learning from high-noise datasets. We re-purpose prioritized experience sampling to locate expert-level demonstrations among millions of low-performance samples. This modification enables offline agents to learn state-of-the-art policies in benchmark tasks using datasets where expert actions are outnumbered nearly 65:1.