SOTAVerified

Exploring and Addressing Reward Confusion in Offline Preference Learning

2024-07-22Unverified0· sign in to hype

Xin Chen, Sam Toyer, Florian Shkurti

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Spurious correlations in a reward model's training data can prevent Reinforcement Learning from Human Feedback (RLHF) from identifying the desired goal and induce unwanted behaviors. This paper shows that offline RLHF is susceptible to reward confusion, especially in the presence of spurious correlations in offline data. We create a benchmark to study this problem and propose a method that can significantly reduce reward confusion by leveraging transitivity of preferences while building a global preference chain with active learning.

Tasks

Reproductions