SOTAVerified

Inference-Time Language Model Alignment via Integrated Value Guidance

2024-09-26Unverified0· sign in to hype

Zhixuan Liu, Zhanhui Zhou, Yuanfu Wang, Chao Yang, Yu Qiao

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Large language models are typically fine-tuned to align with human preferences, but tuning large models is computationally intensive and complex. In this work, we introduce Integrated Value Guidance (IVG), a method that uses implicit and explicit value functions to guide language model decoding at token and chunk-level respectively, efficiently aligning large language models purely at inference time. This approach circumvents the complexities of direct fine-tuning and outperforms traditional methods. Empirically, we demonstrate the versatility of IVG across various tasks. In controlled sentiment generation and summarization tasks, our method significantly improves the alignment of large models using inference-time guidance from gpt2-based value functions. Moreover, in a more challenging instruction-following benchmark AlpacaEval 2.0, we show that both specifically tuned and off-the-shelf value functions greatly improve the length-controlled win rates of large models against gpt-4-turbo (e.g., 19.51\% 26.51\% for Mistral-7B-Instruct-v0.2 and 25.58\% 33.75\% for Mixtral-8x7B-Instruct-v0.1 with Tulu guidance).

Tasks

Reproductions