1 Introduction
Reinforcement Learning from Human Feedback (RLHF) integrates human preferences into AI systems to solve objectives that are hard to specify directly, especially in interactive settings. It rose to prominence with chat assistants and is now a core part of post-training—the suite of methods applied after large-scale pretraining to make models useful, safe, and engaging. Within post-training, instruction/supervised fine-tuning teaches format and basic behaviors, preference fine-tuning (where RLHF dominates) aligns models to subtle human values and style, and reinforcement learning with verifiable rewards targets domains with objective signals. Framed this way, post-training is about eliciting and shaping the latent capabilities of a pretrained base model into reliable, user-centered performance.
The chapter outlines the standard RLHF pipeline—train an instruction-following model, collect preference data to build a reward model, then optimize the policy against that reward (or learn directly from preferences via “direct alignment” methods). RLHF operates at the response level with contrastive signals, guiding models toward better answers and away from worse ones. This shifts outputs from unfocused next-token continuations to concise, helpful, well-formatted, and often warmer responses, and it tends to generalize more broadly than instruction tuning alone. The approach is powerful but delicate: reward models are proxies, optimization can overfit (e.g., to length), and success depends on strong starting models, high-quality data, and careful regularization—making RLHF more complex and costly than simple fine-tuning.
Historically, open efforts first leaned on instruction tuning before skepticism about RLHF gave way to widespread adoption, including simplified preference-optimization methods such as Direct Preference Optimization. Meanwhile, large closed labs advanced multi-stage post-training pipelines at scale, highlighting that effective systems combine instruction tuning, RLHF, prompt and format design, and increasingly, verifiable-reward RL and reasoning-focused training. The chapter situates RLHF as both a mature pillar of preference fine-tuning and a bridge to newer reinforcement-learning approaches, while emphasizing practical goals: clarify trade-offs, provide starting points for implementation, and equip readers with the intuition and tools to contribute to modern post-training in research and industry.
A rendition of the early, three stage RLHF process: first training via supervised fine-tuning (SFT, chapter 4), building a reward model (RM, chapter 5), and then optimizing with reinforcement learning (RL, chapter 6).
Summary
- RLHF incorporates human preferences into AI systems to solve problems that are hard-to-specify programmatically, and became widely known through ChatGPT’s breakout, which made the capabilities of language models more approachable.
- The basic RLHF pipeline has three steps: instruction fine-tuning to teach the model to follow the question-answering format, training a reward model on human preferences, and optimizing the model with RL against that reward.
- RLHF is known to primarily change the style, tone, and format of model responses – making them more helpful, warm, and engaging. But it’s not “just style transfer”: RLHF also improves benchmark performance, though over-optimization (e.g., excessive length or chattiness) can harm capabilities in other domains.
- The elicitation theory of post-training suggests that base models contain latent potential, and post-training’s job is to extract and cultivate that intelligence into useful behaviors.
- RLHF is one component of modern post-training, alongside instruction fine-tuning (IFT/SFT) and reinforcement learning with verifiable rewards (RLVR), used together in an intertwined manner to craft particular training recipes.
FAQ
What is RLHF and why did it become important?
Reinforcement Learning from Human Feedback (RLHF) integrates human preferences into AI systems, helping solve hard-to-specify objectives that arise in human-facing applications. It became widely known through ChatGPT, where preference-guided optimization made language models more helpful, safe, and usable across many tasks.How did RLHF originate and where was it first successful?
Early RLHF efforts targeted classic RL/control settings and then moved to language tasks like summarization, instruction following, web question-answering, and broader “alignment.” These successes showed that simple preference signals can guide powerful models toward desired behavior.What is the classic three-stage RLHF pipeline?
The basic pipeline: (1) train an instruction-following model via supervised fine-tuning (SFT/IFT), (2) collect human preference data to train a reward model of preferences, and (3) optimize the policy with RL by sampling responses and using the reward model to guide updates.How does RLHF fit into modern post-training?
Post-training is a multi-stage process that typically includes: (1) Instruction/Supervised Fine-Tuning (IFT/SFT) to teach format and basic behaviors, (2) Preference Fine-Tuning (PreFT)—where RLHF largely lives—to align style and subtle human preferences, and (3) Reinforcement Learning with Verifiable Rewards (RLVR) for tasks with checkable outcomes.What does RLHF change about a model’s outputs?
RLHF shapes response-level behavior—improving qualities like reliability, helpfulness, and warmth—and teaches preferred styles and formats. Instead of merely predicting the next token, models learn which whole responses are better or worse, leading to concise, user-oriented answers that generalize across domains.How does RLHF differ from instruction fine-tuning (SFT/IFT)?
SFT updates per token to imitate target text, emphasizing specific features and formats. RLHF uses preference signals at the response level, applying contrastive objectives to prefer better completions and avoid worse ones. This generally yields stronger cross-domain generalization and more robust behavior.What are the main challenges and costs of RLHF?
- Reward models are proxies for human goals and can be mis-specified- Optimization can overfit to the proxy (“over-optimization”), requiring regularization
- Known issues include length bias and noisy preference data
- It is more expensive than SFT in compute, data collection, and engineering time, and it benefits from a strong base model
The RLHF Book ebook for free