One-line definition
The post-training pipeline turning a base LLM (fluent, raw next-token predictor) into a useful assistant. Standard 2026 stack: supervised fine-tuning (SFT) -> preference optimization (DPO or RLHF) plus optional safety / RL.
Why it matters
A base LLM trained on internet text is fluent but not useful. It will continue your prompt as if it were the next bit of internet text, not as a helpful answerer. Post-training is what turns the base model into Claude, ChatGPT, or any other production assistant.
Understanding the post-training stack matters because:
- Most of the differences between LLMs in 2026 come from post-training, not from architecture or pretraining.
- Many production fine-tuning decisions are about which post-training stage to invest in.
- Interviewers test this because it’s the live frontier of LLM development.
The stages
Stage 0: Pretraining (for context)
Train on trillions of tokens of internet text with next-token prediction. Produces a base model that’s fluent in many domains but doesn’t follow instructions or refuse harmful requests.
Stage 1: Supervised Fine-Tuning (SFT)
Train the base model on examples of (prompt, ideal response) pairs. The model learns the format of being a helpful assistant.
- Data: 10K-1M instruction-response pairs, often a mix of human-written and model-generated.
- Loss: standard cross-entropy on the response tokens (mask the prompt).
- Effect: model now follows instructions, uses formatting, refuses obvious bad requests.
- Cost: a few thousand GPU-hours at most for a 70B model; trivially feasible.
The quality of SFT data is the dominant factor. A small (~10K) high-quality SFT dataset typically beats a large (~1M) noisy one.
Stage 2: Preference optimization
Train the SFT model to prefer responses that humans (or model judges) prefer over alternatives.
Two main approaches:
RLHF (Reinforcement Learning from Human Feedback): the original (Christiano et al. 2017, OpenAI 2022):
- Collect pairs (prompt, response_A, response_B) with a human label of which is better.
- Train a reward model to predict preferences.
- Use PPO (or similar RL algorithm) to fine-tune the LLM to maximize the reward, with a KL penalty to stay close to the SFT model.
Pros: well-studied, flexible. Cons: complex pipeline, RL stability issues, requires careful reward model maintenance.
DPO (Direct Preference Optimization) (Rafailov et al. 2023):
- Collect the same pairwise preference data.
- Optimize the LLM directly with a closed-form loss derived from the RLHF objective.
Pros: no separate reward model, no RL, much simpler pipeline. Cons: requires offline preference data; can be less stable than RLHF in some regimes.
The 2026 picture: DPO and its variants (IPO, KTO, ORPO) have become the default for most teams because of simplicity. Big labs (Anthropic, OpenAI) still use RL-based methods for the most polished models, but the gap is narrowing.
Stage 3 (optional): On-policy RL with verifiable rewards
For tasks where correctness is checkable (math, code, factual Q&A), use RL with the verifier as the reward signal:
- Sample multiple responses from the model.
- Verify each (run the code, check the math, look up the fact).
- Use the verification signal as a reward (PPO, GRPO, or similar).
This is what produced the recent jumps in math/reasoning capability (DeepSeek-R1 style training, OpenAI’s o-series, etc.). Not yet universal but clearly the direction the frontier is going.
Stage 4 (optional): Safety and constitutional training
Additional fine-tuning specifically on safety, refusal behavior, and adherence to a “constitution” (a set of principles the model should follow). Anthropic uses this prominently; most other labs have analogues.
What an interviewer expects you to say
If asked about the LLM post-training stack:
- Distinguish pretraining from post-training.
- Describe SFT (instruction tuning) as the first step.
- Describe preference optimization (RLHF or DPO) as the second step.
- Mention DPO has largely displaced RLHF in 2026 due to simplicity.
- Mention that on-policy RL with verifiable rewards is the new frontier for reasoning capabilities.
Bonus: discuss data quality (SFT data quality >> SFT data quantity), reward hacking (models gaming the reward model), and KL penalties (preventing the policy from deviating too far from the SFT model).
DPO vs RLHF in detail
The DPO loss is:
L_DPO = -log sigmoid(beta * [log(pi_theta(y_w|x) / pi_ref(y_w|x)) - log(pi_theta(y_l|x) / pi_ref(y_l|x))])
where y_w is the winning response, y_l is the losing response, pi_theta is the policy, pi_ref is the SFT model, beta is a temperature.
Intuition: increase the model’s likelihood of y_w relative to y_l, normalized by the SFT model’s likelihoods (the implicit reward). The KL constraint to the SFT model is built into the loss form.
vs RLHF: separate reward model, sampling from the policy, gradient updates via PPO with the reward model providing rewards. More moving parts; more flexible.
The empirical picture: comparable quality, much simpler implementation. For most teams, DPO is the right choice unless you have specific reasons to need RL.
Common confusions
- “RLHF is required for alignment.” Just SFT goes a long way. Most of the helpful-assistant behavior comes from SFT. RLHF/DPO adds polish.
- “DPO is just RLHF without RL.” True high-level, but DPO has different optimization dynamics, different data requirements (offline pairs vs on-policy samples), and different failure modes.
- “Alignment fixes hallucinations.” It changes refusal patterns and helpfulness but doesn’t eliminate hallucinations. Hallucinations require separate techniques (RAG, verification).
- “You can’t fine-tune RLHF’d models.” You can; the standard pattern is SFT + DPO can be redone or extended. But it’s tricky, further fine-tuning can degrade alignment.
Open problems in 2026
- Reward hacking: model finds ways to get high reward that don’t correspond to actual quality (e.g., sycophancy, length bias, format gaming).
- Calibration: post-trained models are typically less calibrated than base models. Hard to recover.
- Long-horizon RL: hard to do RL on tasks requiring many tokens of reasoning before the reward is observable.
- Generalization of safety: a model trained to refuse one category of harmful request may or may not generalize to related categories.
Why interviewers ask
This question tests:
- Whether you’ve kept up with the field (the stack changed substantially in 2023-2024 with DPO and verifiable rewards).
- Whether you understand the why behind each stage (raw model needs SFT for format; SFT model needs preference optimization for nuance).
- Whether you can discuss reward hacking and other operational concerns.
Senior LLM-team interviews often probe deeply on this, because it’s where the interesting work is happening in 2026.
Related: Transformer architecture, How would you evaluate an LLM application?, Calibration.