The Analogy
A pretrained LLM is like a brilliant scholar who only knows how to finish sentences. Ask “What is the capital of France?” and it might continue with “What is the capital of Germany? What is the capital of Italy?” — because in its training data, questions are often followed by more questions (like a quiz). It has all the knowledge but doesn’t know it should answer the question. Fine-tuning teaches it the format of being helpful.
Key insight: Fine-tuning doesn’t add new knowledge — it unlocks existing knowledge by teaching the model how to respond. InstructGPT (Ouyang et al., 2022) showed that a 1.3B fine-tuned model was preferred by humans over a 175B base model. The knowledge was already there; it just needed to be formatted as helpful responses rather than text completions.
Base Model vs Fine-Tuned
Base Model (Llama 3 8B)
User: What is photosynthesis?
Model: What is cellular respiration? What is the difference between aerobic and anaerobic respiration? What are the stages of...
(Continues generating quiz questions)
Fine-Tuned (Llama 3 8B Instruct)
User: What is photosynthesis?
Model: Photosynthesis is the process by which plants convert sunlight, water, and CO&sub2; into glucose and oxygen. It occurs in chloroplasts...
(Helpful, structured answer)
The training pipeline: Pretraining (Ch 6) → SFT (this chapter) → RLHF/DPO (Ch 8). Pretraining gives knowledge. SFT teaches format. RLHF teaches quality and safety. Each stage uses less data but has outsized impact on behavior.