Sign in
Scaling & Data

Fine-tuning

Adapting a pre-trained model to a specific task or style with targeted training

What it is

Fine-tuning is the process of continuing training on a pre-trained model with a smaller, task-specific dataset. Rather than training from scratch, you start from a model that already understands language and update its weights to specialize for your use case.

Common applications: adjusting model style and persona, adding domain-specific knowledge, improving performance on a narrow task, and safety fine-tuning. Parameter-efficient fine-tuning methods like LoRA update only a small fraction of weights, making fine-tuning feasible on consumer hardware.

The risk: catastrophic forgetting, fine-tuning on new data can overwrite previously learned capabilities if not done carefully.

Why it matters

Fine-tuning is the practical alternative to prompting when prompt engineering hits its limits. If a client needs a model that consistently responds in a specific format, maintains a brand voice, or performs reliably on a narrow task, fine-tuning may be the right tool. Understanding when to fine-tune vs. prompt engineer vs. use RAG is a key architectural decision in AI product development.

Related concepts

Resources

What is Fine-Tuning?
ibm.com· Covers full fine-tuning vs parameter-efficient methods, supervised vs RLHF, practical use cases
10 min
Fine-tuning (deep learning)
en.wikipedia.org· Good overview, conceptual framework, common techniques, relationship to transfer learning
8 min
Finetuning Large Language Models
deeplearning.ai· Covers what fine-tuning is, when it's helpful, how it differs from prompt engineering and RAG, and instruction fine-tuning
60 min
RAG vs fine-tuning vs. prompt engineering
ibm.com· Compares all three approaches, what each does, requirements, ideal use cases, how they affect outputs
8 min
LLMs: Fine-tuning, distillation, and prompt engineering
developers.google.com· Official Google course material, foundation models → fine-tuning → distillation → prompt engineering in clear progression
10 min
PreviousSynthetic Data
NextEnd of section