Overfitting
- Definition
- When a model performs well on training data but poorly on new, unseen data because it memorized patterns rather than learning generalizable features. Overfitting is a constant risk in fine-tuning and custom model development.
- Why it matters
- Overfitting is the most common failure mode in custom model training. A model that has memorized its training examples will ace your eval set but fail in production, creating a dangerous false sense of confidence. The risk is especially high when fine-tuning on small datasets, which is exactly the situation most enterprises face: they have hundreds or thousands of examples, not millions. Detecting overfitting requires disciplined train/eval splits and monitoring for divergence between training and evaluation metrics. Preventing it requires regularization techniques, data augmentation, and knowing when to stop training. Teams that skip these steps waste their fine-tuning budget on models that do not generalize.
- In practice
- In practice, overfitting manifests as a model that perfectly handles scenarios it saw during training but breaks on slight variations. For example, a customer support model fine-tuned on 500 examples might perfectly recite answers to those 500 questions but fail to generalize to the 501st. LoRA mitigates overfitting risk by limiting the number of trainable parameters. Standard practice includes: holding out 10-20% of data for evaluation, monitoring eval loss for divergence from training loss, using early stopping when eval performance plateaus, and testing on fully held-out examples that were not part of any training decision. Dropout and weight decay remain important regularization tools.
We cover data & training every week.
Get the 5 AI stories that matter — free, every Friday.
Related terms
Fine-tuning
The process of continuing to train a pre-trained model on a smaller, task-specific dataset. Fine-tuning customizes model behavior for specific domains or formats and is a key part of most enterprise AI deployments.
Catastrophic forgetting
When a model loses previously learned knowledge after being trained on new data. Catastrophic forgetting is a key challenge in continual learning and one reason fine-tuning must be done carefully.
Pre-training
The initial phase of model training where the network learns general knowledge from a massive dataset. Pre-training is the most expensive phase, often costing tens or hundreds of millions of dollars for frontier models.
Evals
Systematic evaluation frameworks that measure AI model performance on specific tasks relevant to your use case, going beyond generic benchmarks to test the behaviors that actually matter for your application.
Know the terms. Know the moves.
Get the 5 AI stories that matter every Friday — free.
Free forever. No spam.