The process of continuing the training of a pretrained model on a smaller, task-specific dataset to specialize its capabilities. Fine-tuning is the primary method for customizing multimodal AI models to specific domains, data types, and use cases.
Fine-tuning initializes a model with pretrained weights and trains it further on domain-specific data with a task-specific objective. The pretrained features serve as a strong starting point, and training adjusts them to better capture domain-specific patterns. This requires significantly less data and compute than training from scratch because the model already understands general patterns.
Full fine-tuning updates all model parameters on the new data. Parameter-efficient methods include LoRA (Low-Rank Adaptation, adding small trainable matrices), QLoRA (quantized LoRA for memory efficiency), prefix tuning, and adapter layers. Typical fine-tuning uses 100-100K labeled examples, 1-10 epochs, and learning rates 10-100x smaller than pretraining. Evaluation should use held-out data from the target domain to measure actual task performance.