— Ch. 1 · The Adaptation Process —
Fine-tuning (deep learning).
~3 min read · Ch. 1 of 6
Fine-tuning begins when a neural network trained for one task meets new data. The original model, known as the upstream task, must now perform a different downstream task. This process reuses knowledge from the initial training objective to solve specific problems. Engineers apply additional training on fresh datasets to adjust the parameters of the pre-trained system. They can choose to update every single layer or freeze some parts entirely. When layers are frozen, they do not change during backpropagation. Some architectures insert lightweight modules called adapters into the existing structure. These adapters nudge the embedding space for domain adaptation without altering the core weights.
Efficient Parameter Updates
Low-rank adaptation LoRA emerged as a technique to fine-tune models with minimal parameter updates. Designers create a low-rank matrix and add it to the original matrix. An adapter in this context is a collection of these matrices that produces a fine-tuned model when combined with a base model. A language model containing billions of parameters might be LoRA fine-tuned using only several millions of parameters. This approach allows performance to approach full-model fine-tuning while requiring significantly less storage space. Support for LoRA was integrated into the diffusers library from Hugging Face. Similar techniques are available through Hugging Face's parameter-efficient fine-tuning package for a wide range of other models.