- Home
- /
- TIL
- /
- Deep Learning
LoRA vs. Fine-tuning
Table of Contents
Generated by Google Gemini.
Traditional Fine-tuning
- Involves updating all or a large portion of the pre-trained model’s weights during training on the new dataset.
- Aims to adapt the entire model’s knowledge to the target task.
- Can be more computationally expensive and require more data.
LoRA (Low-Rank Adaptation)
- Modifies only a small set of parameters (rank decomposition matrices) within specific layers.
- Focuses on adapting the model’s behavior by injecting new information, rather than overwriting existing knowledge.
- More efficient in terms of memory and training time.
So, is it appropriate to call it “fine-tuning”?
- In a broad sense, yes. LoRA is a technique for adapting a pre-trained model to a new task, which aligns with the general goal of fine-tuning.
- However, it’s more precise to refer to it as “LoRA fine-tuning” or “fine-tuning with LoRA” to distinguish it from traditional full fine-tuning. This clarifies that you’re using a specific method that differs from updating all weights.
Ultimately, the terminology can be flexible, but providing context is key
- If you’re discussing LoRA among researchers or practitioners familiar with the technique, simply saying “fine-tuning” might be understood.
- In other contexts, it’s best to be more specific and explicitly mention LoRA to avoid ambiguity and ensure clear communication.