Главная
Study mode:
on
1
PEFT LoRA
2
Structural Differences
3
Rank vs Size
4
Spectral properties in SVD
5
LoRA introduces new vectors
6
MIT pre-print October 28
7
Example simplification
8
Weight tensor update functions
9
LoRA Rank stabilization
10
Key Takeaways
Description:
Explore a 28-minute technical video analyzing MIT's research on the fundamental differences between Parameter Efficient Fine-Tuning (PEFT) Low-Rank Adaptation (LoRA) and full fine-tuning methods for language models. Dive deep into the spectral properties and structural distinctions between these approaches, understanding how LoRA introduces "intruder dimensions" - new high-ranking singular vectors orthogonal to pre-trained weights - while full fine-tuning makes subtle adjustments to existing singular vectors. Learn about the trade-offs between computational efficiency and model adaptability, examining how LoRA's rank settings influence generalization behavior and retention of pre-trained knowledge. Follow along with detailed segments covering structural differences, rank versus size relationships, SVD spectral properties, weight tensor update functions, and practical examples, concluding with insights on LoRA rank stabilization and essential takeaways for implementing these fine-tuning methods. Read more

Spectral Analysis of LoRA vs Full Fine-Tuning in Language Models

Discover AI
Add to list