PEFT (Parameter Efficient Fine-Tuning)
PEFT (Parameter-Efficient Fine-Tuning): PEFT is a technique that focuses on updating only a small subset of the model’s parameters during fine-tuning.
Text Generation • 0.3B • UpdatedNote LoRA (Low-Rank Adaptation): LoRA reduces computational demands by updating low-rank decompositions of weight matrices during fine-tuning instead of the full matrices. This minimizes trainable parameters and preserves pre-trained weights, enabling efficient task-specific adaptation. It's particularly useful in memory and resource-constrained environments.
-
LoRA: Low-Rank Adaptation of Large Language Models
Paper • 2106.09685 • Published • 53
thsluck/llm-course-hw3-dora
Text Generation • 0.3B • UpdatedNote DoRA (Weight-Decomposed Low-Rank Adaptation): DoRA fine-tunes large models by decomposing weight updates into low-rank components, separating magnitude and direction. This targeted adaptation reduces computational and memory demands, making it ideal for resource-constrained environments while maintaining performance.
-
DoRA: Weight-Decomposed Low-Rank Adaptation
Paper • 2402.09353 • Published • 30
thsluck/llm-course-hw3-tinyllama-qlora
UpdatedNote QLoRA (Quantized Low-Rank Adaptation): QLoRA efficiently fine-tunes large models by integrating weight quantization with low-rank decomposition. Quantization lowers memory usage by using reduced precision, while low-rank adaptation targets essential weight components. This combination significantly cuts down on memory and computational needs, maintaining accuracy and making it ideal for environments with limited resources.
-
QLoRA: Efficient Finetuning of Quantized LLMs
Paper • 2305.14314 • Published • 57