Expand description
LoRA (Low-Rank Adaptation)
Implements parameter-efficient fine-tuning:
- LoRA: Low-Rank Adaptation of Large Language Models
- QLoRA: Quantized LoRA for even more efficiency
- Adapter layers with low-rank decomposition
- Merge and unmerge LoRA weights
Structs§
- LoRA
Config - LoRA configuration
- LoRA
Linear - LoRA layer wrapping a linear layer
- QLoRA
Config - QLoRA (Quantized LoRA) configuration
- QLoRA
Linear - QLoRA layer with quantized base weights
Enums§
- Quant
Type - Quantization type for QLoRA