References¶
PEFT and Fine-Tuning¶
- Hu, E. J., et al. LoRA: Low-Rank Adaptation of Large Language Models (2021). arXiv:2106.09685
- Dettmers, T., et al. QLoRA: Efficient Finetuning of Quantized LLMs (2023). arXiv:2305.14314
- Hugging Face PEFT Documentation: https://huggingface.co/docs/peft
- bitsandbytes Library: https://github.com/TimDettmers/bitsandbytes
- EmergentMind. 4-bit NormalFloat (NF4) Quantization. Link