A simple custom QLoRA implementation for fine-tuning a language model (LLM) with basic tools such as PyTorch and Bitsandbytes, completely decoupled from Hugging Face.
-
Updated
Jan 29, 2024 - Python
A simple custom QLoRA implementation for fine-tuning a language model (LLM) with basic tools such as PyTorch and Bitsandbytes, completely decoupled from Hugging Face.
Text Generation on Pre-Trained NousResearch’s Llama-2-7b-chat-hf using guanaco-llama2-1k dataset
Add a description, image, and links to the 4bit-quantize topic page so that developers can more easily learn about it.
To associate your repository with the 4bit-quantize topic, visit your repo's landing page and select "manage topics."