Training procedure

This adapter has been fine-tuned using quantization-aware LoRA (QA-LoRA).

More details on the training procedure here: Fine-tune Quantized Llama 2 on Your GPU with QA-LoRA

The base model was quantized with AutoGPTQ INT4. You can find it here:

kaitchup/Llama-2-7b-4bit-32g-autogptq

Framework versions