Llama 2 7B quantized with AutoGPTQ V0.3.0.

This model is compatible with the first version of QA-LoRA.

To fine-tune it with QA-LoRA, follow this tutorial: Fine-tune Quantized Llama 2 on Your GPU with QA-LoRA