Training procedure

The following bitsandbytes quantization config was used during training:

Framework versions

PROMPT FORMAT

### Instruction:
<prompt>

Input
  
### Output:
We will begin by duplicating the repository and then utilize the generate.py script to test the model:

!git clone https://github.com/tloen/alpaca-lora.git
%cd alpaca-lora
!git checkout a48d947

The Gradio app launched by the script will allow us to utilize the weights of our model:

!python generate.py \
    --load_8bit \
    --base_model 'decapoda-research/llama-7b-hf' \
    --lora_weights 'Andyrasika/lora-bitcoin-tweets-sentiment' \
    --share_gradio