Training procedure

The following bitsandbytes quantization config was used during training:

Model Description

For more information on how it was created, check out the following link: https://github.com/DunnBC22/NLP_Projects/blob/main/OPT%20Models/Essays%20With%20Instructions%20-%20Fine-Tune%20-%20OPT.ipynb

Intended uses & limitations

This is intended to show the possibilities. It is mainly limited by the input data.

Training & Evaluation Dataset

Dataset Source: https://huggingface.co/datasets/ChristophSchuhmann/essays-with-instructions

Hyperparameters Used

Hyperperameter Value
Model Checkpoint facebook/opt-2.7b
per_device_train_batch_size 8
gradient_accumulation_steps 4
fp16 True
warmup_steps 75
learning_rate 2e-4
Training Steps 150

Framework versions

Library Version
Python 3.10.1
Torch 2.0.1+cu118
Datasets 2.14.4
Transformer 4.31.0
PEFT 0.4.0

Metric

Perplexity = 9.46