not-for-all-audiences nsfw

What is PetrolLoRA?

PetrolLoRA is the LoRA equivalent of PetrolLM, without any of the instruction-tuning of the prior.

The dataset consists of 2800 samples, with the composition as follows:

These samples were then back-filled using gpt-4/gpt-3.5-turbo-16k or otherwise converted to fit the prompt format.

Prompt Format

The LoRA was finetuned with a prompt format similar to the original SuperHOT prototype:

---
style: roleplay
characters:
  [char]: [description]
summary: [scenario]
---
<chat_history>
Format:
[char]: [message]
Human: [message]

Training procedure

The following bitsandbytes quantization config was used during training:

The following bitsandbytes quantization config was used during training:

Framework versions