Training procedure

The following bitsandbytes quantization config was used during training:

Framework versions

I'm NOT the author of this work.

I cite anon :

Storytelling-V2 Qlora. Trained on base Llama-2-13B, works on every L2 13B.
150.5MB of books. Over ten thousand 4096 token samples.
*** for separating chapters, ⁂ for separating books.

Credit to "anon49"