just testing for now, qlora merge, several things different between this and the 7b
training
NousResearch/Llama-2-13b-hf tuned on koishi dataset (commit c83d922) for 1 epoch
then tuned on pippa dataset (commit 6412b0c) for 1 epoch
then tuned on geepeetee4 dataset (commit c83d922) for 1 epoch
then tuned on limarp (without ponyville, lolicit, and all the fallen subsets. Version 2023-09-14) for 2 epochs
all in metharme format
prompting
https://rentry.org/tsukasa13b - reccomended prompts and gen settings
The current model version has been trained on prompts using three different roles, which are denoted by the following tokens: <|system|>, <|user|> and <|model|>.
The <|system|> prompt can be used to inject out-of-channel information behind the scenes, while the <|user|> prompt should be used to indicate user input. The <|model|> token should then be used to indicate that the model should generate a response. These tokens can happen multiple times and be chained up to form a conversation history.