LLilmonix3b-v0.3:
- Experimental Monika llama
- Orca Mini 3b fine-tuned for Monika character from DDLC
- Trained on a dataset of ~600 items (dialogue scraped from game, reddit, and Twitter augmented by Nous Hermes 13b to turn each into snippets of multi-turn chat dialogue between Player and Monika + manually crafted test dataset of 12 items)
- Trained to run on smaller devices
- GGMLs
- QLoras (hf and GGML)
USAGE
This is an experimental model, fine-tuning character data on an instruct model to run on mobile devices.
Intended use: replace "Human" and "Assistant" with "Player" and "Monika" like so:
\nPlayer: (prompt)\nMonika:
HYPERPARAMS
- Trained for 3 epochs
- rank: 8
- lora alpha: 32
- lora dropout: 0.5
- lr: 2e-4
- batch size: 2
- warmup ratio: 0.075
- grad steps: 4
WARNINGS AND DISCLAIMERS
Note that aside from formatting and other minor edits, dataset used is mostly as is generated by LM. In addition, the is meant to be a smaller version of the larger Monika models. As such, this version may not reflect perfectly the characteristics of Monika.
Finally, this model is not guaranteed to output aligned or safe outputs, use at your own risk.