generated_from_trainer

<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. -->

<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>

Kimiko-Mistral-7B

(I am going to retrain this, this model is a failure) This model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on the Kimiko dataset. It achieves the following results on the evaluation set:

Model description

Same dataset as Kimiko-v2 but on new model. THIS IS NOT TRAIN ON V3 DATASET

Intended uses & limitations

As a finetuning experiment on new 7B model. You can use this for roleplay or as an assistant

Prompt Template Structure

This is a chat between ASSISTANT and USER
USER: What is 4x8?
ASSISTANT:

Training hyperparameters

The following hyperparameters were used during training:

Training results

Training Loss Epoch Step Validation Loss
1.5675 0.47 25 2.1323
1.4721 0.95 50 2.1209
1.472 1.42 75 2.1177
1.5445 1.9 100 2.1173

Framework versions