generated_from_trainer text-generation OPT non-commercial dialogue chatbot ai-msgbot

pszemraj/opt-peter-2.7B

<a href="https://colab.research.google.com/gist/pszemraj/4068382a40bbf7aab50638b062bd97a9/opt-peter-2-7b-example-csearch-generation.ipynb"> <img src="https://colab.research.google.com/assets/colab-badge.svg" alt="Open In Colab"/> </a>

This model is a fine-tuned version of facebook/opt-2.7b on about 80k WhatsApp/text messages (mine). Please use responsibly :)

Test it out on Google Colab by clicking the button above.

chatdemo

Model description

sharded checkpoint

As this model file is 10+ GB, it can impose some constraints with lower RAM runtimes and/or download speeds. To help with this issue, a sharded checkpoint of this model is available here.

The pszemraj/opt-peter-2.7B-sharded model can be used as a drop-in replacement for this one for all use cases.

Intended uses & limitations

The base model has a custom license that propagates to this one. Most importantly, it cannot be used commercially. Read more here: facebook/opt-2.7b

Training and evaluation data

WhatsApp & iMessage data were parsed using ai-msgbot and then fed as a text dataset to the HF trainer.

Training procedure

Training hyperparameters

SESSION ONE

The following hyperparameters were used during training:

SESSION TWO

The following hyperparameters were used during training:

Framework versions