text-generation gpt2 gpt

<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. -->

pszemraj/gpt2-medium-vaguely-human-dialogue

This model is a fine-tuned version of gpt2-medium on a parsed version of Wizard of Wikipedia. Because the batch size was so large, it learned a general understanding of words that makes sense together but does not specifically respond to anything - sort of like an alien learning to imitate human words to convince others that it is human.

It achieves the following results on the evaluation set:

Model description

Intended uses & limitations

Training and evaluation data

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

Training results

Training Loss Epoch Step Validation Loss
34.991 1.0 837 14.8359
12.2881 2.0 1674 9.375
8.5071 3.0 2511 7.2148
7.6031 4.0 3348 6.1758
6.4808 5.0 4185 5.5820
5.8562 6.0 5022 5.0977
5.6094 7.0 5859 4.8203
5.2591 8.0 6696 4.5977
5.0031 9.0 7533 4.4219
4.8837 10.0 8370 4.3281

Framework versions