generated_from_trainer

<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. -->

22_12_13_luther_blocks_larger_fp16_20ep

This model is a fine-tuned version of stefan-it/german-gpt2-larger on a dataset of texts by Martin Luther. It achieves the following results on the evaluation set:

Model description

This is a language model used to generate wishes for a happy new year to the readers of "reformiert" a journal in Switzerland (https://www.reformiert.info)

Intended uses & limitations

This is to test the capabilities of the GPT-2 transformer architecture.

Training and evaluation data

Automatic split of an edited and "cleaned" version of parts of Luther's writing. Cleaning refers here to the process of eliminating para-texts like page numbering, footnotes, etc.

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

Training results

Training Loss Epoch Step Validation Loss Accuracy
No log 1.6 50 4.6218 0.2156
8.1175 3.22 100 4.0404 0.2633
8.1175 4.83 150 3.8120 0.2871
3.734 6.44 200 3.7062 0.2997
3.734 8.06 250 3.6382 0.3082
3.3639 9.67 300 3.6108 0.3128
3.3639 11.29 350 3.6012 0.3148
3.1363 12.89 400 3.5847 0.3168
3.1363 14.51 450 3.5914 0.3180
2.9884 16.13 500 3.5954 0.3177
2.9884 17.73 550 3.6001 0.3176
2.8748 19.35 600 3.6048 0.3188

Framework versions