generated_from_keras_callback

chunwoolee0/distilroberta-base-finetuned-wikitext2

This model is a fine-tuned version of distilroberta-base on an wikitext,wikitext-2-raw-v1 dataset. It achieves the following results on the evaluation set:

Model description

This model is a distilled version of the RoBERTa-base model. It follows the same training procedure as DistilBERT.

Intended uses & limitations

This is an exercise for finetuning of nlp language modeling for fill-mask.

Training and evaluation data

Wikitext, wikitext-2-raw-v1 is used

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

Training results

Train Loss Validation Loss Epoch
2.1557 1.8964 0

Framework versions