generated_from_trainer

📋 BUOD: bert2bert Transformer Model

Model:Bert2Bert This model is a fine-tuned version of patrickvonplaten/bert2bert-cnn_dailymail-fp16 on on KAMI-3000 for the task of Filipino Text Summarization. Bert2Bert is a EncoderDecoderModel, meaning that both the encoder and the decoder are bert-base-uncased BERT models.

It achieves the following results on the evaluation set:

🔧 Finetuning/ Training procedure

Training hyperparameters

The following hyperparameters were used during training:

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum
2.8263 1.0 586 2.4478 45.3367 18.3604 29.713 41.2805
2.1264 2.0 1172 2.3346 46.3609 18.8105 30.215 42.3642

Framework versions