<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. -->
bart-base-finetuned-xsum
This model is a fine-tuned version of facebook/bart-base on an unknown dataset. It achieves the following results on the evaluation set:
- Loss: 2.2029
- Rouge1: 31.9426
- Rouge2: 16.4413
- Rougel: 26.9094
- Rougelsum: 27.0631
- Gen Len: 19.9551
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 8
Training results
Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
---|---|---|---|---|---|---|---|---|
2.5212 | 1.0 | 668 | 2.1627 | 31.1883 | 15.4628 | 25.7196 | 25.9468 | 19.7635 |
2.1882 | 2.0 | 1336 | 2.1573 | 32.0929 | 16.022 | 26.3848 | 26.5647 | 19.7994 |
1.9206 | 3.0 | 2004 | 2.1447 | 32.1449 | 15.6461 | 26.4625 | 26.6191 | 19.8802 |
1.7712 | 4.0 | 2672 | 2.1602 | 31.4352 | 15.675 | 26.656 | 26.7887 | 19.8443 |
1.6686 | 5.0 | 3340 | 2.1731 | 32.189 | 16.4349 | 27.0203 | 27.104 | 19.8922 |
1.5526 | 6.0 | 4008 | 2.1775 | 32.4991 | 16.8565 | 27.4952 | 27.7015 | 19.9012 |
1.4832 | 7.0 | 4676 | 2.1964 | 31.8655 | 16.3919 | 26.9015 | 27.0513 | 19.8982 |
1.4485 | 8.0 | 5344 | 2.2029 | 31.9426 | 16.4413 | 26.9094 | 27.0631 | 19.9551 |
Framework versions
- Transformers 4.33.0
- Pytorch 2.0.1+cu118
- Datasets 2.14.4
- Tokenizers 0.13.3