summarization generated_from_trainer

mt5-small-finetuned-amazon-en-es

This model is a fine-tuned version of google/mt5-small on the amazon_reviews_multi dataset. It achieves the following results on the evaluation set:

Model Description

This model is a fine-tuned version of the mT5-small model, a multilingual Transformer-based model pretrained for various NLP tasks. It has been further refined using a dataset consisting of book reviews and review titles, making it particularly well-suited for tasks similar to summarizing book reviews.

Intended Uses & Limitations

Intended Uses:

Limitations:

Training and Evaluation Data

The training and evaluation data used for fine-tuning this model encompassed a diverse collection of textual reviews and their corresponding titles. This dataset serves as the foundation for the model's text summarization capabilities. Below are the key aspects of the dataset preparation:

Baseline: Lead-3 Summarization

A common baseline for text summarization tasks is the "Lead-3" baseline, which simply extracts the first three sentences from the source text as the summary. This baseline helps provide a reference point for evaluating the model's performance. On the validation set, the Lead-3 baseline achieved the following ROUGE scores:

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum
3.4071 1.0 1209 3.1603 17.3175 8.3009 16.7074 16.755
3.0542 2.0 2418 3.1411 18.3538 9.0086 17.8745 17.8275
3.3216 3.0 3627 3.0424 15.7882 7.908 15.5215 15.5397
3.2157 4.0 4836 3.0497 15.6788 7.7739 15.3788 15.4032
3.1488 5.0 6045 3.0347 15.8221 7.8918 15.6714 15.6797
3.0838 6.0 7254 3.0254 16.2869 8.2442 16.1594 16.1527
3.0639 7.0 8463 3.0197 17.1527 8.4248 16.9826 16.9533
3.0388 8.0 9672 3.0205 16.4636 8.2233 16.3489 16.3382

Framework versions