generated_from_trainer

<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. -->

text_shortening_model_v1

This model is a fine-tuned version of t5-small on a dataset of 699 original-shortened texts pairs of advertising texts. It achieves the following results on the evaluation set:

Model description

Data is cleaned and preprocessed: "summarize" prefix added for each original text input.

Loss is a combination of:

Loss = theta * Custom loss + (1 - theta) * CrossEntropy

(theta = 0.3)

Intended uses & limitations

More information needed

Training and evaluation data

699 original-shortened texts pairs of advertising texts of various lengths.

Splitting amongst sub-datasets:

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Bert precision Bert recall Average word count Max word count Min word count Average token count
1.7188 1.0 8 1.9266 0.4797 0.2787 0.4325 0.4321 0.8713 0.8594 10.0714 18 1 15.45

Framework versions