mt5-base-finetuned-test_30483_prefix_summarize-finetuned-test_21911_prefix_summarize_V12159

This model is a fine-tuned version of emilstabil/mt5-base-finetuned-test_30483_prefix_summarize-finetuned-test_21911_prefix_summarize on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 2.6592
  • Rouge1: 28.7045
  • Rouge2: 7.6597
  • Rougel: 17.7911
  • Rougelsum: 26.2935
  • Gen Len: 89.74

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 2
  • eval_batch_size: 2
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 30

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
1.8662 1.25 500 2.3721 27.5477 7.8163 17.0749 25.3524 90.59
1.826 2.5 1000 2.3866 27.1584 7.2417 16.6686 24.744 88.96
1.7636 3.75 1500 2.3649 27.1333 8.0536 17.3248 24.7674 84.37
1.7248 5.0 2000 2.3946 26.1796 7.2833 16.7034 23.9083 82.43
1.6394 6.25 2500 2.4257 27.996 7.9177 17.3958 25.8586 89.47
1.6023 7.5 3000 2.4280 27.6978 7.8114 17.6006 25.6393 89.44
1.5606 8.75 3500 2.4460 28.2604 7.7071 17.2627 26.0593 98.12
1.5173 10.0 4000 2.4752 26.6534 7.5213 17.2009 24.433 80.49
1.4655 11.25 4500 2.4787 28.6229 8.5816 17.6534 26.2693 90.87
1.4405 12.5 5000 2.4944 28.8101 8.5897 17.964 26.6235 89.81
1.4015 13.75 5500 2.5129 27.4546 7.67 17.5582 25.2634 86.03
1.3729 15.0 6000 2.5276 28.7309 8.2089 18.2854 26.5967 91.97
1.3293 16.25 6500 2.5434 27.4011 7.4927 17.5076 25.068 84.97
1.3279 17.5 7000 2.5583 28.4385 7.6384 17.5373 26.0499 92.71
1.2727 18.75 7500 2.5929 28.9833 7.9095 17.4616 26.7396 94.69
1.2693 20.0 8000 2.5798 27.9046 7.7553 17.2599 25.705 85.16
1.2416 21.25 8500 2.6174 27.7 7.8456 17.4919 25.6451 84.51
1.2174 22.5 9000 2.6227 28.5114 7.6761 17.4532 26.1349 91.37
1.2088 23.75 9500 2.6222 28.2178 7.5203 17.5631 25.8688 87.71
1.1982 25.0 10000 2.6334 28.9268 7.7591 17.8701 26.5417 94.73
1.1839 26.25 10500 2.6403 29.1922 7.691 17.8536 26.9276 92.78
1.1589 27.5 11000 2.6541 28.9328 7.9311 17.8917 26.81 92.8
1.1633 28.75 11500 2.6658 28.7038 7.618 17.9869 26.4802 94.12
1.1605 30.0 12000 2.6592 28.7045 7.6597 17.7911 26.2935 89.74

Framework versions

  • Transformers 4.32.1
  • Pytorch 2.1.0
  • Datasets 2.12.0
  • Tokenizers 0.13.3
Downloads last month
15
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for emilstabil/mt5-base-finetuned-test_30483_prefix_summarize-finetuned-test_21911_prefix_summarize_V12159