mt5-small-finetuned-amazon-en-es

This model is a fine-tuned version of google/mt5-small on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 3.0163
  • Rouge1: 16.4647
  • Rouge2: 7.4149
  • Rougel: 16.2555
  • Rougelsum: 16.1836

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5.6e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 8

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum
3.688 1.0 1209 3.2134 17.0085 8.8944 16.3547 16.4524
3.6702 2.0 2418 3.0993 17.397 8.6213 17.0284 16.9829
3.4419 3.0 3627 3.0667 16.9732 7.4209 16.3507 16.3457
3.3136 4.0 4836 3.0511 17.6261 8.0141 16.9829 17.0179
3.2299 5.0 6045 3.0376 16.6415 7.1214 16.2463 16.2489
3.171 6.0 7254 3.0240 16.7934 7.3099 16.4251 16.3639
3.1366 7.0 8463 3.0216 16.488 7.2649 16.282 16.1983
3.1123 8.0 9672 3.0163 16.4647 7.4149 16.2555 16.1836

Framework versions

  • Transformers 4.40.0
  • Pytorch 2.5.1+cu124
  • Datasets 4.8.4
  • Tokenizers 0.19.1
Downloads last month
147
Safetensors
Model size
0.3B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Sennnnnn/mt5-small-finetuned-amazon-en-es

Base model

google/mt5-small
Finetuned
(675)
this model