Vit-GPT2-UCA-UCF-04

This model is a fine-tuned version of NourFakih/Vit-GPT2-COCO2017Flickr-85k-09 on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5019
  • Rouge1: 26.9427
  • Rouge2: 7.5369
  • Rougel: 22.8442
  • Rougelsum: 23.3737
  • Gen Len: 15.8150

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 4
  • eval_batch_size: 4
  • seed: 42
  • optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • num_epochs: 18

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
0.1528 1.1710 500 0.2809 26.7658 7.6288 22.7989 23.3636 15.8422
0.0701 2.3419 1000 0.3340 26.0581 6.9431 22.1094 22.5715 15.7842
0.0394 3.5129 1500 0.3520 25.9985 7.2471 22.2292 22.6947 15.9556
0.024 4.6838 2000 0.3995 26.9298 8.4206 22.9054 23.3302 15.0592
0.0144 5.8548 2500 0.4325 25.4177 6.9663 21.3437 21.7631 14.6091
0.0104 7.0258 3000 0.4389 26.6544 7.3818 22.81 23.0804 15.4291
0.0067 8.1967 3500 0.4620 26.6154 7.5924 22.7463 23.0765 15.6745
0.005 9.3677 4000 0.4657 27.7378 7.6741 23.69 24.1869 15.6424
0.0037 10.5386 4500 0.4729 27.5305 7.6016 23.2043 23.6397 16.7053
0.0069 11.7096 5000 0.4756 27.5112 7.8019 23.6743 24.2136 15.3255
0.0027 12.8806 5500 0.4899 26.6969 7.4515 22.8885 23.2666 15.2996
0.0024 14.0515 6000 0.4887 26.5269 7.1568 22.6349 23.0376 15.8138
0.0018 15.2225 6500 0.4937 26.9342 7.3399 23.1986 23.6486 15.3317
0.0016 16.3934 7000 0.5019 27.1042 7.4545 23.1031 23.5834 15.6942
0.0014 17.5644 7500 0.5019 26.9427 7.5369 22.8442 23.3737 15.8150

Framework versions

  • Transformers 4.47.0
  • Pytorch 2.5.1+cu121
  • Datasets 3.3.1
  • Tokenizers 0.21.0
Downloads last month
3
Safetensors
Model size
0.2B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for NourFakih/Vit-GPT2-UCA-UCF-04

Unable to build the model tree, the base model loops to the model itself. Learn more.