idt5-base-qaqg-v1.12-TydiQA-id

This model is a fine-tuned version of muchad/idt5-base on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8697
  • Rouge1: 0.5240
  • Rouge2: 0.3498
  • Rougel: 0.5227
  • Rougelsum: 0.5223
  • Bleu: 0.3432
  • Rouge All: {'rouge1': 0.5240346007536814, 'rouge2': 0.34978996264301754, 'rougeL': 0.522710963680047, 'rougeLsum': 0.5223208797539063}
  • Bleu All: {'bleu': 0.34315353770773166, 'precisions': [0.5807064513343121, 0.4227170054856405, 0.3432937644972029, 0.3066427289048474], 'brevity_penalty': 0.855878745494461, 'length_ratio': 0.8653314401016367, 'translation_length': 11579, 'reference_length': 13381}

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0001
  • train_batch_size: 4
  • eval_batch_size: 4
  • seed: 42
  • optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • num_epochs: 5

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Bleu Rouge All Bleu All
1.4087 1.0 2281 0.9875 0.4623 0.2939 0.4613 0.4606 0.3128 {'rouge1': 0.46229438264513656, 'rouge2': 0.2938529297746283, 'rougeL': 0.4612646163282077, 'rougeLsum': 0.4605582862662171} {'bleu': 0.3127745735621603, 'precisions': [0.5066238458450422, 0.352172203656379, 0.2824858757062147, 0.25564860167483017], 'brevity_penalty': 0.9283488958054709, 'length_ratio': 0.9307973992975114, 'translation_length': 12455, 'reference_length': 13381}
1.0619 2.0 4562 0.8708 0.5134 0.3414 0.5112 0.5108 0.3338 {'rouge1': 0.513394588591981, 'rouge2': 0.34136043067199673, 'rougeL': 0.5112324089779485, 'rougeLsum': 0.5108241780391356} {'bleu': 0.33378331860538335, 'precisions': [0.56817986784519, 0.4075338811226123, 0.3296777687516924, 0.29425369151396547], 'brevity_penalty': 0.8621827752794875, 'length_ratio': 0.8708616695314252, 'translation_length': 11653, 'reference_length': 13381}
0.8994 3.0 6843 0.8689 0.5195 0.3484 0.5181 0.5180 0.3432 {'rouge1': 0.5194791000460289, 'rouge2': 0.34835227420426074, 'rougeL': 0.5181144219443232, 'rougeLsum': 0.5179952388777813} {'bleu': 0.34316067855090626, 'precisions': [0.5818499127399651, 0.4252560470690782, 0.3474458634092171, 0.31538320161734973], 'brevity_penalty': 0.8456696104377565, 'length_ratio': 0.8564382333158956, 'translation_length': 11460, 'reference_length': 13381}
0.7231 4.0 9124 0.8536 0.5132 0.3361 0.5122 0.5117 0.3347 {'rouge1': 0.5131997890831617, 'rouge2': 0.33609967105633076, 'rougeL': 0.512183530122801, 'rougeLsum': 0.5117135378316945} {'bleu': 0.33466149600472417, 'precisions': [0.579715271023079, 0.42084856541486654, 0.3420865862313698, 0.31277117524995285], 'brevity_penalty': 0.8325875421766096, 'length_ratio': 0.8451535759659218, 'translation_length': 11309, 'reference_length': 13381}
0.6618 5.0 11405 0.8697 0.5240 0.3498 0.5227 0.5223 0.3432 {'rouge1': 0.5240346007536814, 'rouge2': 0.34978996264301754, 'rougeL': 0.522710963680047, 'rougeLsum': 0.5223208797539063} {'bleu': 0.34315353770773166, 'precisions': [0.5807064513343121, 0.4227170054856405, 0.3432937644972029, 0.3066427289048474], 'brevity_penalty': 0.855878745494461, 'length_ratio': 0.8653314401016367, 'translation_length': 11579, 'reference_length': 13381}

Framework versions

  • Transformers 4.57.1
  • Pytorch 2.9.0+cu129
  • Datasets 4.2.0
  • Tokenizers 0.22.1
Downloads last month
2
Safetensors
Model size
0.2B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for hawalurahman/idt5-base-qaqg-v1.12-TydiQA-id

Base model

muchad/idt5-base
Finetuned
(26)
this model