xlm-roberta-bambara-01

This model is a fine-tuned version of oza75/xlm-roberta-bambara-base-01 on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 1.1160
  • Accuracy: 0.7599

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 32
  • eval_batch_size: 32
  • seed: 42
  • gradient_accumulation_steps: 2
  • total_train_batch_size: 64
  • optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 0.06
  • num_epochs: 100.0

Training results

Training Loss Epoch Step Validation Loss Accuracy
2.9609 1.8458 1000 1.4658 0.6942
2.8570 3.6907 2000 1.4204 0.7014
2.7516 5.5355 3000 1.3654 0.7116
2.6739 7.3804 4000 1.3460 0.7167
2.6104 9.2253 5000 1.3267 0.7195
2.5437 11.0702 6000 1.3044 0.7239
2.5264 12.9160 7000 1.2751 0.7296
2.4831 14.7608 8000 1.2648 0.7314
2.4471 16.6057 9000 1.2492 0.7345
2.4112 18.4506 10000 1.2466 0.7352
2.3945 20.2955 11000 1.2400 0.7364
2.3746 22.1404 12000 1.2290 0.7380
2.3774 23.9861 13000 1.2189 0.7400
2.3462 25.8310 14000 1.2187 0.7391
2.3205 27.6759 15000 1.2137 0.7405
2.3026 29.5208 16000 1.2014 0.7425
2.2770 31.3657 17000 1.1973 0.7446
2.2654 33.2105 18000 1.1902 0.7461
2.2580 35.0554 19000 1.1864 0.7461
2.2299 36.9012 20000 1.1864 0.7455
2.2272 38.7461 21000 1.1784 0.7480
2.2221 40.5910 22000 1.1819 0.7470
2.2076 42.4358 23000 1.1661 0.7496
2.1860 44.2807 24000 1.1729 0.7494
2.1880 46.1256 25000 1.1617 0.7510
2.1815 47.9714 26000 1.1598 0.7516
2.1620 49.8163 27000 1.1568 0.7512
2.1703 51.6611 28000 1.1516 0.7525
2.1563 53.5060 29000 1.1529 0.7518
2.1396 55.3509 30000 1.1537 0.7524
2.1182 57.1958 31000 1.1410 0.7544
2.1404 59.0406 32000 1.1506 0.7538
2.1401 60.8864 33000 1.1422 0.7540
2.1146 62.7313 34000 1.1428 0.7538
2.1252 64.5762 35000 1.1358 0.7544
2.1011 66.4211 36000 1.1356 0.7565
2.1142 68.2659 37000 1.1321 0.7557
2.0950 70.1108 38000 1.1362 0.7562
2.0930 71.9566 39000 1.1387 0.7556
2.0881 73.8015 40000 1.1272 0.7573
2.0858 75.6464 41000 1.1327 0.7566
2.0640 77.4912 42000 1.1252 0.7575
2.0456 79.3361 43000 1.1283 0.7579
2.0573 81.1810 44000 1.1283 0.7565
2.0496 83.0259 45000 1.1317 0.7568
2.0701 84.8717 46000 1.1320 0.7567
2.0779 86.7165 47000 1.1219 0.7579
2.0462 88.5614 48000 1.1201 0.7580
2.0726 90.4063 49000 1.1282 0.7580
2.0605 92.2512 50000 1.1221 0.7578
2.0428 94.0960 51000 1.1202 0.7580
2.0565 95.9418 52000 1.1143 0.7592
2.0513 97.7867 53000 1.1258 0.7579
2.0514 99.6316 54000 1.1242 0.7577

Framework versions

  • Transformers 5.0.1.dev0
  • Pytorch 2.4.1+cu124
  • Datasets 4.5.0
  • Tokenizers 0.22.2
Downloads last month
1
Safetensors
Model size
0.3B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for oza75/xlm-roberta-bambara-01

Finetuned
(1)
this model