tibetan-CS-detector_mbert-tibetan-continual-wylie_all_data_no_labels_no_partial

This model is a fine-tuned version of OMRIDRORI/mbert-tibetan-continual-wylie-final on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 17.6968
  • Accuracy: 0.9266
  • Switch Precision: 0.5017
  • Switch Recall: 0.9091
  • Switch F1: 0.6466
  • True Switches: 165
  • Pred Switches: 299
  • Exact Matches: 144
  • Proximity Matches: 6
  • To Auto Precision: 0.6372
  • To Auto Recall: 0.9
  • To Allo Precision: 0.4194
  • To Allo Recall: 0.9176
  • True To Auto: 80
  • True To Allo: 85
  • Matched To Auto: 72
  • Matched To Allo: 78

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 1e-05
  • train_batch_size: 4
  • eval_batch_size: 4
  • seed: 42
  • gradient_accumulation_steps: 4
  • total_train_batch_size: 16
  • optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 200
  • num_epochs: 35
  • mixed_precision_training: Native AMP
  • label_smoothing_factor: 0.05

Training results

Training Loss Epoch Step Validation Loss Accuracy Switch Precision Switch Recall Switch F1 True Switches Pred Switches Exact Matches Proximity Matches To Auto Precision To Auto Recall To Allo Precision To Allo Recall True To Auto True To Allo Matched To Auto Matched To Allo
9.3149 1.5789 30 3.6306 0.5806 0.0 0.0 0.0 165 2 0 0 0.0 0.0 0.0 0.0 80 85 0 0
4.2352 3.1579 60 3.0508 0.7829 0.0 0.0 0.0 165 1 0 0 0.0 0.0 0.0 0.0 80 85 0 0
3.8336 4.7368 90 2.6432 0.7946 0.7963 0.2606 0.3927 165 54 42 1 0.82 0.5125 0.5 0.0235 80 85 41 2
7.8192 6.3158 120 3.5075 0.7951 0.4525 0.4909 0.4709 165 179 79 2 0.67 0.8375 0.1772 0.1647 80 85 67 14
13.6795 7.8947 150 3.5598 0.8145 0.5064 0.4788 0.4922 165 156 77 2 0.75 0.7875 0.2222 0.1882 80 85 63 16
16.8577 9.4737 180 2.6579 0.8798 0.5753 0.5091 0.5402 165 146 83 1 0.6731 0.875 0.3333 0.1647 80 85 70 14
6.7643 11.0526 210 3.0767 0.8965 0.5067 0.6848 0.5825 165 223 111 2 0.6698 0.8875 0.3590 0.4941 80 85 71 42
2.2893 12.6316 240 2.4180 0.9070 0.5642 0.7455 0.6423 165 218 120 3 0.6372 0.9 0.4857 0.6 80 85 72 51
5.0292 14.2105 270 17.3441 0.9111 0.5292 0.8242 0.6445 165 257 131 5 0.6261 0.9 0.4507 0.7529 80 85 72 64
5.1075 15.7895 300 17.8003 0.9127 0.4768 0.8727 0.6167 165 302 140 4 0.6050 0.9 0.3934 0.8471 80 85 72 72
1.7376 17.3684 330 18.0361 0.9128 0.4719 0.8667 0.6111 165 303 137 6 0.6207 0.9 0.3797 0.8353 80 85 72 71
1.7352 18.9474 360 32.0882 0.9217 0.4727 0.8909 0.6176 165 311 141 6 0.5581 0.9 0.4121 0.8824 80 85 72 75
3.8194 20.5263 390 30.3409 0.9216 0.5620 0.8788 0.6856 165 258 142 3 0.6827 0.8875 0.4805 0.8706 80 85 71 74
4.0106 22.1053 420 48.1283 0.9250 0.4654 0.8970 0.6128 165 318 140 8 0.6154 0.9 0.3781 0.8941 80 85 72 76
3.2054 23.6842 450 17.7213 0.9222 0.5121 0.8970 0.6520 165 289 143 5 0.6486 0.9 0.4270 0.8941 80 85 72 76
6.8048 25.2632 480 16.7449 0.9240 0.5068 0.9030 0.6492 165 294 142 7 0.6261 0.9 0.4302 0.9059 80 85 72 77
1.2021 26.8421 510 43.0137 0.9240 0.4760 0.9030 0.6234 165 313 142 7 0.6102 0.9 0.3949 0.9059 80 85 72 77
1.3674 28.4211 540 41.8534 0.9249 0.4715 0.9030 0.6195 165 316 142 7 0.6261 0.9 0.3831 0.9059 80 85 72 77
1.2001 30.0 570 42.0665 0.9260 0.4717 0.9091 0.6211 165 318 143 7 0.6372 0.9 0.3805 0.9176 80 85 72 78
1.1834 31.5789 600 32.0717 0.9214 0.4792 0.9091 0.6276 165 313 143 7 0.6102 0.9 0.4 0.9176 80 85 72 78
1.1969 33.1579 630 17.6968 0.9266 0.5017 0.9091 0.6466 165 299 144 6 0.6372 0.9 0.4194 0.9176 80 85 72 78

Framework versions

  • Transformers 4.46.3
  • Pytorch 2.4.1+cu121
  • Datasets 2.0.0
  • Tokenizers 0.20.3
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for levshechter/tibetan-CS-detector_mbert-tibetan-continual-wylie_all_data_no_labels_no_partial

Finetuned
(10)
this model