mbert-base-finetuned-ner-chim-v1
This model is a fine-tuned version of google-bert/bert-base-multilingual-cased on an unknown dataset. It achieves the following results on the evaluation set:
- Loss: 0.4824
- Accuracy: 0.8639
- Precision: 0.5430
- Recall: 0.6330
- F1: 0.5818
- Date Precision: 0.4146
- Date Recall: 0.425
- Date F1-score: 0.4198
- Habitat Precision: 0.552
- Habitat Recall: 0.7263
- Habitat F1-score: 0.6273
- Id Feature Precision: 0.1871
- Id Feature Recall: 0.3295
- Id Feature F1-score: 0.2387
- Location Precision: 0.7885
- Location Recall: 0.8662
- Location F1-score: 0.8255
- Organization Precision: 0.6027
- Organization Recall: 0.6769
- Organization F1-score: 0.6377
- Species Precision: 0.7129
- Species Recall: 0.7740
- Species F1-score: 0.7422
- Micro avg Precision: 0.5859
- Micro avg Recall: 0.7036
- Micro avg F1-score: 0.6394
- Macro avg Precision: 0.5430
- Macro avg Recall: 0.6330
- Macro avg F1-score: 0.5818
- Weighted avg Precision: 0.6161
- Weighted avg Recall: 0.7036
- Weighted avg F1-score: 0.6548
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- num_epochs: 10
Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 | Date Precision | Date Recall | Date F1-score | Habitat Precision | Habitat Recall | Habitat F1-score | Id Feature Precision | Id Feature Recall | Id Feature F1-score | Location Precision | Location Recall | Location F1-score | Organization Precision | Organization Recall | Organization F1-score | Species Precision | Species Recall | Species F1-score | Micro avg Precision | Micro avg Recall | Micro avg F1-score | Macro avg Precision | Macro avg Recall | Macro avg F1-score | Weighted avg Precision | Weighted avg Recall | Weighted avg F1-score |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| No log | 1.0 | 19 | 1.0313 | 0.7163 | 0.2262 | 0.1318 | 0.1526 | 0.0 | 0.0 | 0.0 | 0.4375 | 0.0737 | 0.1261 | 0.0862 | 0.0568 | 0.0685 | 0.4786 | 0.4718 | 0.4752 | 0.0 | 0.0 | 0.0 | 0.3548 | 0.1884 | 0.2461 | 0.3631 | 0.1856 | 0.2456 | 0.2262 | 0.1318 | 0.1526 | 0.3057 | 0.1856 | 0.2179 |
| No log | 2.0 | 38 | 0.6467 | 0.8117 | 0.4076 | 0.4255 | 0.4093 | 0.3913 | 0.225 | 0.2857 | 0.5876 | 0.6 | 0.5938 | 0.1761 | 0.2841 | 0.2174 | 0.5805 | 0.7113 | 0.6392 | 0.08 | 0.0615 | 0.0696 | 0.6302 | 0.6712 | 0.6501 | 0.4918 | 0.5429 | 0.5161 | 0.4076 | 0.4255 | 0.4093 | 0.4967 | 0.5429 | 0.5154 |
| No log | 3.0 | 57 | 0.5422 | 0.8383 | 0.4642 | 0.5099 | 0.4809 | 0.4333 | 0.325 | 0.3714 | 0.5810 | 0.6421 | 0.6100 | 0.2 | 0.3182 | 0.2456 | 0.6782 | 0.8310 | 0.7468 | 0.2830 | 0.2308 | 0.2542 | 0.6100 | 0.7123 | 0.6572 | 0.5255 | 0.6136 | 0.5661 | 0.4642 | 0.5099 | 0.4809 | 0.5304 | 0.6136 | 0.5663 |
| No log | 4.0 | 76 | 0.4842 | 0.8612 | 0.4950 | 0.5324 | 0.5100 | 0.3611 | 0.325 | 0.3421 | 0.6020 | 0.6211 | 0.6114 | 0.2256 | 0.3409 | 0.2715 | 0.6923 | 0.8239 | 0.7524 | 0.3818 | 0.3231 | 0.3500 | 0.7070 | 0.7603 | 0.7327 | 0.5739 | 0.6399 | 0.6051 | 0.4950 | 0.5324 | 0.5100 | 0.5832 | 0.6399 | 0.6083 |
| No log | 5.0 | 95 | 0.4704 | 0.8610 | 0.5148 | 0.6097 | 0.5560 | 0.5 | 0.5 | 0.5 | 0.5546 | 0.6947 | 0.6168 | 0.1892 | 0.3182 | 0.2373 | 0.7241 | 0.8873 | 0.7975 | 0.4559 | 0.4769 | 0.4662 | 0.6647 | 0.7808 | 0.7181 | 0.5594 | 0.6911 | 0.6183 | 0.5148 | 0.6097 | 0.5560 | 0.5760 | 0.6911 | 0.6270 |
| No log | 6.0 | 114 | 0.4701 | 0.8678 | 0.5421 | 0.6292 | 0.5801 | 0.425 | 0.425 | 0.425 | 0.5702 | 0.7263 | 0.6389 | 0.2340 | 0.375 | 0.2882 | 0.7515 | 0.8732 | 0.8078 | 0.5758 | 0.5846 | 0.5802 | 0.6958 | 0.7911 | 0.7404 | 0.5919 | 0.7091 | 0.6452 | 0.5421 | 0.6292 | 0.5801 | 0.6081 | 0.7091 | 0.6533 |
| No log | 7.0 | 133 | 0.4692 | 0.8640 | 0.5482 | 0.6385 | 0.5869 | 0.4762 | 0.5 | 0.4878 | 0.5469 | 0.7368 | 0.6278 | 0.2171 | 0.375 | 0.2750 | 0.7961 | 0.8521 | 0.8231 | 0.5417 | 0.6 | 0.5693 | 0.7111 | 0.7671 | 0.7381 | 0.5889 | 0.7022 | 0.6406 | 0.5482 | 0.6385 | 0.5869 | 0.6177 | 0.7022 | 0.6548 |
| No log | 8.0 | 152 | 0.4742 | 0.8650 | 0.5439 | 0.6309 | 0.5822 | 0.475 | 0.475 | 0.4750 | 0.55 | 0.6947 | 0.6140 | 0.2014 | 0.3295 | 0.25 | 0.7640 | 0.8662 | 0.8119 | 0.5753 | 0.6462 | 0.6087 | 0.6975 | 0.7740 | 0.7338 | 0.5858 | 0.6994 | 0.6376 | 0.5439 | 0.6309 | 0.5822 | 0.6074 | 0.6994 | 0.6488 |
| No log | 9.0 | 171 | 0.4847 | 0.8629 | 0.5419 | 0.6246 | 0.5780 | 0.4390 | 0.45 | 0.4444 | 0.5484 | 0.7158 | 0.6210 | 0.1722 | 0.2955 | 0.2176 | 0.7785 | 0.8662 | 0.82 | 0.5915 | 0.6462 | 0.6176 | 0.7220 | 0.7740 | 0.7471 | 0.5862 | 0.6967 | 0.6367 | 0.5419 | 0.6246 | 0.5780 | 0.6158 | 0.6967 | 0.6519 |
| No log | 10.0 | 190 | 0.4824 | 0.8639 | 0.5430 | 0.6330 | 0.5818 | 0.4146 | 0.425 | 0.4198 | 0.552 | 0.7263 | 0.6273 | 0.1871 | 0.3295 | 0.2387 | 0.7885 | 0.8662 | 0.8255 | 0.6027 | 0.6769 | 0.6377 | 0.7129 | 0.7740 | 0.7422 | 0.5859 | 0.7036 | 0.6394 | 0.5430 | 0.6330 | 0.5818 | 0.6161 | 0.7036 | 0.6548 |
Framework versions
- Transformers 4.51.3
- Pytorch 2.6.0+cu124
- Datasets 3.6.0
- Tokenizers 0.21.1
- Downloads last month
- 1