roberta-large-ner-ghtk-cs-add-3label-5-new-data-3090-14Sep-1

This model was trained from scratch on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.2554
  • Tk: {'precision': 0.8137254901960784, 'recall': 0.7155172413793104, 'f1': 0.761467889908257, 'number': 116}
  • A: {'precision': 0.9300225733634312, 'recall': 0.9559164733178654, 'f1': 0.94279176201373, 'number': 431}
  • Gày: {'precision': 0.7142857142857143, 'recall': 0.8823529411764706, 'f1': 0.7894736842105262, 'number': 34}
  • Gày trừu tượng: {'precision': 0.9030927835051547, 'recall': 0.8975409836065574, 'f1': 0.9003083247687564, 'number': 488}
  • Iền: {'precision': 0.74, 'recall': 0.9487179487179487, 'f1': 0.8314606741573033, 'number': 39}
  • Iờ: {'precision': 0.5576923076923077, 'recall': 0.7631578947368421, 'f1': 0.6444444444444444, 'number': 38}
  • Ã đơn: {'precision': 0.873015873015873, 'recall': 0.812807881773399, 'f1': 0.8418367346938775, 'number': 203}
  • Đt: {'precision': 0.9357601713062098, 'recall': 0.9954441913439636, 'f1': 0.9646799116997792, 'number': 878}
  • Đt trừu tượng: {'precision': 0.8291666666666667, 'recall': 0.8540772532188842, 'f1': 0.8414376321353066, 'number': 233}
  • Overall Precision: 0.8936
  • Overall Recall: 0.9215
  • Overall F1: 0.9073
  • Overall Accuracy: 0.9592

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2.5e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Tk A Gày Gày trừu tượng Iền Iờ Ã đơn Đt Đt trừu tượng Overall Precision Overall Recall Overall F1 Overall Accuracy
No log 1.0 272 0.1712 {'precision': 0.7191780821917808, 'recall': 0.9051724137931034, 'f1': 0.8015267175572519, 'number': 116} {'precision': 0.8742268041237113, 'recall': 0.9837587006960556, 'f1': 0.9257641921397378, 'number': 431} {'precision': 0.717948717948718, 'recall': 0.8235294117647058, 'f1': 0.767123287671233, 'number': 34} {'precision': 0.8707070707070707, 'recall': 0.8831967213114754, 'f1': 0.8769074262461851, 'number': 488} {'precision': 0.7333333333333333, 'recall': 0.8461538461538461, 'f1': 0.7857142857142856, 'number': 39} {'precision': 0.6571428571428571, 'recall': 0.6052631578947368, 'f1': 0.6301369863013698, 'number': 38} {'precision': 0.7979274611398963, 'recall': 0.7586206896551724, 'f1': 0.7777777777777777, 'number': 203} {'precision': 0.9670975323149236, 'recall': 0.9373576309794989, 'f1': 0.9519953730480046, 'number': 878} {'precision': 0.7473684210526316, 'recall': 0.9141630901287554, 'f1': 0.8223938223938223, 'number': 233} 0.8679 0.9081 0.8876 0.9507
0.1009 2.0 544 0.2061 {'precision': 0.8085106382978723, 'recall': 0.3275862068965517, 'f1': 0.46625766871165636, 'number': 116} {'precision': 0.9246575342465754, 'recall': 0.9396751740139211, 'f1': 0.9321058688147296, 'number': 431} {'precision': 0.673469387755102, 'recall': 0.9705882352941176, 'f1': 0.7951807228915663, 'number': 34} {'precision': 0.8606870229007634, 'recall': 0.9241803278688525, 'f1': 0.891304347826087, 'number': 488} {'precision': 0.6666666666666666, 'recall': 0.9230769230769231, 'f1': 0.7741935483870968, 'number': 39} {'precision': 0.5737704918032787, 'recall': 0.9210526315789473, 'f1': 0.7070707070707071, 'number': 38} {'precision': 0.7119341563786008, 'recall': 0.8522167487684729, 'f1': 0.7757847533632286, 'number': 203} {'precision': 0.9004106776180698, 'recall': 0.9988610478359908, 'f1': 0.9470842332613392, 'number': 878} {'precision': 0.8418803418803419, 'recall': 0.8454935622317596, 'f1': 0.8436830835117772, 'number': 233} 0.8556 0.9126 0.8832 0.9490
0.1009 3.0 816 0.1502 {'precision': 0.7703703703703704, 'recall': 0.896551724137931, 'f1': 0.8286852589641435, 'number': 116} {'precision': 0.9377880184331797, 'recall': 0.9443155452436195, 'f1': 0.9410404624277456, 'number': 431} {'precision': 0.7857142857142857, 'recall': 0.6470588235294118, 'f1': 0.7096774193548386, 'number': 34} {'precision': 0.8856015779092702, 'recall': 0.9200819672131147, 'f1': 0.9025125628140703, 'number': 488} {'precision': 0.7551020408163265, 'recall': 0.9487179487179487, 'f1': 0.8409090909090908, 'number': 39} {'precision': 0.5714285714285714, 'recall': 0.9473684210526315, 'f1': 0.7128712871287128, 'number': 38} {'precision': 0.8532608695652174, 'recall': 0.7733990147783252, 'f1': 0.8113695090439278, 'number': 203} {'precision': 0.960308710033076, 'recall': 0.9920273348519362, 'f1': 0.9759103641456582, 'number': 878} {'precision': 0.8810572687224669, 'recall': 0.8583690987124464, 'f1': 0.8695652173913043, 'number': 233} 0.9009 0.9280 0.9143 0.9606
0.0434 4.0 1088 0.1833 {'precision': 0.808695652173913, 'recall': 0.8017241379310345, 'f1': 0.8051948051948051, 'number': 116} {'precision': 0.9311926605504587, 'recall': 0.9419953596287703, 'f1': 0.936562860438293, 'number': 431} {'precision': 0.7142857142857143, 'recall': 0.8823529411764706, 'f1': 0.7894736842105262, 'number': 34} {'precision': 0.89738430583501, 'recall': 0.9139344262295082, 'f1': 0.9055837563451776, 'number': 488} {'precision': 0.74, 'recall': 0.9487179487179487, 'f1': 0.8314606741573033, 'number': 39} {'precision': 0.6111111111111112, 'recall': 0.5789473684210527, 'f1': 0.5945945945945946, 'number': 38} {'precision': 0.9075144508670521, 'recall': 0.7733990147783252, 'f1': 0.8351063829787234, 'number': 203} {'precision': 0.9663299663299664, 'recall': 0.9806378132118451, 'f1': 0.9734313171283211, 'number': 878} {'precision': 0.8448275862068966, 'recall': 0.8412017167381974, 'f1': 0.8430107526881722, 'number': 233} 0.9094 0.9138 0.9116 0.9614
0.0434 5.0 1360 0.2314 {'precision': 0.896551724137931, 'recall': 0.6724137931034483, 'f1': 0.768472906403941, 'number': 116} {'precision': 0.9359267734553776, 'recall': 0.9489559164733179, 'f1': 0.9423963133640553, 'number': 431} {'precision': 0.7073170731707317, 'recall': 0.8529411764705882, 'f1': 0.7733333333333334, 'number': 34} {'precision': 0.90020366598778, 'recall': 0.9057377049180327, 'f1': 0.9029622063329927, 'number': 488} {'precision': 0.74, 'recall': 0.9487179487179487, 'f1': 0.8314606741573033, 'number': 39} {'precision': 0.8, 'recall': 0.5263157894736842, 'f1': 0.6349206349206348, 'number': 38} {'precision': 0.8256410256410256, 'recall': 0.7931034482758621, 'f1': 0.8090452261306532, 'number': 203} {'precision': 0.916754478398314, 'recall': 0.9908883826879271, 'f1': 0.9523809523809524, 'number': 878} {'precision': 0.7811320754716982, 'recall': 0.8884120171673819, 'f1': 0.8313253012048193, 'number': 233} 0.8870 0.9159 0.9012 0.9570
0.0257 6.0 1632 0.2328 {'precision': 0.79, 'recall': 0.6810344827586207, 'f1': 0.7314814814814816, 'number': 116} {'precision': 0.9422632794457275, 'recall': 0.9466357308584686, 'f1': 0.9444444444444444, 'number': 431} {'precision': 0.725, 'recall': 0.8529411764705882, 'f1': 0.7837837837837837, 'number': 34} {'precision': 0.927038626609442, 'recall': 0.8852459016393442, 'f1': 0.9056603773584905, 'number': 488} {'precision': 0.7708333333333334, 'recall': 0.9487179487179487, 'f1': 0.8505747126436781, 'number': 39} {'precision': 0.6666666666666666, 'recall': 0.6842105263157895, 'f1': 0.6753246753246753, 'number': 38} {'precision': 0.8850574712643678, 'recall': 0.7586206896551724, 'f1': 0.8169761273209549, 'number': 203} {'precision': 0.9526431718061674, 'recall': 0.9851936218678815, 'f1': 0.9686450167973124, 'number': 878} {'precision': 0.8857142857142857, 'recall': 0.7982832618025751, 'f1': 0.8397291196388261, 'number': 233} 0.9165 0.9008 0.9086 0.9595
0.0257 7.0 1904 0.2436 {'precision': 0.8282828282828283, 'recall': 0.7068965517241379, 'f1': 0.7627906976744185, 'number': 116} {'precision': 0.9284116331096197, 'recall': 0.962877030162413, 'f1': 0.9453302961275627, 'number': 431} {'precision': 0.7111111111111111, 'recall': 0.9411764705882353, 'f1': 0.8101265822784811, 'number': 34} {'precision': 0.8835341365461847, 'recall': 0.9016393442622951, 'f1': 0.8924949290060852, 'number': 488} {'precision': 0.7254901960784313, 'recall': 0.9487179487179487, 'f1': 0.8222222222222223, 'number': 39} {'precision': 0.5614035087719298, 'recall': 0.8421052631578947, 'f1': 0.6736842105263158, 'number': 38} {'precision': 0.8208955223880597, 'recall': 0.812807881773399, 'f1': 0.8168316831683169, 'number': 203} {'precision': 0.9287991498405951, 'recall': 0.9954441913439636, 'f1': 0.9609675645959318, 'number': 878} {'precision': 0.7984189723320159, 'recall': 0.8669527896995708, 'f1': 0.831275720164609, 'number': 233} 0.8792 0.9264 0.9022 0.9559
0.0164 8.0 2176 0.2475 {'precision': 0.8041237113402062, 'recall': 0.6724137931034483, 'f1': 0.7323943661971831, 'number': 116} {'precision': 0.917960088691796, 'recall': 0.9605568445475638, 'f1': 0.9387755102040817, 'number': 431} {'precision': 0.7317073170731707, 'recall': 0.8823529411764706, 'f1': 0.8, 'number': 34} {'precision': 0.8995901639344263, 'recall': 0.8995901639344263, 'f1': 0.8995901639344263, 'number': 488} {'precision': 0.7708333333333334, 'recall': 0.9487179487179487, 'f1': 0.8505747126436781, 'number': 39} {'precision': 0.5660377358490566, 'recall': 0.7894736842105263, 'f1': 0.6593406593406593, 'number': 38} {'precision': 0.8358974358974359, 'recall': 0.8029556650246306, 'f1': 0.8190954773869348, 'number': 203} {'precision': 0.9297872340425531, 'recall': 0.9954441913439636, 'f1': 0.9614961496149614, 'number': 878} {'precision': 0.8291666666666667, 'recall': 0.8540772532188842, 'f1': 0.8414376321353066, 'number': 233} 0.8868 0.9203 0.9033 0.9577
0.0164 9.0 2448 0.2498 {'precision': 0.8058252427184466, 'recall': 0.7155172413793104, 'f1': 0.7579908675799087, 'number': 116} {'precision': 0.9280898876404494, 'recall': 0.9582366589327146, 'f1': 0.9429223744292237, 'number': 431} {'precision': 0.7142857142857143, 'recall': 0.8823529411764706, 'f1': 0.7894736842105262, 'number': 34} {'precision': 0.9066390041493776, 'recall': 0.8954918032786885, 'f1': 0.9010309278350516, 'number': 488} {'precision': 0.74, 'recall': 0.9487179487179487, 'f1': 0.8314606741573033, 'number': 39} {'precision': 0.5636363636363636, 'recall': 0.8157894736842105, 'f1': 0.6666666666666666, 'number': 38} {'precision': 0.8549222797927462, 'recall': 0.812807881773399, 'f1': 0.8333333333333334, 'number': 203} {'precision': 0.9377682403433476, 'recall': 0.9954441913439636, 'f1': 0.9657458563535912, 'number': 878} {'precision': 0.8468085106382979, 'recall': 0.8540772532188842, 'f1': 0.8504273504273505, 'number': 233} 0.8944 0.9224 0.9081 0.9592
0.0088 10.0 2720 0.2554 {'precision': 0.8137254901960784, 'recall': 0.7155172413793104, 'f1': 0.761467889908257, 'number': 116} {'precision': 0.9300225733634312, 'recall': 0.9559164733178654, 'f1': 0.94279176201373, 'number': 431} {'precision': 0.7142857142857143, 'recall': 0.8823529411764706, 'f1': 0.7894736842105262, 'number': 34} {'precision': 0.9030927835051547, 'recall': 0.8975409836065574, 'f1': 0.9003083247687564, 'number': 488} {'precision': 0.74, 'recall': 0.9487179487179487, 'f1': 0.8314606741573033, 'number': 39} {'precision': 0.5576923076923077, 'recall': 0.7631578947368421, 'f1': 0.6444444444444444, 'number': 38} {'precision': 0.873015873015873, 'recall': 0.812807881773399, 'f1': 0.8418367346938775, 'number': 203} {'precision': 0.9357601713062098, 'recall': 0.9954441913439636, 'f1': 0.9646799116997792, 'number': 878} {'precision': 0.8291666666666667, 'recall': 0.8540772532188842, 'f1': 0.8414376321353066, 'number': 233} 0.8936 0.9215 0.9073 0.9592

Framework versions

  • Transformers 4.44.0
  • Pytorch 2.3.1+cu121
  • Datasets 2.19.1
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.6B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support