GSM8K-Binary_Llama-3.2-1B-gt3yb3cm

This model is a fine-tuned version of meta-llama/Llama-3.2-1B on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 1.4064
  • Model Preparation Time: 0.0055
  • Mdl: 5021.6739
  • Accumulated Loss: 3480.7591
  • Correct Preds: 1544.0
  • Total Preds: 2475.0
  • Accuracy: 0.6238
  • Correct Gen Preds: 261.0
  • Gen Accuracy: 0.1055
  • Correct Gen Preds 34192: 99.0
  • Correct Preds 34192: 991.0
  • Total Labels 34192: 1196.0
  • Accuracy 34192: 0.8286
  • Gen Accuracy 34192: 0.0828
  • Correct Gen Preds 41568: 155.0
  • Correct Preds 41568: 553.0
  • Total Labels 41568: 1267.0
  • Accuracy 41568: 0.4365
  • Gen Accuracy 41568: 0.1223

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 32
  • eval_batch_size: 64
  • seed: 42
  • optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_ratio: 0.01
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Model Preparation Time Mdl Accumulated Loss Correct Preds Total Preds Accuracy Correct Gen Preds Gen Accuracy Correct Gen Preds 34192 Correct Preds 34192 Total Labels 34192 Accuracy 34192 Gen Accuracy 34192 Correct Gen Preds 41568 Correct Preds 41568 Total Labels 41568 Accuracy 41568 Gen Accuracy 41568
No log 0 0 1.4656 0.0055 5233.1723 3627.3586 1196.0 2475.0 0.4832 1204.0 0.4865 1196.0 1196.0 1196.0 1.0 1.0 0.0 0.0 1267.0 0.0 0.0
2.5224 1.0 2 2.3937 0.0055 8547.2902 5924.5301 1266.0 2475.0 0.5115 1271.0 0.5135 0.0 0.0 1196.0 0.0 0.0 1263.0 1266.0 1267.0 0.9992 0.9968
1.7841 2.0 4 3.2020 0.0055 11433.4129 7925.0379 1196.0 2475.0 0.4832 8.0 0.0032 0.0 1196.0 1196.0 1.0 0.0 0.0 0.0 1267.0 0.0 0.0
0.8475 3.0 6 0.8616 0.0055 3076.3645 2132.3734 1266.0 2475.0 0.5115 7.0 0.0028 0.0 0.0 1196.0 0.0 0.0 0.0 1266.0 1267.0 0.9992 0.0
0.9462 4.0 8 0.8428 0.0055 3009.1929 2085.8135 1307.0 2475.0 0.5281 7.0 0.0028 0.0 520.0 1196.0 0.4348 0.0 0.0 787.0 1267.0 0.6212 0.0
0.9051 5.0 10 0.7595 0.0055 2711.8650 1879.7216 1330.0 2475.0 0.5374 7.0 0.0028 0.0 182.0 1196.0 0.1522 0.0 0.0 1148.0 1267.0 0.9061 0.0
1.812 6.0 12 0.8649 0.0055 3088.4505 2140.7508 1196.0 2475.0 0.4832 7.0 0.0028 0.0 1196.0 1196.0 1.0 0.0 0.0 0.0 1267.0 0.0 0.0
1.6264 7.0 14 0.7345 0.0055 2622.7585 1817.9576 1357.0 2475.0 0.5483 7.0 0.0028 0.0 858.0 1196.0 0.7174 0.0 0.0 499.0 1267.0 0.3938 0.0
0.5291 8.0 16 0.7757 0.0055 2769.7376 1919.8358 1403.0 2475.0 0.5669 7.0 0.0028 0.0 1022.0 1196.0 0.8545 0.0 0.0 381.0 1267.0 0.3007 0.0
0.237 9.0 18 0.8259 0.0055 2949.0298 2044.1117 1514.0 2475.0 0.6117 20.0 0.0081 0.0 973.0 1196.0 0.8135 0.0 12.0 541.0 1267.0 0.4270 0.0095
0.0421 10.0 20 1.3945 0.0055 4979.2689 3451.3662 1390.0 2475.0 0.5616 478.0 0.1931 1.0 367.0 1196.0 0.3069 0.0008 470.0 1023.0 1267.0 0.8074 0.3710
0.1491 11.0 22 1.4601 0.0055 5213.3876 3613.6449 1524.0 2475.0 0.6158 59.0 0.0238 13.0 1061.0 1196.0 0.8871 0.0109 39.0 463.0 1267.0 0.3654 0.0308
1.1113 12.0 24 1.4064 0.0055 5021.6739 3480.7591 1544.0 2475.0 0.6238 261.0 0.1055 99.0 991.0 1196.0 0.8286 0.0828 155.0 553.0 1267.0 0.4365 0.1223
0.0007 13.0 26 1.9100 0.0055 6819.9622 4727.2376 1436.0 2475.0 0.5802 858.0 0.3467 74.0 398.0 1196.0 0.3328 0.0619 776.0 1038.0 1267.0 0.8193 0.6125
0.0082 14.0 28 1.7320 0.0055 6184.5284 4286.7884 1502.0 2475.0 0.6069 960.0 0.3879 355.0 735.0 1196.0 0.6145 0.2968 597.0 767.0 1267.0 0.6054 0.4712
1.0707 15.0 30 2.0597 0.0055 7354.4002 5097.6817 1511.0 2475.0 0.6105 1248.0 0.5042 725.0 917.0 1196.0 0.7667 0.6062 516.0 594.0 1267.0 0.4688 0.4073
1.0696 16.0 32 2.1505 0.0055 7678.6998 5322.4691 1515.0 2475.0 0.6121 1359.0 0.5491 767.0 878.0 1196.0 0.7341 0.6413 584.0 637.0 1267.0 0.5028 0.4609
1.0693 17.0 34 2.1829 0.0055 7794.3632 5402.6409 1523.0 2475.0 0.6154 1395.0 0.5636 749.0 845.0 1196.0 0.7065 0.6263 638.0 678.0 1267.0 0.5351 0.5036
0.0 18.0 36 2.2042 0.0055 7870.3884 5455.3375 1521.0 2475.0 0.6145 1413.0 0.5709 732.0 810.0 1196.0 0.6773 0.6120 673.0 711.0 1267.0 0.5612 0.5312
0.0 19.0 38 2.2162 0.0055 7913.3941 5485.1468 1523.0 2475.0 0.6154 1425.0 0.5758 715.0 785.0 1196.0 0.6564 0.5978 703.0 738.0 1267.0 0.5825 0.5549
0.0 20.0 40 2.2249 0.0055 7944.3650 5506.6142 1513.0 2475.0 0.6113 1417.0 0.5725 697.0 763.0 1196.0 0.6380 0.5828 713.0 750.0 1267.0 0.5919 0.5627
0.0 21.0 42 2.2290 0.0055 7958.9258 5516.7070 1516.0 2475.0 0.6125 1426.0 0.5762 693.0 754.0 1196.0 0.6304 0.5794 726.0 762.0 1267.0 0.6014 0.5730
0.0 22.0 44 2.2335 0.0055 7975.2389 5528.0144 1517.0 2475.0 0.6129 1432.0 0.5786 691.0 750.0 1196.0 0.6271 0.5778 734.0 767.0 1267.0 0.6054 0.5793
0.0 23.0 46 2.2385 0.0055 7992.7922 5540.1814 1518.0 2475.0 0.6133 1435.0 0.5798 690.0 748.0 1196.0 0.6254 0.5769 738.0 770.0 1267.0 0.6077 0.5825
0.0 24.0 48 2.2436 0.0055 8011.1401 5552.8992 1515.0 2475.0 0.6121 1439.0 0.5814 694.0 748.0 1196.0 0.6254 0.5803 737.0 767.0 1267.0 0.6054 0.5817
0.0 25.0 50 2.2498 0.0055 8033.4137 5568.3380 1512.0 2475.0 0.6109 1431.0 0.5782 691.0 746.0 1196.0 0.6237 0.5778 733.0 766.0 1267.0 0.6046 0.5785
0.0 26.0 52 2.2514 0.0055 8039.0644 5572.2548 1520.0 2475.0 0.6141 1443.0 0.5830 699.0 753.0 1196.0 0.6296 0.5844 736.0 767.0 1267.0 0.6054 0.5809
0.0 27.0 54 2.2563 0.0055 8056.4092 5584.2773 1516.0 2475.0 0.6125 1438.0 0.5810 697.0 751.0 1196.0 0.6279 0.5828 734.0 765.0 1267.0 0.6038 0.5793
0.0 28.0 56 2.2602 0.0055 8070.5630 5594.0880 1517.0 2475.0 0.6129 1443.0 0.5830 701.0 753.0 1196.0 0.6296 0.5861 734.0 764.0 1267.0 0.6030 0.5793
0.0 29.0 58 2.2622 0.0055 8077.6162 5598.9769 1513.0 2475.0 0.6113 1433.0 0.5790 697.0 752.0 1196.0 0.6288 0.5828 729.0 761.0 1267.0 0.6006 0.5754
0.0 30.0 60 2.2649 0.0055 8087.1150 5605.5610 1518.0 2475.0 0.6133 1440.0 0.5818 704.0 757.0 1196.0 0.6329 0.5886 729.0 761.0 1267.0 0.6006 0.5754
0.0 31.0 62 2.2656 0.0055 8089.6775 5607.3372 1519.0 2475.0 0.6137 1441.0 0.5822 704.0 758.0 1196.0 0.6338 0.5886 730.0 761.0 1267.0 0.6006 0.5762
0.0 32.0 64 2.2689 0.0055 8101.4802 5615.5182 1514.0 2475.0 0.6117 1439.0 0.5814 703.0 756.0 1196.0 0.6321 0.5878 728.0 758.0 1267.0 0.5983 0.5746
0.0 33.0 66 2.2710 0.0055 8108.8518 5620.6278 1516.0 2475.0 0.6125 1436.0 0.5802 702.0 758.0 1196.0 0.6338 0.5870 727.0 758.0 1267.0 0.5983 0.5738
0.0 34.0 68 2.2741 0.0055 8119.9634 5628.3297 1518.0 2475.0 0.6133 1438.0 0.5810 704.0 760.0 1196.0 0.6355 0.5886 727.0 758.0 1267.0 0.5983 0.5738
1.0693 35.0 70 2.2735 0.0055 8117.9366 5626.9249 1517.0 2475.0 0.6129 1437.0 0.5806 705.0 760.0 1196.0 0.6355 0.5895 725.0 757.0 1267.0 0.5975 0.5722
0.0 36.0 72 2.2753 0.0055 8124.4509 5631.4402 1516.0 2475.0 0.6125 1436.0 0.5802 704.0 762.0 1196.0 0.6371 0.5886 725.0 754.0 1267.0 0.5951 0.5722
0.0 37.0 74 2.2753 0.0055 8124.3745 5631.3873 1514.0 2475.0 0.6117 1436.0 0.5802 707.0 762.0 1196.0 0.6371 0.5911 721.0 752.0 1267.0 0.5935 0.5691
0.0 38.0 76 2.2788 0.0055 8136.6842 5639.9197 1521.0 2475.0 0.6145 1441.0 0.5822 710.0 766.0 1196.0 0.6405 0.5936 724.0 755.0 1267.0 0.5959 0.5714
0.0 39.0 78 2.2790 0.0055 8137.4584 5640.4564 1516.0 2475.0 0.6125 1438.0 0.5810 710.0 764.0 1196.0 0.6388 0.5936 720.0 752.0 1267.0 0.5935 0.5683
0.0 40.0 80 2.2802 0.0055 8141.8020 5643.4671 1517.0 2475.0 0.6129 1438.0 0.5810 710.0 766.0 1196.0 0.6405 0.5936 721.0 751.0 1267.0 0.5927 0.5691
0.0 41.0 82 2.2808 0.0055 8144.0353 5645.0151 1519.0 2475.0 0.6137 1439.0 0.5814 711.0 767.0 1196.0 0.6413 0.5945 720.0 752.0 1267.0 0.5935 0.5683
0.0 42.0 84 2.2831 0.0055 8152.0846 5650.5945 1516.0 2475.0 0.6125 1439.0 0.5814 710.0 765.0 1196.0 0.6396 0.5936 721.0 751.0 1267.0 0.5927 0.5691

Framework versions

  • Transformers 4.51.3
  • Pytorch 2.6.0+cu124
  • Datasets 3.5.0
  • Tokenizers 0.21.1
Downloads last month
2
Safetensors
Model size
1B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for donoway/GSM8K-Binary_Llama-3.2-1B-gt3yb3cm

Finetuned
(903)
this model