GSM8K-Binary_Llama-3.2-1B-28w952gf

This model is a fine-tuned version of meta-llama/Llama-3.2-1B on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 1.5835
  • Model Preparation Time: 0.0058
  • Mdl: 5653.9903
  • Accumulated Loss: 3919.0474
  • Correct Preds: 2011.0
  • Total Preds: 2475.0
  • Accuracy: 0.8125
  • Correct Gen Preds: 2009.0
  • Gen Accuracy: 0.8117
  • Correct Gen Preds 34192: 1030.0
  • Correct Preds 34192: 1035.0
  • Total Labels 34192: 1196.0
  • Accuracy 34192: 0.8654
  • Gen Accuracy 34192: 0.8612
  • Correct Gen Preds 41568: 970.0
  • Correct Preds 41568: 976.0
  • Total Labels 41568: 1267.0
  • Accuracy 41568: 0.7703
  • Gen Accuracy 41568: 0.7656

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 32
  • eval_batch_size: 64
  • seed: 42
  • optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_ratio: 0.01
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Model Preparation Time Mdl Accumulated Loss Correct Preds Total Preds Accuracy Correct Gen Preds Gen Accuracy Correct Gen Preds 34192 Correct Preds 34192 Total Labels 34192 Accuracy 34192 Gen Accuracy 34192 Correct Gen Preds 41568 Correct Preds 41568 Total Labels 41568 Accuracy 41568 Gen Accuracy 41568
No log 0 0 1.4656 0.0058 5233.1723 3627.3586 1196.0 2475.0 0.4832 1204.0 0.4865 1196.0 1196.0 1196.0 1.0 1.0 0.0 0.0 1267.0 0.0 0.0
1.0365 1.0 41 0.7442 0.0058 2657.1555 1841.7998 1698.0 2475.0 0.6861 1609.0 0.6501 552.0 633.0 1196.0 0.5293 0.4615 1049.0 1065.0 1267.0 0.8406 0.8279
0.8382 2.0 82 0.5588 0.0058 1995.1639 1382.9423 1815.0 2475.0 0.7333 354.0 0.1430 15.0 1146.0 1196.0 0.9582 0.0125 331.0 669.0 1267.0 0.5280 0.2612
0.8522 3.0 123 0.6851 0.0058 2446.3901 1695.7084 1847.0 2475.0 0.7463 349.0 0.1410 111.0 1149.0 1196.0 0.9607 0.0928 230.0 698.0 1267.0 0.5509 0.1815
0.4153 4.0 164 0.5204 0.0058 1858.1740 1287.9881 1977.0 2475.0 0.7988 1351.0 0.5459 499.0 944.0 1196.0 0.7893 0.4172 844.0 1033.0 1267.0 0.8153 0.6661
0.7977 5.0 205 0.7190 0.0058 2567.2475 1779.4804 1980.0 2475.0 0.8 1245.0 0.5030 561.0 1080.0 1196.0 0.9030 0.4691 675.0 900.0 1267.0 0.7103 0.5328
0.001 6.0 246 1.2247 0.0058 4372.8577 3031.0340 1947.0 2475.0 0.7867 1648.0 0.6659 936.0 1104.0 1196.0 0.9231 0.7826 703.0 843.0 1267.0 0.6654 0.5549
0.0026 7.0 287 1.0541 0.0058 3763.8213 2608.8821 1966.0 2475.0 0.7943 1568.0 0.6335 873.0 1066.0 1196.0 0.8913 0.7299 686.0 900.0 1267.0 0.7103 0.5414
0.0366 8.0 328 1.2770 0.0058 4559.6621 3160.5169 1948.0 2475.0 0.7871 1751.0 0.7075 847.0 950.0 1196.0 0.7943 0.7082 895.0 998.0 1267.0 0.7877 0.7064
0.0 9.0 369 1.6923 0.0058 6042.5959 4188.4083 1954.0 2475.0 0.7895 1837.0 0.7422 1026.0 1117.0 1196.0 0.9339 0.8579 801.0 837.0 1267.0 0.6606 0.6322
0.0002 10.0 410 1.8498 0.0058 6605.0876 4578.2979 1946.0 2475.0 0.7863 1926.0 0.7782 1104.0 1118.0 1196.0 0.9348 0.9231 813.0 828.0 1267.0 0.6535 0.6417
0.7352 11.0 451 1.9100 0.0058 6819.8267 4727.1437 1969.0 2475.0 0.7956 1964.0 0.7935 1099.0 1106.0 1196.0 0.9247 0.9189 855.0 863.0 1267.0 0.6811 0.6748
0.0 12.0 492 1.5599 0.0058 5570.0379 3860.8561 2009.0 2475.0 0.8117 2008.0 0.8113 1022.0 1026.0 1196.0 0.8579 0.8545 977.0 983.0 1267.0 0.7758 0.7711
0.0 13.0 533 1.5753 0.0058 5624.9198 3898.8973 2008.0 2475.0 0.8113 2005.0 0.8101 1027.0 1033.0 1196.0 0.8637 0.8587 969.0 975.0 1267.0 0.7695 0.7648
0.0 14.0 574 1.5762 0.0058 5628.2132 3901.1801 2008.0 2475.0 0.8113 2006.0 0.8105 1029.0 1034.0 1196.0 0.8645 0.8604 968.0 974.0 1267.0 0.7687 0.7640
1.4702 15.0 615 1.5799 0.0058 5641.1668 3910.1588 2009.0 2475.0 0.8117 2007.0 0.8109 1027.0 1032.0 1196.0 0.8629 0.8587 971.0 977.0 1267.0 0.7711 0.7664
0.0 16.0 656 1.5812 0.0058 5645.8715 3913.4199 2009.0 2475.0 0.8117 2007.0 0.8109 1029.0 1035.0 1196.0 0.8654 0.8604 969.0 974.0 1267.0 0.7687 0.7648
0.0 17.0 697 1.5816 0.0058 5647.2796 3914.3960 2009.0 2475.0 0.8117 2008.0 0.8113 1029.0 1034.0 1196.0 0.8645 0.8604 970.0 975.0 1267.0 0.7695 0.7656
0.0 18.0 738 1.5835 0.0058 5653.9903 3919.0474 2011.0 2475.0 0.8125 2009.0 0.8117 1030.0 1035.0 1196.0 0.8654 0.8612 970.0 976.0 1267.0 0.7703 0.7656
0.0 19.0 779 1.5847 0.0058 5658.5138 3922.1829 2006.0 2475.0 0.8105 2003.0 0.8093 1027.0 1033.0 1196.0 0.8637 0.8587 967.0 973.0 1267.0 0.7680 0.7632
0.0 20.0 820 1.5855 0.0058 5661.4129 3924.1924 2009.0 2475.0 0.8117 2007.0 0.8109 1029.0 1035.0 1196.0 0.8654 0.8604 969.0 974.0 1267.0 0.7687 0.7648
0.0 21.0 861 1.5872 0.0058 5667.5375 3928.4376 2008.0 2475.0 0.8113 2006.0 0.8105 1028.0 1034.0 1196.0 0.8645 0.8595 969.0 974.0 1267.0 0.7687 0.7648
0.0 22.0 902 1.5892 0.0058 5674.6021 3933.3344 2009.0 2475.0 0.8117 2008.0 0.8113 1030.0 1035.0 1196.0 0.8654 0.8612 969.0 974.0 1267.0 0.7687 0.7648
0.0 23.0 943 1.5898 0.0058 5676.7926 3934.8528 2010.0 2475.0 0.8121 2008.0 0.8113 1030.0 1036.0 1196.0 0.8662 0.8612 969.0 974.0 1267.0 0.7687 0.7648
0.0 24.0 984 1.5899 0.0058 5676.8491 3934.8920 2011.0 2475.0 0.8125 2010.0 0.8121 1031.0 1037.0 1196.0 0.8671 0.8620 970.0 974.0 1267.0 0.7687 0.7656
0.7351 25.0 1025 1.5905 0.0058 5679.0066 3936.3874 2006.0 2475.0 0.8105 2006.0 0.8105 1029.0 1034.0 1196.0 0.8645 0.8604 968.0 972.0 1267.0 0.7672 0.7640
0.0 26.0 1066 1.5921 0.0058 5684.9764 3940.5253 2008.0 2475.0 0.8113 2006.0 0.8105 1029.0 1035.0 1196.0 0.8654 0.8604 968.0 973.0 1267.0 0.7680 0.7640
0.0 27.0 1107 1.5941 0.0058 5691.8627 3945.2986 2008.0 2475.0 0.8113 2007.0 0.8109 1029.0 1034.0 1196.0 0.8645 0.8604 969.0 974.0 1267.0 0.7687 0.7648
0.0 28.0 1148 1.5959 0.0058 5698.4968 3949.8970 2011.0 2475.0 0.8125 2009.0 0.8117 1031.0 1037.0 1196.0 0.8671 0.8620 969.0 974.0 1267.0 0.7687 0.7648
0.0 29.0 1189 1.5953 0.0058 5696.1814 3948.2921 2010.0 2475.0 0.8121 2007.0 0.8109 1029.0 1035.0 1196.0 0.8654 0.8604 969.0 975.0 1267.0 0.7695 0.7648
0.0 30.0 1230 1.5954 0.0058 5696.5394 3948.5402 2009.0 2475.0 0.8117 2006.0 0.8105 1029.0 1035.0 1196.0 0.8654 0.8604 968.0 974.0 1267.0 0.7687 0.7640

Framework versions

  • Transformers 4.51.3
  • Pytorch 2.6.0+cu124
  • Datasets 3.5.0
  • Tokenizers 0.21.1
Downloads last month
2
Safetensors
Model size
1B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for donoway/GSM8K-Binary_Llama-3.2-1B-28w952gf

Finetuned
(903)
this model