GSM8K-Binary_Llama-3.2-1B-fal8is8l

This model is a fine-tuned version of meta-llama/Llama-3.2-1B on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9292
  • Model Preparation Time: 0.0056
  • Mdl: 3317.6903
  • Accumulated Loss: 2299.6477
  • Correct Preds: 1831.0
  • Total Preds: 2475.0
  • Accuracy: 0.7398
  • Correct Gen Preds: 1492.0
  • Gen Accuracy: 0.6028
  • Correct Gen Preds 34192: 688.0
  • Correct Preds 34192: 931.0
  • Total Labels 34192: 1196.0
  • Accuracy 34192: 0.7784
  • Gen Accuracy 34192: 0.5753
  • Correct Gen Preds 41568: 796.0
  • Correct Preds 41568: 900.0
  • Total Labels 41568: 1267.0
  • Accuracy 41568: 0.7103
  • Gen Accuracy 41568: 0.6283

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 32
  • eval_batch_size: 64
  • seed: 42
  • optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_ratio: 0.01
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Model Preparation Time Mdl Accumulated Loss Correct Preds Total Preds Accuracy Correct Gen Preds Gen Accuracy Correct Gen Preds 34192 Correct Preds 34192 Total Labels 34192 Accuracy 34192 Gen Accuracy 34192 Correct Gen Preds 41568 Correct Preds 41568 Total Labels 41568 Accuracy 41568 Gen Accuracy 41568
No log 0 0 1.4656 0.0056 5233.1723 3627.3586 1196.0 2475.0 0.4832 1204.0 0.4865 1196.0 1196.0 1196.0 1.0 1.0 0.0 0.0 1267.0 0.0 0.0
1.3358 1.0 6 0.7779 0.0056 2777.7209 1925.3694 1292.0 2475.0 0.5220 8.0 0.0032 0.0 1091.0 1196.0 0.9122 0.0 0.0 201.0 1267.0 0.1586 0.0
0.7578 2.0 12 0.6802 0.0056 2428.6444 1683.4080 1649.0 2475.0 0.6663 7.0 0.0028 0.0 1004.0 1196.0 0.8395 0.0 0.0 645.0 1267.0 0.5091 0.0
1.201 3.0 18 0.7200 0.0056 2570.8615 1781.9854 1488.0 2475.0 0.6012 7.0 0.0028 0.0 1168.0 1196.0 0.9766 0.0 0.0 320.0 1267.0 0.2526 0.0
0.8365 4.0 24 0.6922 0.0056 2471.7795 1713.3070 1648.0 2475.0 0.6659 71.0 0.0287 0.0 1129.0 1196.0 0.9440 0.0 63.0 519.0 1267.0 0.4096 0.0497
0.4388 5.0 30 0.7937 0.0056 2834.1220 1964.4637 1674.0 2475.0 0.6764 117.0 0.0473 1.0 1118.0 1196.0 0.9348 0.0008 108.0 556.0 1267.0 0.4388 0.0852
0.1168 6.0 36 0.6699 0.0056 2392.0669 1658.0544 1758.0 2475.0 0.7103 242.0 0.0978 24.0 969.0 1196.0 0.8102 0.0201 210.0 789.0 1267.0 0.6227 0.1657
0.0328 7.0 42 0.9292 0.0056 3317.6903 2299.6477 1831.0 2475.0 0.7398 1492.0 0.6028 688.0 931.0 1196.0 0.7784 0.5753 796.0 900.0 1267.0 0.7103 0.6283
0.4624 8.0 48 1.4861 0.0056 5306.3463 3678.0790 1754.0 2475.0 0.7087 1681.0 0.6792 1018.0 1072.0 1196.0 0.8963 0.8512 656.0 682.0 1267.0 0.5383 0.5178
0.0041 9.0 54 1.3956 0.0056 4983.3753 3454.2125 1713.0 2475.0 0.6921 1328.0 0.5366 800.0 1080.0 1196.0 0.9030 0.6689 521.0 633.0 1267.0 0.4996 0.4112
0.0 10.0 60 1.2059 0.0056 4305.9186 2984.6353 1776.0 2475.0 0.7176 1744.0 0.7046 973.0 996.0 1196.0 0.8328 0.8135 764.0 780.0 1267.0 0.6156 0.6030
0.0001 11.0 66 1.2530 0.0056 4474.1803 3101.2654 1797.0 2475.0 0.7261 1797.0 0.7261 959.0 961.0 1196.0 0.8035 0.8018 830.0 836.0 1267.0 0.6598 0.6551
0.0019 12.0 72 1.2914 0.0056 4611.0049 3196.1051 1806.0 2475.0 0.7297 1812.0 0.7321 831.0 832.0 1196.0 0.6957 0.6948 973.0 974.0 1267.0 0.7687 0.7680
0.0 13.0 78 1.4823 0.0056 5292.8256 3668.7071 1752.0 2475.0 0.7079 1756.0 0.7095 1028.0 1029.0 1196.0 0.8604 0.8595 720.0 723.0 1267.0 0.5706 0.5683
0.9048 14.0 84 1.5439 0.0056 5512.8836 3821.2397 1744.0 2475.0 0.7046 1748.0 0.7063 1041.0 1042.0 1196.0 0.8712 0.8704 699.0 702.0 1267.0 0.5541 0.5517
0.4524 15.0 90 1.5327 0.0056 5472.8181 3793.4684 1741.0 2475.0 0.7034 1744.0 0.7046 1029.0 1031.0 1196.0 0.8620 0.8604 707.0 710.0 1267.0 0.5604 0.5580
0.0001 16.0 96 1.5159 0.0056 5412.8149 3751.8774 1750.0 2475.0 0.7071 1752.0 0.7079 1027.0 1029.0 1196.0 0.8604 0.8587 717.0 721.0 1267.0 0.5691 0.5659
0.0 17.0 102 1.4979 0.0056 5348.5086 3707.3037 1760.0 2475.0 0.7111 1762.0 0.7119 1022.0 1024.0 1196.0 0.8562 0.8545 732.0 736.0 1267.0 0.5809 0.5777
0.4524 18.0 108 1.4874 0.0056 5311.1411 3681.4025 1760.0 2475.0 0.7111 1764.0 0.7127 1018.0 1018.0 1196.0 0.8512 0.8512 738.0 742.0 1267.0 0.5856 0.5825
0.0 19.0 114 1.4810 0.0056 5288.0565 3665.4015 1771.0 2475.0 0.7156 1774.0 0.7168 1018.0 1018.0 1196.0 0.8512 0.8512 748.0 753.0 1267.0 0.5943 0.5904
0.0 20.0 120 1.4760 0.0056 5270.4369 3653.1885 1773.0 2475.0 0.7164 1776.0 0.7176 1014.0 1014.0 1196.0 0.8478 0.8478 754.0 759.0 1267.0 0.5991 0.5951
0.0 21.0 126 1.4728 0.0056 5258.7289 3645.0731 1773.0 2475.0 0.7164 1776.0 0.7176 1013.0 1013.0 1196.0 0.8470 0.8470 755.0 760.0 1267.0 0.5998 0.5959
0.0 22.0 132 1.4693 0.0056 5246.4381 3636.5537 1777.0 2475.0 0.7180 1779.0 0.7188 1013.0 1014.0 1196.0 0.8478 0.8470 758.0 763.0 1267.0 0.6022 0.5983
0.9048 23.0 138 1.4666 0.0056 5236.6019 3629.7358 1774.0 2475.0 0.7168 1776.0 0.7176 1012.0 1013.0 1196.0 0.8470 0.8462 756.0 761.0 1267.0 0.6006 0.5967
0.4524 24.0 144 1.4641 0.0056 5227.7933 3623.6302 1779.0 2475.0 0.7188 1782.0 0.72 1014.0 1014.0 1196.0 0.8478 0.8478 760.0 765.0 1267.0 0.6038 0.5998
0.0 25.0 150 1.4629 0.0056 5223.5871 3620.7147 1777.0 2475.0 0.7180 1780.0 0.7192 1012.0 1012.0 1196.0 0.8462 0.8462 760.0 765.0 1267.0 0.6038 0.5998
0.0 26.0 156 1.4613 0.0056 5217.9147 3616.7828 1781.0 2475.0 0.7196 1784.0 0.7208 1013.0 1013.0 1196.0 0.8470 0.8470 763.0 768.0 1267.0 0.6062 0.6022
0.0 27.0 162 1.4584 0.0056 5207.4129 3609.5036 1781.0 2475.0 0.7196 1784.0 0.7208 1011.0 1011.0 1196.0 0.8453 0.8453 765.0 770.0 1267.0 0.6077 0.6038
0.0 28.0 168 1.4583 0.0056 5207.2841 3609.4143 1779.0 2475.0 0.7188 1782.0 0.72 1009.0 1009.0 1196.0 0.8436 0.8436 765.0 770.0 1267.0 0.6077 0.6038
0.4524 29.0 174 1.4559 0.0056 5198.4690 3603.3041 1786.0 2475.0 0.7216 1790.0 0.7232 1012.0 1012.0 1196.0 0.8462 0.8462 770.0 774.0 1267.0 0.6109 0.6077
0.0 30.0 180 1.4550 0.0056 5195.2049 3601.0416 1785.0 2475.0 0.7212 1788.0 0.7224 1011.0 1011.0 1196.0 0.8453 0.8453 769.0 774.0 1267.0 0.6109 0.6069
0.0 31.0 186 1.4533 0.0056 5189.0929 3596.8051 1783.0 2475.0 0.7204 1787.0 0.7220 1011.0 1011.0 1196.0 0.8453 0.8453 768.0 772.0 1267.0 0.6093 0.6062
0.4524 32.0 192 1.4546 0.0056 5193.8924 3600.1319 1785.0 2475.0 0.7212 1789.0 0.7228 1009.0 1009.0 1196.0 0.8436 0.8436 772.0 776.0 1267.0 0.6125 0.6093
0.4524 33.0 198 1.4522 0.0056 5185.2381 3594.1332 1786.0 2475.0 0.7216 1789.0 0.7228 1011.0 1011.0 1196.0 0.8453 0.8453 770.0 775.0 1267.0 0.6117 0.6077
0.0 34.0 204 1.4512 0.0056 5181.7853 3591.7399 1785.0 2475.0 0.7212 1788.0 0.7224 1008.0 1008.0 1196.0 0.8428 0.8428 772.0 777.0 1267.0 0.6133 0.6093
0.4524 35.0 210 1.4503 0.0056 5178.6494 3589.5662 1786.0 2475.0 0.7216 1789.0 0.7228 1009.0 1009.0 1196.0 0.8436 0.8436 772.0 777.0 1267.0 0.6133 0.6093
0.0 36.0 216 1.4496 0.0056 5175.8705 3587.6401 1788.0 2475.0 0.7224 1792.0 0.7240 1007.0 1007.0 1196.0 0.8420 0.8420 777.0 781.0 1267.0 0.6164 0.6133
0.4524 37.0 222 1.4509 0.0056 5180.7493 3591.0218 1783.0 2475.0 0.7204 1787.0 0.7220 1005.0 1005.0 1196.0 0.8403 0.8403 774.0 778.0 1267.0 0.6140 0.6109

Framework versions

  • Transformers 4.51.3
  • Pytorch 2.6.0+cu124
  • Datasets 3.5.0
  • Tokenizers 0.21.1
Downloads last month
2
Safetensors
Model size
1B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for donoway/GSM8K-Binary_Llama-3.2-1B-fal8is8l

Finetuned
(903)
this model