ARC-Easy_Llama-3.2-1B-szzhws4z

This model is a fine-tuned version of meta-llama/Llama-3.2-1B on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 2.5495
  • Model Preparation Time: 0.006
  • Mdl: 2096.5226
  • Accumulated Loss: 1453.1987
  • Correct Preds: 411.0
  • Total Preds: 570.0
  • Accuracy: 0.7211
  • Correct Gen Preds: 410.0
  • Gen Accuracy: 0.7193
  • Correct Gen Preds 32: 112.0
  • Correct Preds 32: 113.0
  • Total Labels 32: 158.0
  • Accuracy 32: 0.7152
  • Gen Accuracy 32: 0.7089
  • Correct Gen Preds 33: 117.0
  • Correct Preds 33: 117.0
  • Total Labels 33: 152.0
  • Accuracy 33: 0.7697
  • Gen Accuracy 33: 0.7697
  • Correct Gen Preds 34: 108.0
  • Correct Preds 34: 108.0
  • Total Labels 34: 142.0
  • Accuracy 34: 0.7606
  • Gen Accuracy 34: 0.7606
  • Correct Gen Preds 35: 73.0
  • Correct Preds 35: 73.0
  • Total Labels 35: 118.0
  • Accuracy 35: 0.6186
  • Gen Accuracy 35: 0.6186
  • Correct Gen Preds 36: 0.0
  • Correct Preds 36: 0.0
  • Total Labels 36: 0.0
  • Accuracy 36: 0.0
  • Gen Accuracy 36: 0.0

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 64
  • eval_batch_size: 112
  • seed: 42
  • optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_ratio: 0.01
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Model Preparation Time Mdl Accumulated Loss Correct Preds Total Preds Accuracy Correct Gen Preds Gen Accuracy Correct Gen Preds 32 Correct Preds 32 Total Labels 32 Accuracy 32 Gen Accuracy 32 Correct Gen Preds 33 Correct Preds 33 Total Labels 33 Accuracy 33 Gen Accuracy 33 Correct Gen Preds 34 Correct Preds 34 Total Labels 34 Accuracy 34 Gen Accuracy 34 Correct Gen Preds 35 Correct Preds 35 Total Labels 35 Accuracy 35 Gen Accuracy 35 Correct Gen Preds 36 Correct Preds 36 Total Labels 36 Accuracy 36 Gen Accuracy 36
No log 0 0 1.5354 0.006 1262.6022 875.1692 172.0 570.0 0.3018 170.0 0.2982 154.0 154.0 158.0 0.9747 0.9747 0.0 0.0 152.0 0.0 0.0 15.0 17.0 142.0 0.1197 0.1056 1.0 1.0 118.0 0.0085 0.0085 0.0 0.0 0.0 0.0 0.0
0.6626 1.0 11 0.9446 0.006 776.7691 538.4153 380.0 570.0 0.6667 380.0 0.6667 122.0 122.0 158.0 0.7722 0.7722 103.0 103.0 152.0 0.6776 0.6776 94.0 94.0 142.0 0.6620 0.6620 61.0 61.0 118.0 0.5169 0.5169 0.0 0.0 0.0 0.0 0.0
0.4951 2.0 22 0.8259 0.006 679.1973 470.7837 406.0 570.0 0.7123 406.0 0.7123 100.0 100.0 158.0 0.6329 0.6329 111.0 111.0 152.0 0.7303 0.7303 115.0 115.0 142.0 0.8099 0.8099 80.0 80.0 118.0 0.6780 0.6780 0.0 0.0 0.0 0.0 0.0
0.1373 3.0 33 0.9945 0.006 817.8361 566.8808 408.0 570.0 0.7158 408.0 0.7158 115.0 115.0 158.0 0.7278 0.7278 114.0 114.0 152.0 0.75 0.75 96.0 96.0 142.0 0.6761 0.6761 83.0 83.0 118.0 0.7034 0.7034 0.0 0.0 0.0 0.0 0.0
0.0001 4.0 44 2.5495 0.006 2096.5226 1453.1987 411.0 570.0 0.7211 410.0 0.7193 112.0 113.0 158.0 0.7152 0.7089 117.0 117.0 152.0 0.7697 0.7697 108.0 108.0 142.0 0.7606 0.7606 73.0 73.0 118.0 0.6186 0.6186 0.0 0.0 0.0 0.0 0.0
0.0002 5.0 55 2.4175 0.006 1988.0273 1377.9955 406.0 570.0 0.7123 406.0 0.7123 98.0 98.0 158.0 0.6203 0.6203 117.0 117.0 152.0 0.7697 0.7697 111.0 111.0 142.0 0.7817 0.7817 80.0 80.0 118.0 0.6780 0.6780 0.0 0.0 0.0 0.0 0.0
0.0 6.0 66 2.4839 0.006 2042.5931 1415.8176 404.0 570.0 0.7088 404.0 0.7088 117.0 117.0 158.0 0.7405 0.7405 108.0 108.0 152.0 0.7105 0.7105 107.0 107.0 142.0 0.7535 0.7535 72.0 72.0 118.0 0.6102 0.6102 0.0 0.0 0.0 0.0 0.0
0.0001 7.0 77 2.6310 0.006 2163.5625 1499.6673 406.0 570.0 0.7123 406.0 0.7123 123.0 123.0 158.0 0.7785 0.7785 106.0 106.0 152.0 0.6974 0.6974 103.0 103.0 142.0 0.7254 0.7254 74.0 74.0 118.0 0.6271 0.6271 0.0 0.0 0.0 0.0 0.0
0.0 8.0 88 2.7352 0.006 2249.2482 1559.0600 400.0 570.0 0.7018 400.0 0.7018 112.0 112.0 158.0 0.7089 0.7089 108.0 108.0 152.0 0.7105 0.7105 101.0 101.0 142.0 0.7113 0.7113 79.0 79.0 118.0 0.6695 0.6695 0.0 0.0 0.0 0.0 0.0
0.0 9.0 99 2.7916 0.006 2295.6207 1591.2030 400.0 570.0 0.7018 400.0 0.7018 111.0 111.0 158.0 0.7025 0.7025 109.0 109.0 152.0 0.7171 0.7171 101.0 101.0 142.0 0.7113 0.7113 79.0 79.0 118.0 0.6695 0.6695 0.0 0.0 0.0 0.0 0.0
0.0 10.0 110 2.8267 0.006 2324.5097 1611.2273 396.0 570.0 0.6947 396.0 0.6947 109.0 109.0 158.0 0.6899 0.6899 109.0 109.0 152.0 0.7171 0.7171 99.0 99.0 142.0 0.6972 0.6972 79.0 79.0 118.0 0.6695 0.6695 0.0 0.0 0.0 0.0 0.0
0.0 11.0 121 2.8506 0.006 2344.1674 1624.8530 398.0 570.0 0.6982 398.0 0.6982 110.0 110.0 158.0 0.6962 0.6962 110.0 110.0 152.0 0.7237 0.7237 100.0 100.0 142.0 0.7042 0.7042 78.0 78.0 118.0 0.6610 0.6610 0.0 0.0 0.0 0.0 0.0
0.0 12.0 132 2.8582 0.006 2350.3907 1629.1667 401.0 570.0 0.7035 401.0 0.7035 111.0 111.0 158.0 0.7025 0.7025 110.0 110.0 152.0 0.7237 0.7237 102.0 102.0 142.0 0.7183 0.7183 78.0 78.0 118.0 0.6610 0.6610 0.0 0.0 0.0 0.0 0.0
0.0 13.0 143 2.8602 0.006 2352.0259 1630.3001 400.0 570.0 0.7018 400.0 0.7018 109.0 109.0 158.0 0.6899 0.6899 111.0 111.0 152.0 0.7303 0.7303 102.0 102.0 142.0 0.7183 0.7183 78.0 78.0 118.0 0.6610 0.6610 0.0 0.0 0.0 0.0 0.0
0.0 14.0 154 2.8996 0.006 2384.4652 1652.7854 401.0 570.0 0.7035 401.0 0.7035 112.0 112.0 158.0 0.7089 0.7089 111.0 111.0 152.0 0.7303 0.7303 99.0 99.0 142.0 0.6972 0.6972 79.0 79.0 118.0 0.6695 0.6695 0.0 0.0 0.0 0.0 0.0
0.0 15.0 165 2.8866 0.006 2373.7384 1645.3501 401.0 570.0 0.7035 401.0 0.7035 110.0 110.0 158.0 0.6962 0.6962 111.0 111.0 152.0 0.7303 0.7303 100.0 100.0 142.0 0.7042 0.7042 80.0 80.0 118.0 0.6780 0.6780 0.0 0.0 0.0 0.0 0.0
0.0 16.0 176 2.8719 0.006 2361.6528 1636.9730 403.0 570.0 0.7070 403.0 0.7070 111.0 111.0 158.0 0.7025 0.7025 112.0 112.0 152.0 0.7368 0.7368 100.0 100.0 142.0 0.7042 0.7042 80.0 80.0 118.0 0.6780 0.6780 0.0 0.0 0.0 0.0 0.0
0.0 17.0 187 2.9309 0.006 2410.2127 1670.6321 400.0 570.0 0.7018 400.0 0.7018 111.0 111.0 158.0 0.7025 0.7025 110.0 110.0 152.0 0.7237 0.7237 100.0 100.0 142.0 0.7042 0.7042 79.0 79.0 118.0 0.6695 0.6695 0.0 0.0 0.0 0.0 0.0
0.0 18.0 198 2.9284 0.006 2408.1119 1669.1760 400.0 570.0 0.7018 400.0 0.7018 110.0 110.0 158.0 0.6962 0.6962 111.0 111.0 152.0 0.7303 0.7303 100.0 100.0 142.0 0.7042 0.7042 79.0 79.0 118.0 0.6695 0.6695 0.0 0.0 0.0 0.0 0.0
0.0 19.0 209 2.9469 0.006 2423.3530 1679.7403 399.0 570.0 0.7 399.0 0.7 111.0 111.0 158.0 0.7025 0.7025 110.0 110.0 152.0 0.7237 0.7237 100.0 100.0 142.0 0.7042 0.7042 78.0 78.0 118.0 0.6610 0.6610 0.0 0.0 0.0 0.0 0.0
0.0 20.0 220 2.9537 0.006 2428.9387 1683.6120 401.0 570.0 0.7035 401.0 0.7035 110.0 110.0 158.0 0.6962 0.6962 111.0 111.0 152.0 0.7303 0.7303 101.0 101.0 142.0 0.7113 0.7113 79.0 79.0 118.0 0.6695 0.6695 0.0 0.0 0.0 0.0 0.0
0.0 21.0 231 2.9368 0.006 2415.0125 1673.9591 401.0 570.0 0.7035 401.0 0.7035 112.0 112.0 158.0 0.7089 0.7089 110.0 110.0 152.0 0.7237 0.7237 100.0 100.0 142.0 0.7042 0.7042 79.0 79.0 118.0 0.6695 0.6695 0.0 0.0 0.0 0.0 0.0
0.0 22.0 242 2.9519 0.006 2427.4650 1682.5905 397.0 570.0 0.6965 397.0 0.6965 110.0 110.0 158.0 0.6962 0.6962 109.0 109.0 152.0 0.7171 0.7171 99.0 99.0 142.0 0.6972 0.6972 79.0 79.0 118.0 0.6695 0.6695 0.0 0.0 0.0 0.0 0.0
0.0 23.0 253 2.9601 0.006 2434.1860 1687.2492 403.0 570.0 0.7070 403.0 0.7070 111.0 111.0 158.0 0.7025 0.7025 112.0 112.0 152.0 0.7368 0.7368 101.0 101.0 142.0 0.7113 0.7113 79.0 79.0 118.0 0.6695 0.6695 0.0 0.0 0.0 0.0 0.0
0.0 24.0 264 2.9746 0.006 2446.1592 1695.5484 400.0 570.0 0.7018 400.0 0.7018 110.0 110.0 158.0 0.6962 0.6962 110.0 110.0 152.0 0.7237 0.7237 101.0 101.0 142.0 0.7113 0.7113 79.0 79.0 118.0 0.6695 0.6695 0.0 0.0 0.0 0.0 0.0
0.0 25.0 275 2.9806 0.006 2451.0232 1698.9198 398.0 570.0 0.6982 398.0 0.6982 109.0 109.0 158.0 0.6899 0.6899 111.0 111.0 152.0 0.7303 0.7303 99.0 99.0 142.0 0.6972 0.6972 79.0 79.0 118.0 0.6695 0.6695 0.0 0.0 0.0 0.0 0.0
0.0 26.0 286 3.0022 0.006 2468.8163 1711.2530 400.0 570.0 0.7018 400.0 0.7018 110.0 110.0 158.0 0.6962 0.6962 110.0 110.0 152.0 0.7237 0.7237 101.0 101.0 142.0 0.7113 0.7113 79.0 79.0 118.0 0.6695 0.6695 0.0 0.0 0.0 0.0 0.0
0.0 27.0 297 2.9606 0.006 2434.6001 1687.5362 405.0 570.0 0.7105 405.0 0.7105 113.0 113.0 158.0 0.7152 0.7152 111.0 111.0 152.0 0.7303 0.7303 102.0 102.0 142.0 0.7183 0.7183 79.0 79.0 118.0 0.6695 0.6695 0.0 0.0 0.0 0.0 0.0
0.0 28.0 308 2.9773 0.006 2448.3431 1697.0621 403.0 570.0 0.7070 403.0 0.7070 112.0 112.0 158.0 0.7089 0.7089 110.0 110.0 152.0 0.7237 0.7237 102.0 102.0 142.0 0.7183 0.7183 79.0 79.0 118.0 0.6695 0.6695 0.0 0.0 0.0 0.0 0.0
0.0 29.0 319 3.0146 0.006 2478.9969 1718.3097 400.0 570.0 0.7018 400.0 0.7018 110.0 110.0 158.0 0.6962 0.6962 112.0 112.0 152.0 0.7368 0.7368 100.0 100.0 142.0 0.7042 0.7042 78.0 78.0 118.0 0.6610 0.6610 0.0 0.0 0.0 0.0 0.0
0.0 30.0 330 2.9987 0.006 2465.9668 1709.2779 399.0 570.0 0.7 399.0 0.7 109.0 109.0 158.0 0.6899 0.6899 111.0 111.0 152.0 0.7303 0.7303 100.0 100.0 142.0 0.7042 0.7042 79.0 79.0 118.0 0.6695 0.6695 0.0 0.0 0.0 0.0 0.0
0.0 31.0 341 3.0034 0.006 2469.7829 1711.9230 399.0 570.0 0.7 399.0 0.7 110.0 110.0 158.0 0.6962 0.6962 110.0 110.0 152.0 0.7237 0.7237 100.0 100.0 142.0 0.7042 0.7042 79.0 79.0 118.0 0.6695 0.6695 0.0 0.0 0.0 0.0 0.0
0.0 32.0 352 3.0066 0.006 2472.4048 1713.7404 402.0 570.0 0.7053 402.0 0.7053 109.0 109.0 158.0 0.6899 0.6899 112.0 112.0 152.0 0.7368 0.7368 102.0 102.0 142.0 0.7183 0.7183 79.0 79.0 118.0 0.6695 0.6695 0.0 0.0 0.0 0.0 0.0
0.0 33.0 363 3.0106 0.006 2475.7556 1716.0630 402.0 570.0 0.7053 402.0 0.7053 112.0 112.0 158.0 0.7089 0.7089 111.0 111.0 152.0 0.7303 0.7303 99.0 99.0 142.0 0.6972 0.6972 80.0 80.0 118.0 0.6780 0.6780 0.0 0.0 0.0 0.0 0.0
0.0 34.0 374 3.0176 0.006 2481.4695 1720.0236 401.0 570.0 0.7035 401.0 0.7035 112.0 112.0 158.0 0.7089 0.7089 111.0 111.0 152.0 0.7303 0.7303 100.0 100.0 142.0 0.7042 0.7042 78.0 78.0 118.0 0.6610 0.6610 0.0 0.0 0.0 0.0 0.0

Framework versions

  • Transformers 4.51.3
  • Pytorch 2.6.0+cu124
  • Datasets 3.5.0
  • Tokenizers 0.21.1
Downloads last month
3
Safetensors
Model size
1B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for donoway/ARC-Easy_Llama-3.2-1B-szzhws4z

Finetuned
(900)
this model