tlocvsdyspneaTask-Llama-3.2-1B-Instruct-all

This model is a fine-tuned version of meta-llama/Llama-3.2-1B-Instruct on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 2.7712
  • F1 Micro: 0.6712
  • F1 Macro: 0.4036
  • F1 Weighted: 0.8029

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0003
  • train_batch_size: 8
  • eval_batch_size: 32
  • seed: 42
  • distributed_type: multi-GPU
  • gradient_accumulation_steps: 8
  • total_train_batch_size: 64
  • optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-07 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_ratio: 0.1
  • num_epochs: 3

Training results

Training Loss Epoch Step Validation Loss F1 Micro F1 Macro F1 Weighted
4.4852 0.1067 2 4.5015 0.0 0.0 0.0
4.4852 0.2133 4 4.2641 0.0 0.0 0.0
4.4158 0.32 6 3.9209 0.0 0.0 0.0
4.4158 0.4267 8 3.7165 0.0 0.0 0.0
3.8219 0.5333 10 3.5813 0.0 0.0 0.0
3.8219 0.64 12 3.4265 0.0058 0.0001 0.0115
3.8219 0.7467 14 3.3091 0.0370 0.0004 0.0683
3.4588 0.8533 16 3.2119 0.6323 0.0916 0.6747
3.4588 0.96 18 3.1466 0.6284 0.2871 0.6593
3.186 1.0533 20 3.1164 0.6868 0.3884 0.8022
3.186 1.16 22 3.0644 0.4747 0.3149 0.5625
3.186 1.2667 24 3.0140 0.6537 0.3237 0.7150
3.0434 1.3733 26 2.9700 0.6868 0.3896 0.8034
3.0434 1.48 28 2.9330 0.5039 0.3309 0.6000
2.9228 1.5867 30 2.8991 0.6440 0.3076 0.6911
2.9228 1.6933 32 2.8724 0.6673 0.3402 0.7401
2.9228 1.8 34 2.8495 0.5389 0.3504 0.6478
2.8011 1.9067 36 2.8293 0.6089 0.3847 0.7393
2.8011 2.0 38 2.8130 0.6693 0.3371 0.7365
2.846 2.1067 40 2.8003 0.6401 0.3926 0.6721
2.846 2.2133 42 2.7901 0.6673 0.3266 0.7220
2.846 2.32 44 2.7828 0.7023 0.3992 0.8179
2.7493 2.4267 46 2.7777 0.6479 0.3982 0.7826
2.7493 2.5333 48 2.7743 0.6459 0.3981 0.7792
2.7241 2.64 50 2.7722 0.6420 0.3956 0.7791
2.7241 2.7467 52 2.7714 0.6595 0.4006 0.7937
2.7241 2.8533 54 2.7712 0.6712 0.4036 0.8029

Framework versions

  • PEFT 0.18.1
  • Transformers 4.51.0
  • Pytorch 2.8.0+cu128
  • Datasets 3.6.0
  • Tokenizers 0.21.0
Downloads last month
43
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for ferrazzipietro/tlocvsdyspneaTask-Llama-3.2-1B-Instruct-all

Adapter
(602)
this model