dv347 commited on
Commit
5f0ba68
·
verified ·
1 Parent(s): f7e74d2

Model save

Browse files
Files changed (2) hide show
  1. README.md +18 -18
  2. model.safetensors +1 -1
README.md CHANGED
@@ -16,11 +16,13 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 0.8715
20
  - Exact Match: 0.0
21
- - Micro F1: 0.3416
22
- - Macro F1: 0.0445
23
- - Hamming Accuracy: 0.8703
 
 
24
 
25
  ## Model description
26
 
@@ -39,29 +41,27 @@ More information needed
39
  ### Training hyperparameters
40
 
41
  The following hyperparameters were used during training:
42
- - learning_rate: 5e-06
43
  - train_batch_size: 32
44
  - eval_batch_size: 64
45
  - seed: 42
 
 
46
  - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
47
  - lr_scheduler_type: cosine
48
  - lr_scheduler_warmup_steps: 0.2
49
- - num_epochs: 5
50
 
51
  ### Training results
52
 
53
- | Training Loss | Epoch | Step | Validation Loss | Exact Match | Micro F1 | Macro F1 | Hamming Accuracy |
54
- |:-------------:|:-----:|:----:|:---------------:|:-----------:|:--------:|:--------:|:----------------:|
55
- | 0.4892 | 0.5 | 328 | 1.3449 | 0.0 | 0.3826 | 0.0421 | 0.8956 |
56
- | 0.5243 | 1.0 | 656 | 0.9100 | 0.0 | 0.3610 | 0.0367 | 0.8968 |
57
- | 0.5423 | 1.5 | 984 | 0.8660 | 0.0 | 0.3658 | 0.0484 | 0.8723 |
58
- | 0.5105 | 2.0 | 1312 | 0.8797 | 0.0 | 0.3385 | 0.0421 | 0.8740 |
59
- | 0.5304 | 2.5 | 1640 | 0.8877 | 0.0 | 0.3756 | 0.0464 | 0.8833 |
60
- | 0.5187 | 3.0 | 1968 | 0.9012 | 0.0 | 0.3827 | 0.0410 | 0.8981 |
61
- | 0.5122 | 3.5 | 2296 | 0.8840 | 0.0 | 0.3708 | 0.0458 | 0.8824 |
62
- | 0.5080 | 4.0 | 2624 | 0.8731 | 0.0 | 0.3919 | 0.0519 | 0.8776 |
63
- | 0.5403 | 4.5 | 2952 | 0.8727 | 0.0 | 0.3398 | 0.0443 | 0.8695 |
64
- | 0.5245 | 5.0 | 3280 | 0.8715 | 0.0 | 0.3416 | 0.0445 | 0.8703 |
65
 
66
 
67
  ### Framework versions
 
16
 
17
  This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 4.0967
20
  - Exact Match: 0.0
21
+ - Micro F1: 0.3075
22
+ - Macro F1: 0.0334
23
+ - Hamming Accuracy: 0.8806
24
+ - Avg Pred Positives: 34.0
25
+ - Avg Gold Positives: 13.5736
26
 
27
  ## Model description
28
 
 
41
  ### Training hyperparameters
42
 
43
  The following hyperparameters were used during training:
44
+ - learning_rate: 2e-05
45
  - train_batch_size: 32
46
  - eval_batch_size: 64
47
  - seed: 42
48
+ - gradient_accumulation_steps: 2
49
+ - total_train_batch_size: 64
50
  - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
51
  - lr_scheduler_type: cosine
52
  - lr_scheduler_warmup_steps: 0.2
53
+ - num_epochs: 10
54
 
55
  ### Training results
56
 
57
+ | Training Loss | Epoch | Step | Validation Loss | Exact Match | Micro F1 | Macro F1 | Hamming Accuracy | Avg Pred Positives | Avg Gold Positives |
58
+ |:-------------:|:-----:|:----:|:---------------:|:-----------:|:--------:|:--------:|:----------------:|:------------------:|:------------------:|
59
+ | 0.2930 | 0.5 | 164 | 0.5663 | 0.0 | 0.3139 | 0.0288 | 0.9041 | 25.0 | 13.5736 |
60
+ | 0.1928 | 1.0 | 328 | 0.2720 | 0.0 | 0.4392 | 0.0256 | 0.9460 | 13.0 | 13.5736 |
61
+ | 0.0751 | 1.5 | 492 | 0.0559 | 0.0 | 0.5244 | 0.0234 | 0.9628 | 8.0 | 13.5736 |
62
+ | 33.3599 | 2.0 | 656 | 13.5265 | 0.0 | 0.2931 | 0.0226 | 0.9114 | 21.0 | 13.5736 |
63
+ | 19.0859 | 2.5 | 820 | 10.5864 | 0.0 | 0.2214 | 0.0302 | 0.8376 | 44.0 | 13.5736 |
64
+ | 8.6145 | 3.0 | 984 | 4.0967 | 0.0 | 0.3075 | 0.0334 | 0.8806 | 34.0 | 13.5736 |
 
 
 
 
65
 
66
 
67
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a94b77ae063bb226caaac9a70cba275eab900d8227cc136a6ddc27941f7d855f
3
  size 870738024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd8e1f4b5b9413d593b5ba27ab1c4bb8c6921bda8faec79056c481955b303dc6
3
  size 870738024