Deaksh commited on
Commit
9e18434
·
verified ·
1 Parent(s): ba26c9a

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .DS_Store +0 -0
  2. checkpoint-104/config.json +46 -0
  3. checkpoint-104/model.safetensors +3 -0
  4. checkpoint-104/optimizer.pt +3 -0
  5. checkpoint-104/rng_state.pth +3 -0
  6. checkpoint-104/scheduler.pt +3 -0
  7. checkpoint-104/tokenizer.json +0 -0
  8. checkpoint-104/tokenizer_config.json +14 -0
  9. checkpoint-104/trainer_state.json +92 -0
  10. checkpoint-104/training_args.bin +3 -0
  11. checkpoint-130/config.json +46 -0
  12. checkpoint-130/model.safetensors +3 -0
  13. checkpoint-130/optimizer.pt +3 -0
  14. checkpoint-130/rng_state.pth +3 -0
  15. checkpoint-130/scheduler.pt +3 -0
  16. checkpoint-130/tokenizer.json +0 -0
  17. checkpoint-130/tokenizer_config.json +14 -0
  18. checkpoint-130/trainer_state.json +103 -0
  19. checkpoint-130/training_args.bin +3 -0
  20. checkpoint-156/config.json +46 -0
  21. checkpoint-156/model.safetensors +3 -0
  22. checkpoint-156/optimizer.pt +3 -0
  23. checkpoint-156/rng_state.pth +3 -0
  24. checkpoint-156/scheduler.pt +3 -0
  25. checkpoint-156/tokenizer.json +0 -0
  26. checkpoint-156/tokenizer_config.json +14 -0
  27. checkpoint-156/trainer_state.json +121 -0
  28. checkpoint-156/training_args.bin +3 -0
  29. checkpoint-26/model.safetensors +1 -1
  30. checkpoint-26/optimizer.pt +2 -2
  31. checkpoint-26/rng_state.pth +1 -1
  32. checkpoint-26/scheduler.pt +1 -1
  33. checkpoint-26/tokenizer.json +2 -2
  34. checkpoint-26/trainer_state.json +12 -12
  35. checkpoint-26/training_args.bin +1 -1
  36. checkpoint-52/model.safetensors +1 -1
  37. checkpoint-52/optimizer.pt +2 -2
  38. checkpoint-52/rng_state.pth +1 -1
  39. checkpoint-52/scheduler.pt +1 -1
  40. checkpoint-52/tokenizer.json +2 -2
  41. checkpoint-52/trainer_state.json +22 -22
  42. checkpoint-52/training_args.bin +1 -1
  43. checkpoint-78/model.safetensors +1 -1
  44. checkpoint-78/optimizer.pt +2 -2
  45. checkpoint-78/rng_state.pth +1 -1
  46. checkpoint-78/scheduler.pt +1 -1
  47. checkpoint-78/tokenizer.json +2 -2
  48. checkpoint-78/trainer_state.json +30 -30
  49. checkpoint-78/training_args.bin +1 -1
  50. model.safetensors +1 -1
.DS_Store ADDED
Binary file (8.2 kB). View file
 
checkpoint-104/config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_cross_attention": false,
3
+ "architectures": [
4
+ "BertForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": null,
8
+ "classifier_dropout": null,
9
+ "dtype": "float32",
10
+ "eos_token_id": null,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "id2label": {
15
+ "0": "B-ADE",
16
+ "1": "B-AGE",
17
+ "2": "B-DRUG",
18
+ "3": "I-ADE",
19
+ "4": "I-AGE",
20
+ "5": "I-DRUG",
21
+ "6": "O"
22
+ },
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 3072,
25
+ "is_decoder": false,
26
+ "label2id": {
27
+ "B-ADE": 0,
28
+ "B-AGE": 1,
29
+ "B-DRUG": 2,
30
+ "I-ADE": 3,
31
+ "I-AGE": 4,
32
+ "I-DRUG": 5,
33
+ "O": 6
34
+ },
35
+ "layer_norm_eps": 1e-12,
36
+ "max_position_embeddings": 512,
37
+ "model_type": "bert",
38
+ "num_attention_heads": 12,
39
+ "num_hidden_layers": 12,
40
+ "pad_token_id": 0,
41
+ "tie_word_embeddings": true,
42
+ "transformers_version": "5.3.0",
43
+ "type_vocab_size": 2,
44
+ "use_cache": false,
45
+ "vocab_size": 28996
46
+ }
checkpoint-104/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be524d20e72a49452e9c17b768a95af44cbcd1be2136c087389ecde98e54978a
3
+ size 430923564
checkpoint-104/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ddfa9681c3a7a174a931529b85f7f2c0d9b26beb70b541f46ee0014025cbb639
3
+ size 226911819
checkpoint-104/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35f9ae804b67082c0f5e66aaffc58dbaf19219b300adf8a8537f84981da0a4ee
3
+ size 14645
checkpoint-104/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:211440b956b84618efb5795f4adc8c7b304b277a8f154839104dd8279ad04d0f
3
+ size 1465
checkpoint-104/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-104/tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "cls_token": "[CLS]",
4
+ "do_lower_case": true,
5
+ "is_local": false,
6
+ "mask_token": "[MASK]",
7
+ "model_max_length": 1000000000000000019884624838656,
8
+ "pad_token": "[PAD]",
9
+ "sep_token": "[SEP]",
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "BertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
checkpoint-104/trainer_state.json ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 104,
3
+ "best_metric": 0.44260814785957336,
4
+ "best_model_checkpoint": "/content/drive/MyDrive/ADEGuard/ner_biobert_final_v2/checkpoint-104",
5
+ "epoch": 4.0,
6
+ "eval_steps": 500,
7
+ "global_step": 104,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 1.0,
14
+ "eval_f1": 0.00783944810285356,
15
+ "eval_loss": 1.139850378036499,
16
+ "eval_precision": 0.0039488232506713,
17
+ "eval_recall": 0.5319148936170213,
18
+ "eval_runtime": 0.5321,
19
+ "eval_samples_per_second": 48.861,
20
+ "eval_steps_per_second": 7.517,
21
+ "step": 26
22
+ },
23
+ {
24
+ "epoch": 1.9230769230769231,
25
+ "grad_norm": 2.009291410446167,
26
+ "learning_rate": 2.057692307692308e-05,
27
+ "loss": 1.1273328399658202,
28
+ "step": 50
29
+ },
30
+ {
31
+ "epoch": 2.0,
32
+ "eval_f1": 0.009548896937767533,
33
+ "eval_loss": 0.6769160032272339,
34
+ "eval_precision": 0.004811680769868923,
35
+ "eval_recall": 0.6170212765957447,
36
+ "eval_runtime": 0.4451,
37
+ "eval_samples_per_second": 58.415,
38
+ "eval_steps_per_second": 8.987,
39
+ "step": 52
40
+ },
41
+ {
42
+ "epoch": 3.0,
43
+ "eval_f1": 0.011583871686344399,
44
+ "eval_loss": 0.5276753902435303,
45
+ "eval_precision": 0.00585321927059883,
46
+ "eval_recall": 0.5531914893617021,
47
+ "eval_runtime": 0.5507,
48
+ "eval_samples_per_second": 47.215,
49
+ "eval_steps_per_second": 7.264,
50
+ "step": 78
51
+ },
52
+ {
53
+ "epoch": 3.8461538461538463,
54
+ "grad_norm": 3.652482748031616,
55
+ "learning_rate": 1.0961538461538462e-05,
56
+ "loss": 0.4301648712158203,
57
+ "step": 100
58
+ },
59
+ {
60
+ "epoch": 4.0,
61
+ "eval_f1": 0.03429923122412773,
62
+ "eval_loss": 0.44260814785957336,
63
+ "eval_precision": 0.017639902676399026,
64
+ "eval_recall": 0.6170212765957447,
65
+ "eval_runtime": 0.4831,
66
+ "eval_samples_per_second": 53.818,
67
+ "eval_steps_per_second": 8.28,
68
+ "step": 104
69
+ }
70
+ ],
71
+ "logging_steps": 50,
72
+ "max_steps": 156,
73
+ "num_input_tokens_seen": 0,
74
+ "num_train_epochs": 6,
75
+ "save_steps": 500,
76
+ "stateful_callbacks": {
77
+ "TrainerControl": {
78
+ "args": {
79
+ "should_epoch_stop": false,
80
+ "should_evaluate": false,
81
+ "should_log": false,
82
+ "should_save": true,
83
+ "should_training_stop": false
84
+ },
85
+ "attributes": {}
86
+ }
87
+ },
88
+ "total_flos": 54352182263808.0,
89
+ "train_batch_size": 4,
90
+ "trial_name": null,
91
+ "trial_params": null
92
+ }
checkpoint-104/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28773adb83572bc94ef9e073a42eb194223d823b5b140ca2b7c77bf0c3f76e9e
3
+ size 5201
checkpoint-130/config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_cross_attention": false,
3
+ "architectures": [
4
+ "BertForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": null,
8
+ "classifier_dropout": null,
9
+ "dtype": "float32",
10
+ "eos_token_id": null,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "id2label": {
15
+ "0": "B-ADE",
16
+ "1": "B-AGE",
17
+ "2": "B-DRUG",
18
+ "3": "I-ADE",
19
+ "4": "I-AGE",
20
+ "5": "I-DRUG",
21
+ "6": "O"
22
+ },
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 3072,
25
+ "is_decoder": false,
26
+ "label2id": {
27
+ "B-ADE": 0,
28
+ "B-AGE": 1,
29
+ "B-DRUG": 2,
30
+ "I-ADE": 3,
31
+ "I-AGE": 4,
32
+ "I-DRUG": 5,
33
+ "O": 6
34
+ },
35
+ "layer_norm_eps": 1e-12,
36
+ "max_position_embeddings": 512,
37
+ "model_type": "bert",
38
+ "num_attention_heads": 12,
39
+ "num_hidden_layers": 12,
40
+ "pad_token_id": 0,
41
+ "tie_word_embeddings": true,
42
+ "transformers_version": "5.3.0",
43
+ "type_vocab_size": 2,
44
+ "use_cache": false,
45
+ "vocab_size": 28996
46
+ }
checkpoint-130/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3952445c17e00ef94784c968198848e7e752ac66bab27f0f790c5b33d02710bf
3
+ size 430923564
checkpoint-130/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1b52c2190e61ededd59d82375ff39255d5514b5f7515890526a252726477dda
3
+ size 226911819
checkpoint-130/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffd56c42ada7f08402b22f05bde9943acf13d0f86760a91a934f02bec06e3f74
3
+ size 14645
checkpoint-130/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6cd97dc6d2b9e48dd48b602ecbfe0bc2440557eb03cb5f43b959e23c524fadba
3
+ size 1465
checkpoint-130/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-130/tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "cls_token": "[CLS]",
4
+ "do_lower_case": true,
5
+ "is_local": false,
6
+ "mask_token": "[MASK]",
7
+ "model_max_length": 1000000000000000019884624838656,
8
+ "pad_token": "[PAD]",
9
+ "sep_token": "[SEP]",
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "BertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
checkpoint-130/trainer_state.json ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 130,
3
+ "best_metric": 0.42207130789756775,
4
+ "best_model_checkpoint": "/content/drive/MyDrive/ADEGuard/ner_biobert_final_v2/checkpoint-130",
5
+ "epoch": 5.0,
6
+ "eval_steps": 500,
7
+ "global_step": 130,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 1.0,
14
+ "eval_f1": 0.00783944810285356,
15
+ "eval_loss": 1.139850378036499,
16
+ "eval_precision": 0.0039488232506713,
17
+ "eval_recall": 0.5319148936170213,
18
+ "eval_runtime": 0.5321,
19
+ "eval_samples_per_second": 48.861,
20
+ "eval_steps_per_second": 7.517,
21
+ "step": 26
22
+ },
23
+ {
24
+ "epoch": 1.9230769230769231,
25
+ "grad_norm": 2.009291410446167,
26
+ "learning_rate": 2.057692307692308e-05,
27
+ "loss": 1.1273328399658202,
28
+ "step": 50
29
+ },
30
+ {
31
+ "epoch": 2.0,
32
+ "eval_f1": 0.009548896937767533,
33
+ "eval_loss": 0.6769160032272339,
34
+ "eval_precision": 0.004811680769868923,
35
+ "eval_recall": 0.6170212765957447,
36
+ "eval_runtime": 0.4451,
37
+ "eval_samples_per_second": 58.415,
38
+ "eval_steps_per_second": 8.987,
39
+ "step": 52
40
+ },
41
+ {
42
+ "epoch": 3.0,
43
+ "eval_f1": 0.011583871686344399,
44
+ "eval_loss": 0.5276753902435303,
45
+ "eval_precision": 0.00585321927059883,
46
+ "eval_recall": 0.5531914893617021,
47
+ "eval_runtime": 0.5507,
48
+ "eval_samples_per_second": 47.215,
49
+ "eval_steps_per_second": 7.264,
50
+ "step": 78
51
+ },
52
+ {
53
+ "epoch": 3.8461538461538463,
54
+ "grad_norm": 3.652482748031616,
55
+ "learning_rate": 1.0961538461538462e-05,
56
+ "loss": 0.4301648712158203,
57
+ "step": 100
58
+ },
59
+ {
60
+ "epoch": 4.0,
61
+ "eval_f1": 0.03429923122412773,
62
+ "eval_loss": 0.44260814785957336,
63
+ "eval_precision": 0.017639902676399026,
64
+ "eval_recall": 0.6170212765957447,
65
+ "eval_runtime": 0.4831,
66
+ "eval_samples_per_second": 53.818,
67
+ "eval_steps_per_second": 8.28,
68
+ "step": 104
69
+ },
70
+ {
71
+ "epoch": 5.0,
72
+ "eval_f1": 0.040502793296089384,
73
+ "eval_loss": 0.42207130789756775,
74
+ "eval_precision": 0.020938628158844765,
75
+ "eval_recall": 0.6170212765957447,
76
+ "eval_runtime": 0.4485,
77
+ "eval_samples_per_second": 57.967,
78
+ "eval_steps_per_second": 8.918,
79
+ "step": 130
80
+ }
81
+ ],
82
+ "logging_steps": 50,
83
+ "max_steps": 156,
84
+ "num_input_tokens_seen": 0,
85
+ "num_train_epochs": 6,
86
+ "save_steps": 500,
87
+ "stateful_callbacks": {
88
+ "TrainerControl": {
89
+ "args": {
90
+ "should_epoch_stop": false,
91
+ "should_evaluate": false,
92
+ "should_log": false,
93
+ "should_save": true,
94
+ "should_training_stop": false
95
+ },
96
+ "attributes": {}
97
+ }
98
+ },
99
+ "total_flos": 67940227829760.0,
100
+ "train_batch_size": 4,
101
+ "trial_name": null,
102
+ "trial_params": null
103
+ }
checkpoint-130/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28773adb83572bc94ef9e073a42eb194223d823b5b140ca2b7c77bf0c3f76e9e
3
+ size 5201
checkpoint-156/config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_cross_attention": false,
3
+ "architectures": [
4
+ "BertForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": null,
8
+ "classifier_dropout": null,
9
+ "dtype": "float32",
10
+ "eos_token_id": null,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "id2label": {
15
+ "0": "B-ADE",
16
+ "1": "B-AGE",
17
+ "2": "B-DRUG",
18
+ "3": "I-ADE",
19
+ "4": "I-AGE",
20
+ "5": "I-DRUG",
21
+ "6": "O"
22
+ },
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 3072,
25
+ "is_decoder": false,
26
+ "label2id": {
27
+ "B-ADE": 0,
28
+ "B-AGE": 1,
29
+ "B-DRUG": 2,
30
+ "I-ADE": 3,
31
+ "I-AGE": 4,
32
+ "I-DRUG": 5,
33
+ "O": 6
34
+ },
35
+ "layer_norm_eps": 1e-12,
36
+ "max_position_embeddings": 512,
37
+ "model_type": "bert",
38
+ "num_attention_heads": 12,
39
+ "num_hidden_layers": 12,
40
+ "pad_token_id": 0,
41
+ "tie_word_embeddings": true,
42
+ "transformers_version": "5.3.0",
43
+ "type_vocab_size": 2,
44
+ "use_cache": false,
45
+ "vocab_size": 28996
46
+ }
checkpoint-156/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8118ef61650dcb39f2a759979b7d4de786801e0fda2c09b835e757c956d92f51
3
+ size 430923564
checkpoint-156/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50eec345bb88d986ff0e79b966613c5623a8099d77948c9eaaeb3e9b82624804
3
+ size 226911819
checkpoint-156/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56b9ffcfea7e20d15051e8936d76c0e26b05a1983ab529e562e8f944b7bd373f
3
+ size 14645
checkpoint-156/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2b973829473ca01589916f767b33fa10f572dbffbc4cab770082eb09d9b6f17
3
+ size 1465
checkpoint-156/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-156/tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "cls_token": "[CLS]",
4
+ "do_lower_case": true,
5
+ "is_local": false,
6
+ "mask_token": "[MASK]",
7
+ "model_max_length": 1000000000000000019884624838656,
8
+ "pad_token": "[PAD]",
9
+ "sep_token": "[SEP]",
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "BertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
checkpoint-156/trainer_state.json ADDED
@@ -0,0 +1,121 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 130,
3
+ "best_metric": 0.42207130789756775,
4
+ "best_model_checkpoint": "/content/drive/MyDrive/ADEGuard/ner_biobert_final_v2/checkpoint-130",
5
+ "epoch": 6.0,
6
+ "eval_steps": 500,
7
+ "global_step": 156,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 1.0,
14
+ "eval_f1": 0.00783944810285356,
15
+ "eval_loss": 1.139850378036499,
16
+ "eval_precision": 0.0039488232506713,
17
+ "eval_recall": 0.5319148936170213,
18
+ "eval_runtime": 0.5321,
19
+ "eval_samples_per_second": 48.861,
20
+ "eval_steps_per_second": 7.517,
21
+ "step": 26
22
+ },
23
+ {
24
+ "epoch": 1.9230769230769231,
25
+ "grad_norm": 2.009291410446167,
26
+ "learning_rate": 2.057692307692308e-05,
27
+ "loss": 1.1273328399658202,
28
+ "step": 50
29
+ },
30
+ {
31
+ "epoch": 2.0,
32
+ "eval_f1": 0.009548896937767533,
33
+ "eval_loss": 0.6769160032272339,
34
+ "eval_precision": 0.004811680769868923,
35
+ "eval_recall": 0.6170212765957447,
36
+ "eval_runtime": 0.4451,
37
+ "eval_samples_per_second": 58.415,
38
+ "eval_steps_per_second": 8.987,
39
+ "step": 52
40
+ },
41
+ {
42
+ "epoch": 3.0,
43
+ "eval_f1": 0.011583871686344399,
44
+ "eval_loss": 0.5276753902435303,
45
+ "eval_precision": 0.00585321927059883,
46
+ "eval_recall": 0.5531914893617021,
47
+ "eval_runtime": 0.5507,
48
+ "eval_samples_per_second": 47.215,
49
+ "eval_steps_per_second": 7.264,
50
+ "step": 78
51
+ },
52
+ {
53
+ "epoch": 3.8461538461538463,
54
+ "grad_norm": 3.652482748031616,
55
+ "learning_rate": 1.0961538461538462e-05,
56
+ "loss": 0.4301648712158203,
57
+ "step": 100
58
+ },
59
+ {
60
+ "epoch": 4.0,
61
+ "eval_f1": 0.03429923122412773,
62
+ "eval_loss": 0.44260814785957336,
63
+ "eval_precision": 0.017639902676399026,
64
+ "eval_recall": 0.6170212765957447,
65
+ "eval_runtime": 0.4831,
66
+ "eval_samples_per_second": 53.818,
67
+ "eval_steps_per_second": 8.28,
68
+ "step": 104
69
+ },
70
+ {
71
+ "epoch": 5.0,
72
+ "eval_f1": 0.040502793296089384,
73
+ "eval_loss": 0.42207130789756775,
74
+ "eval_precision": 0.020938628158844765,
75
+ "eval_recall": 0.6170212765957447,
76
+ "eval_runtime": 0.4485,
77
+ "eval_samples_per_second": 57.967,
78
+ "eval_steps_per_second": 8.918,
79
+ "step": 130
80
+ },
81
+ {
82
+ "epoch": 5.769230769230769,
83
+ "grad_norm": 1.719118595123291,
84
+ "learning_rate": 1.3461538461538462e-06,
85
+ "loss": 0.18102121353149414,
86
+ "step": 150
87
+ },
88
+ {
89
+ "epoch": 6.0,
90
+ "eval_f1": 0.04627249357326478,
91
+ "eval_loss": 0.4319979250431061,
92
+ "eval_precision": 0.024107142857142858,
93
+ "eval_recall": 0.574468085106383,
94
+ "eval_runtime": 0.4511,
95
+ "eval_samples_per_second": 57.634,
96
+ "eval_steps_per_second": 8.867,
97
+ "step": 156
98
+ }
99
+ ],
100
+ "logging_steps": 50,
101
+ "max_steps": 156,
102
+ "num_input_tokens_seen": 0,
103
+ "num_train_epochs": 6,
104
+ "save_steps": 500,
105
+ "stateful_callbacks": {
106
+ "TrainerControl": {
107
+ "args": {
108
+ "should_epoch_stop": false,
109
+ "should_evaluate": false,
110
+ "should_log": false,
111
+ "should_save": true,
112
+ "should_training_stop": true
113
+ },
114
+ "attributes": {}
115
+ }
116
+ },
117
+ "total_flos": 81528273395712.0,
118
+ "train_batch_size": 4,
119
+ "trial_name": null,
120
+ "trial_params": null
121
+ }
checkpoint-156/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28773adb83572bc94ef9e073a42eb194223d823b5b140ca2b7c77bf0c3f76e9e
3
+ size 5201
checkpoint-26/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d9c2496469f79c1fb87c1591444607745d8e6108ca4adb7d45140bf2dff46af5
3
  size 430923564
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb4423e4ddb3f4c26593b0a62f46f10dd1cb278ccfa6e9d8bd79759e1f96b2d4
3
  size 430923564
checkpoint-26/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:845bc460a8c73f457650d48c009c3db103f3efc570a6bf9285776f8bb6992256
3
- size 861970123
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52ddee98d8539dcf61c7551dffe234ddc6c2b750ca0dec5cf3d384974726d72d
3
+ size 226911819
checkpoint-26/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a5a7422dbe51806d882a2479da08afcb5a46f17dd87b6371d4e78b65d56b8ea
3
  size 14645
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5691beddb7ad8a4f355ae33432fde96e4488ff9d27a171ca1d3335528d0b4362
3
  size 14645
checkpoint-26/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ebfcfb4944bd4fb2e03e1e04992fc304c0f65cf8dd985b64119612b6df24f2d
3
  size 1465
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:683d1ed49f868c59cd7d1a26e5f958afa905456ce8fcd55424f0eacbba42da02
3
  size 1465
checkpoint-26/tokenizer.json CHANGED
@@ -2,13 +2,13 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 512,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
  "padding": {
10
  "strategy": {
11
- "Fixed": 512
12
  },
13
  "direction": "Right",
14
  "pad_to_multiple_of": null,
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 256,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
  "padding": {
10
  "strategy": {
11
+ "Fixed": 256
12
  },
13
  "direction": "Right",
14
  "pad_to_multiple_of": null,
checkpoint-26/trainer_state.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "best_global_step": 26,
3
- "best_metric": 0.06289733201265335,
4
- "best_model_checkpoint": "/content/drive/MyDrive/ADEGuard/ner_biobert_final/checkpoint-26",
5
  "epoch": 1.0,
6
  "eval_steps": 500,
7
  "global_step": 26,
@@ -11,20 +11,20 @@
11
  "log_history": [
12
  {
13
  "epoch": 1.0,
14
- "eval_f1": 0.0,
15
- "eval_loss": 0.06289733201265335,
16
- "eval_precision": 0.0,
17
- "eval_recall": 0.0,
18
- "eval_runtime": 0.7698,
19
- "eval_samples_per_second": 33.775,
20
- "eval_steps_per_second": 5.196,
21
  "step": 26
22
  }
23
  ],
24
  "logging_steps": 50,
25
- "max_steps": 78,
26
  "num_input_tokens_seen": 0,
27
- "num_train_epochs": 3,
28
  "save_steps": 500,
29
  "stateful_callbacks": {
30
  "TrainerControl": {
@@ -38,7 +38,7 @@
38
  "attributes": {}
39
  }
40
  },
41
- "total_flos": 27176091131904.0,
42
  "train_batch_size": 4,
43
  "trial_name": null,
44
  "trial_params": null
 
1
  {
2
  "best_global_step": 26,
3
+ "best_metric": 1.139850378036499,
4
+ "best_model_checkpoint": "/content/drive/MyDrive/ADEGuard/ner_biobert_final_v2/checkpoint-26",
5
  "epoch": 1.0,
6
  "eval_steps": 500,
7
  "global_step": 26,
 
11
  "log_history": [
12
  {
13
  "epoch": 1.0,
14
+ "eval_f1": 0.00783944810285356,
15
+ "eval_loss": 1.139850378036499,
16
+ "eval_precision": 0.0039488232506713,
17
+ "eval_recall": 0.5319148936170213,
18
+ "eval_runtime": 0.5321,
19
+ "eval_samples_per_second": 48.861,
20
+ "eval_steps_per_second": 7.517,
21
  "step": 26
22
  }
23
  ],
24
  "logging_steps": 50,
25
+ "max_steps": 156,
26
  "num_input_tokens_seen": 0,
27
+ "num_train_epochs": 6,
28
  "save_steps": 500,
29
  "stateful_callbacks": {
30
  "TrainerControl": {
 
38
  "attributes": {}
39
  }
40
  },
41
+ "total_flos": 13588045565952.0,
42
  "train_batch_size": 4,
43
  "trial_name": null,
44
  "trial_params": null
checkpoint-26/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b2e33e0687b2225b37b3b842824292cee58a877110d3e61b73275919a0a2f4b1
3
  size 5201
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28773adb83572bc94ef9e073a42eb194223d823b5b140ca2b7c77bf0c3f76e9e
3
  size 5201
checkpoint-52/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:06f3e92579bd3ee0f5b4a8dc8d5709068f2ef100851ffb0bba272814494b1170
3
  size 430923564
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7480b024fe7766fab782a3cba7cc8bdb38651840f3cb50826fba71ec204866cc
3
  size 430923564
checkpoint-52/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:356b5d063ccb0d1429b245b503f250811f8e547f4d06bc214165902cb83c3524
3
- size 861970123
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f989e375de6084bc9c7699d0246c4295c0bf24f72fcc0eaf11762b1193179cf
3
+ size 226911819
checkpoint-52/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c7e6c0e8e273c2f6b50af5caac50a1b7f627da4cf5222dd263a292a90c1d573e
3
  size 14645
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21889862d27281783a0fee3f059c1c137f41d59d3720580f58ad5e30dfafb855
3
  size 14645
checkpoint-52/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bea3c51af45d25a7261b4b0df286c324e8be04554f45df38e96d91bc84fdda87
3
  size 1465
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c8949a640b45a3a8cd2569d356a198f561c9b743be4c42b29713a6275a9af3c
3
  size 1465
checkpoint-52/tokenizer.json CHANGED
@@ -2,13 +2,13 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 512,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
  "padding": {
10
  "strategy": {
11
- "Fixed": 512
12
  },
13
  "direction": "Right",
14
  "pad_to_multiple_of": null,
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 256,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
  "padding": {
10
  "strategy": {
11
+ "Fixed": 256
12
  },
13
  "direction": "Right",
14
  "pad_to_multiple_of": null,
checkpoint-52/trainer_state.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "best_global_step": 52,
3
- "best_metric": 0.049389105290174484,
4
- "best_model_checkpoint": "/content/drive/MyDrive/ADEGuard/ner_biobert_final/checkpoint-52",
5
  "epoch": 2.0,
6
  "eval_steps": 500,
7
  "global_step": 52,
@@ -11,38 +11,38 @@
11
  "log_history": [
12
  {
13
  "epoch": 1.0,
14
- "eval_f1": 0.0,
15
- "eval_loss": 0.06289733201265335,
16
- "eval_precision": 0.0,
17
- "eval_recall": 0.0,
18
- "eval_runtime": 0.7698,
19
- "eval_samples_per_second": 33.775,
20
- "eval_steps_per_second": 5.196,
21
  "step": 26
22
  },
23
  {
24
  "epoch": 1.9230769230769231,
25
- "grad_norm": 0.09649301320314407,
26
- "learning_rate": 1.858974358974359e-05,
27
- "loss": 0.15085140228271485,
28
  "step": 50
29
  },
30
  {
31
  "epoch": 2.0,
32
- "eval_f1": 0.1904761904761905,
33
- "eval_loss": 0.049389105290174484,
34
- "eval_precision": 1.0,
35
- "eval_recall": 0.10526315789473684,
36
- "eval_runtime": 0.7878,
37
- "eval_samples_per_second": 33.002,
38
- "eval_steps_per_second": 5.077,
39
  "step": 52
40
  }
41
  ],
42
  "logging_steps": 50,
43
- "max_steps": 78,
44
  "num_input_tokens_seen": 0,
45
- "num_train_epochs": 3,
46
  "save_steps": 500,
47
  "stateful_callbacks": {
48
  "TrainerControl": {
@@ -56,7 +56,7 @@
56
  "attributes": {}
57
  }
58
  },
59
- "total_flos": 54352182263808.0,
60
  "train_batch_size": 4,
61
  "trial_name": null,
62
  "trial_params": null
 
1
  {
2
  "best_global_step": 52,
3
+ "best_metric": 0.6769160032272339,
4
+ "best_model_checkpoint": "/content/drive/MyDrive/ADEGuard/ner_biobert_final_v2/checkpoint-52",
5
  "epoch": 2.0,
6
  "eval_steps": 500,
7
  "global_step": 52,
 
11
  "log_history": [
12
  {
13
  "epoch": 1.0,
14
+ "eval_f1": 0.00783944810285356,
15
+ "eval_loss": 1.139850378036499,
16
+ "eval_precision": 0.0039488232506713,
17
+ "eval_recall": 0.5319148936170213,
18
+ "eval_runtime": 0.5321,
19
+ "eval_samples_per_second": 48.861,
20
+ "eval_steps_per_second": 7.517,
21
  "step": 26
22
  },
23
  {
24
  "epoch": 1.9230769230769231,
25
+ "grad_norm": 2.009291410446167,
26
+ "learning_rate": 2.057692307692308e-05,
27
+ "loss": 1.1273328399658202,
28
  "step": 50
29
  },
30
  {
31
  "epoch": 2.0,
32
+ "eval_f1": 0.009548896937767533,
33
+ "eval_loss": 0.6769160032272339,
34
+ "eval_precision": 0.004811680769868923,
35
+ "eval_recall": 0.6170212765957447,
36
+ "eval_runtime": 0.4451,
37
+ "eval_samples_per_second": 58.415,
38
+ "eval_steps_per_second": 8.987,
39
  "step": 52
40
  }
41
  ],
42
  "logging_steps": 50,
43
+ "max_steps": 156,
44
  "num_input_tokens_seen": 0,
45
+ "num_train_epochs": 6,
46
  "save_steps": 500,
47
  "stateful_callbacks": {
48
  "TrainerControl": {
 
56
  "attributes": {}
57
  }
58
  },
59
+ "total_flos": 27176091131904.0,
60
  "train_batch_size": 4,
61
  "trial_name": null,
62
  "trial_params": null
checkpoint-52/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b2e33e0687b2225b37b3b842824292cee58a877110d3e61b73275919a0a2f4b1
3
  size 5201
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28773adb83572bc94ef9e073a42eb194223d823b5b140ca2b7c77bf0c3f76e9e
3
  size 5201
checkpoint-78/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a4d7f02d1f3246386da560af52aee7ca7aecdb423d0707a4c5b6dd8eef7dabcc
3
  size 430923564
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f9533048d9749d8b17edc37db6af1286e31b0d00052508ec627fcde7989bb2a
3
  size 430923564
checkpoint-78/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:90c4a48b2641c9e4428ed7beef0d4df29041494b046b605b0b37dbe3fd833630
3
- size 861970123
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:760051714cd4e374b1407888a2c29a3cb5992e21d894bae2d20bf62290a78f1a
3
+ size 226911819
checkpoint-78/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:79ed8d45a4499bcffc47aab390786f5185fc3d7b706539fb18806d87c243fb0d
3
  size 14645
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96fc2fc16bde34718774552cb32c5672a6c84857f9839f9a70a7286c4f014f79
3
  size 14645
checkpoint-78/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1bf819bf62e0d94d7d30d87fe36e9317a4be1fad217925e872ff99a8214e2e9d
3
  size 1465
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eca5ca9fe276b1820e6b746310ee88fa238e1e01be76fba2edc08e699cf943f9
3
  size 1465
checkpoint-78/tokenizer.json CHANGED
@@ -2,13 +2,13 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 512,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
  "padding": {
10
  "strategy": {
11
- "Fixed": 512
12
  },
13
  "direction": "Right",
14
  "pad_to_multiple_of": null,
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 256,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
  "padding": {
10
  "strategy": {
11
+ "Fixed": 256
12
  },
13
  "direction": "Right",
14
  "pad_to_multiple_of": null,
checkpoint-78/trainer_state.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "best_global_step": 78,
3
- "best_metric": 0.04243722930550575,
4
- "best_model_checkpoint": "/content/drive/MyDrive/ADEGuard/ner_biobert_final/checkpoint-78",
5
  "epoch": 3.0,
6
  "eval_steps": 500,
7
  "global_step": 78,
@@ -11,49 +11,49 @@
11
  "log_history": [
12
  {
13
  "epoch": 1.0,
14
- "eval_f1": 0.0,
15
- "eval_loss": 0.06289733201265335,
16
- "eval_precision": 0.0,
17
- "eval_recall": 0.0,
18
- "eval_runtime": 0.7698,
19
- "eval_samples_per_second": 33.775,
20
- "eval_steps_per_second": 5.196,
21
  "step": 26
22
  },
23
  {
24
  "epoch": 1.9230769230769231,
25
- "grad_norm": 0.09649301320314407,
26
- "learning_rate": 1.858974358974359e-05,
27
- "loss": 0.15085140228271485,
28
  "step": 50
29
  },
30
  {
31
  "epoch": 2.0,
32
- "eval_f1": 0.1904761904761905,
33
- "eval_loss": 0.049389105290174484,
34
- "eval_precision": 1.0,
35
- "eval_recall": 0.10526315789473684,
36
- "eval_runtime": 0.7878,
37
- "eval_samples_per_second": 33.002,
38
- "eval_steps_per_second": 5.077,
39
  "step": 52
40
  },
41
  {
42
  "epoch": 3.0,
43
- "eval_f1": 0.1904761904761905,
44
- "eval_loss": 0.04243722930550575,
45
- "eval_precision": 1.0,
46
- "eval_recall": 0.10526315789473684,
47
- "eval_runtime": 0.8159,
48
- "eval_samples_per_second": 31.867,
49
- "eval_steps_per_second": 4.903,
50
  "step": 78
51
  }
52
  ],
53
  "logging_steps": 50,
54
- "max_steps": 78,
55
  "num_input_tokens_seen": 0,
56
- "num_train_epochs": 3,
57
  "save_steps": 500,
58
  "stateful_callbacks": {
59
  "TrainerControl": {
@@ -62,12 +62,12 @@
62
  "should_evaluate": false,
63
  "should_log": false,
64
  "should_save": true,
65
- "should_training_stop": true
66
  },
67
  "attributes": {}
68
  }
69
  },
70
- "total_flos": 81528273395712.0,
71
  "train_batch_size": 4,
72
  "trial_name": null,
73
  "trial_params": null
 
1
  {
2
  "best_global_step": 78,
3
+ "best_metric": 0.5276753902435303,
4
+ "best_model_checkpoint": "/content/drive/MyDrive/ADEGuard/ner_biobert_final_v2/checkpoint-78",
5
  "epoch": 3.0,
6
  "eval_steps": 500,
7
  "global_step": 78,
 
11
  "log_history": [
12
  {
13
  "epoch": 1.0,
14
+ "eval_f1": 0.00783944810285356,
15
+ "eval_loss": 1.139850378036499,
16
+ "eval_precision": 0.0039488232506713,
17
+ "eval_recall": 0.5319148936170213,
18
+ "eval_runtime": 0.5321,
19
+ "eval_samples_per_second": 48.861,
20
+ "eval_steps_per_second": 7.517,
21
  "step": 26
22
  },
23
  {
24
  "epoch": 1.9230769230769231,
25
+ "grad_norm": 2.009291410446167,
26
+ "learning_rate": 2.057692307692308e-05,
27
+ "loss": 1.1273328399658202,
28
  "step": 50
29
  },
30
  {
31
  "epoch": 2.0,
32
+ "eval_f1": 0.009548896937767533,
33
+ "eval_loss": 0.6769160032272339,
34
+ "eval_precision": 0.004811680769868923,
35
+ "eval_recall": 0.6170212765957447,
36
+ "eval_runtime": 0.4451,
37
+ "eval_samples_per_second": 58.415,
38
+ "eval_steps_per_second": 8.987,
39
  "step": 52
40
  },
41
  {
42
  "epoch": 3.0,
43
+ "eval_f1": 0.011583871686344399,
44
+ "eval_loss": 0.5276753902435303,
45
+ "eval_precision": 0.00585321927059883,
46
+ "eval_recall": 0.5531914893617021,
47
+ "eval_runtime": 0.5507,
48
+ "eval_samples_per_second": 47.215,
49
+ "eval_steps_per_second": 7.264,
50
  "step": 78
51
  }
52
  ],
53
  "logging_steps": 50,
54
+ "max_steps": 156,
55
  "num_input_tokens_seen": 0,
56
+ "num_train_epochs": 6,
57
  "save_steps": 500,
58
  "stateful_callbacks": {
59
  "TrainerControl": {
 
62
  "should_evaluate": false,
63
  "should_log": false,
64
  "should_save": true,
65
+ "should_training_stop": false
66
  },
67
  "attributes": {}
68
  }
69
  },
70
+ "total_flos": 40764136697856.0,
71
  "train_batch_size": 4,
72
  "trial_name": null,
73
  "trial_params": null
checkpoint-78/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b2e33e0687b2225b37b3b842824292cee58a877110d3e61b73275919a0a2f4b1
3
  size 5201
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28773adb83572bc94ef9e073a42eb194223d823b5b140ca2b7c77bf0c3f76e9e
3
  size 5201
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a4d7f02d1f3246386da560af52aee7ca7aecdb423d0707a4c5b6dd8eef7dabcc
3
  size 430923564
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3134d4e9b31847c551b1b2be4101851ac51d0b57688993699ea34cdf1f6f5560
3
  size 430923564