Farouk commited on
Commit
a4178a3
·
1 Parent(s): 4dbd620

Training in progress, step 200

Browse files
adapter_config.json CHANGED
@@ -14,10 +14,10 @@
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
17
- "o_proj",
18
  "q_proj",
 
19
  "down_proj",
20
- "v_proj",
21
  "k_proj",
22
  "gate_proj",
23
  "up_proj"
 
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
17
+ "v_proj",
18
  "q_proj",
19
+ "o_proj",
20
  "down_proj",
 
21
  "k_proj",
22
  "gate_proj",
23
  "up_proj"
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:efdc2a905cceb1a9c3bfe194c1b9885e9dfdbbe4afd906a34a44a8c1628ee9e5
3
  size 319977229
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2b10da90a336f67a541e93e611dc8cf785e6594bf22c830698a608b4695e8f1
3
  size 319977229
checkpoint-200/README.md CHANGED
@@ -4,6 +4,17 @@ library_name: peft
4
  ## Training procedure
5
 
6
 
 
 
 
 
 
 
 
 
 
 
 
7
  The following `bitsandbytes` quantization config was used during training:
8
  - load_in_8bit: False
9
  - load_in_4bit: True
@@ -16,5 +27,6 @@ The following `bitsandbytes` quantization config was used during training:
16
  - bnb_4bit_compute_dtype: bfloat16
17
  ### Framework versions
18
 
 
19
 
20
  - PEFT 0.4.0
 
4
  ## Training procedure
5
 
6
 
7
+ The following `bitsandbytes` quantization config was used during training:
8
+ - load_in_8bit: False
9
+ - load_in_4bit: True
10
+ - llm_int8_threshold: 6.0
11
+ - llm_int8_skip_modules: None
12
+ - llm_int8_enable_fp32_cpu_offload: False
13
+ - llm_int8_has_fp16_weight: False
14
+ - bnb_4bit_quant_type: nf4
15
+ - bnb_4bit_use_double_quant: True
16
+ - bnb_4bit_compute_dtype: bfloat16
17
+
18
  The following `bitsandbytes` quantization config was used during training:
19
  - load_in_8bit: False
20
  - load_in_4bit: True
 
27
  - bnb_4bit_compute_dtype: bfloat16
28
  ### Framework versions
29
 
30
+ - PEFT 0.4.0
31
 
32
  - PEFT 0.4.0
checkpoint-200/adapter_config.json CHANGED
@@ -14,10 +14,10 @@
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
17
- "o_proj",
18
  "q_proj",
 
19
  "down_proj",
20
- "v_proj",
21
  "k_proj",
22
  "gate_proj",
23
  "up_proj"
 
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
17
+ "v_proj",
18
  "q_proj",
19
+ "o_proj",
20
  "down_proj",
 
21
  "k_proj",
22
  "gate_proj",
23
  "up_proj"
checkpoint-200/adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:951f0410a5b1b895cd7f4ac9717b56d77a66ac1e7f064443ac97a17ac881c0c0
3
  size 319977229
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2b10da90a336f67a541e93e611dc8cf785e6594bf22c830698a608b4695e8f1
3
  size 319977229
checkpoint-200/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4e14849e6ad46b1dc5832e2fd13fe921afb7ac60c46b23a430a6801ffc5761ea
3
  size 1279539525
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b9d08d7292e7785801f823e4c80ca66bf9349510885d5641c81dd2c8402310f
3
  size 1279539525
checkpoint-200/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d440b345926143ca36c3e807d6f422c49a3f1e63c7f15c261f9777fbc7a2bbc4
3
  size 14511
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:494626a050cceac1c0023f61e6deff68d6953b3cd5f8b62db8ab0c97dd59334e
3
  size 14511
checkpoint-200/trainer_state.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "best_metric": 0.9700731039047241,
3
  "best_model_checkpoint": "experts/expert-14/checkpoint-200",
4
  "epoch": 0.06151480199923107,
5
  "global_step": 200,
@@ -10,13 +10,13 @@
10
  {
11
  "epoch": 0.0,
12
  "learning_rate": 0.0002,
13
- "loss": 1.0634,
14
  "step": 10
15
  },
16
  {
17
  "epoch": 0.01,
18
  "learning_rate": 0.0002,
19
- "loss": 1.0357,
20
  "step": 20
21
  },
22
  {
@@ -28,43 +28,43 @@
28
  {
29
  "epoch": 0.01,
30
  "learning_rate": 0.0002,
31
- "loss": 0.9858,
32
  "step": 40
33
  },
34
  {
35
  "epoch": 0.02,
36
  "learning_rate": 0.0002,
37
- "loss": 1.0149,
38
  "step": 50
39
  },
40
  {
41
  "epoch": 0.02,
42
  "learning_rate": 0.0002,
43
- "loss": 1.0292,
44
  "step": 60
45
  },
46
  {
47
  "epoch": 0.02,
48
  "learning_rate": 0.0002,
49
- "loss": 0.9947,
50
  "step": 70
51
  },
52
  {
53
  "epoch": 0.02,
54
  "learning_rate": 0.0002,
55
- "loss": 1.0068,
56
  "step": 80
57
  },
58
  {
59
  "epoch": 0.03,
60
  "learning_rate": 0.0002,
61
- "loss": 1.015,
62
  "step": 90
63
  },
64
  {
65
  "epoch": 0.03,
66
  "learning_rate": 0.0002,
67
- "loss": 1.0055,
68
  "step": 100
69
  },
70
  {
@@ -76,37 +76,37 @@
76
  {
77
  "epoch": 0.04,
78
  "learning_rate": 0.0002,
79
- "loss": 1.0218,
80
  "step": 120
81
  },
82
  {
83
  "epoch": 0.04,
84
  "learning_rate": 0.0002,
85
- "loss": 0.9645,
86
  "step": 130
87
  },
88
  {
89
  "epoch": 0.04,
90
  "learning_rate": 0.0002,
91
- "loss": 1.0059,
92
  "step": 140
93
  },
94
  {
95
  "epoch": 0.05,
96
  "learning_rate": 0.0002,
97
- "loss": 1.0041,
98
  "step": 150
99
  },
100
  {
101
  "epoch": 0.05,
102
  "learning_rate": 0.0002,
103
- "loss": 0.9626,
104
  "step": 160
105
  },
106
  {
107
  "epoch": 0.05,
108
  "learning_rate": 0.0002,
109
- "loss": 0.9682,
110
  "step": 170
111
  },
112
  {
@@ -118,74 +118,74 @@
118
  {
119
  "epoch": 0.06,
120
  "learning_rate": 0.0002,
121
- "loss": 1.0029,
122
  "step": 190
123
  },
124
  {
125
  "epoch": 0.06,
126
  "learning_rate": 0.0002,
127
- "loss": 0.9873,
128
  "step": 200
129
  },
130
  {
131
  "epoch": 0.06,
132
- "eval_loss": 0.9700731039047241,
133
- "eval_runtime": 90.2222,
134
- "eval_samples_per_second": 11.084,
135
- "eval_steps_per_second": 5.542,
136
  "step": 200
137
  },
138
  {
139
  "epoch": 0.06,
140
- "mmlu_eval_accuracy": 0.4896808303512203,
141
  "mmlu_eval_accuracy_abstract_algebra": 0.2727272727272727,
142
- "mmlu_eval_accuracy_anatomy": 0.6428571428571429,
143
  "mmlu_eval_accuracy_astronomy": 0.4375,
144
  "mmlu_eval_accuracy_business_ethics": 0.5454545454545454,
145
- "mmlu_eval_accuracy_clinical_knowledge": 0.5172413793103449,
146
  "mmlu_eval_accuracy_college_biology": 0.3125,
147
  "mmlu_eval_accuracy_college_chemistry": 0.25,
148
  "mmlu_eval_accuracy_college_computer_science": 0.36363636363636365,
149
  "mmlu_eval_accuracy_college_mathematics": 0.18181818181818182,
150
  "mmlu_eval_accuracy_college_medicine": 0.45454545454545453,
151
  "mmlu_eval_accuracy_college_physics": 0.45454545454545453,
152
- "mmlu_eval_accuracy_computer_security": 0.45454545454545453,
153
  "mmlu_eval_accuracy_conceptual_physics": 0.38461538461538464,
154
  "mmlu_eval_accuracy_econometrics": 0.25,
155
  "mmlu_eval_accuracy_electrical_engineering": 0.3125,
156
  "mmlu_eval_accuracy_elementary_mathematics": 0.3902439024390244,
157
- "mmlu_eval_accuracy_formal_logic": 0.14285714285714285,
158
- "mmlu_eval_accuracy_global_facts": 0.6,
159
- "mmlu_eval_accuracy_high_school_biology": 0.40625,
160
  "mmlu_eval_accuracy_high_school_chemistry": 0.22727272727272727,
161
  "mmlu_eval_accuracy_high_school_computer_science": 0.5555555555555556,
162
  "mmlu_eval_accuracy_high_school_european_history": 0.6111111111111112,
163
  "mmlu_eval_accuracy_high_school_geography": 0.9090909090909091,
164
  "mmlu_eval_accuracy_high_school_government_and_politics": 0.5238095238095238,
165
  "mmlu_eval_accuracy_high_school_macroeconomics": 0.46511627906976744,
166
- "mmlu_eval_accuracy_high_school_mathematics": 0.2413793103448276,
167
  "mmlu_eval_accuracy_high_school_microeconomics": 0.46153846153846156,
168
  "mmlu_eval_accuracy_high_school_physics": 0.17647058823529413,
169
- "mmlu_eval_accuracy_high_school_psychology": 0.85,
170
- "mmlu_eval_accuracy_high_school_statistics": 0.43478260869565216,
171
- "mmlu_eval_accuracy_high_school_us_history": 0.6363636363636364,
172
- "mmlu_eval_accuracy_high_school_world_history": 0.6538461538461539,
173
- "mmlu_eval_accuracy_human_aging": 0.6521739130434783,
174
  "mmlu_eval_accuracy_human_sexuality": 0.3333333333333333,
175
  "mmlu_eval_accuracy_international_law": 0.8461538461538461,
176
  "mmlu_eval_accuracy_jurisprudence": 0.36363636363636365,
177
  "mmlu_eval_accuracy_logical_fallacies": 0.6666666666666666,
178
  "mmlu_eval_accuracy_machine_learning": 0.18181818181818182,
179
  "mmlu_eval_accuracy_management": 0.6363636363636364,
180
- "mmlu_eval_accuracy_marketing": 0.88,
181
  "mmlu_eval_accuracy_medical_genetics": 0.9090909090909091,
182
  "mmlu_eval_accuracy_miscellaneous": 0.6511627906976745,
183
- "mmlu_eval_accuracy_moral_disputes": 0.39473684210526316,
184
- "mmlu_eval_accuracy_moral_scenarios": 0.23,
185
- "mmlu_eval_accuracy_nutrition": 0.696969696969697,
186
- "mmlu_eval_accuracy_philosophy": 0.4411764705882353,
187
  "mmlu_eval_accuracy_prehistory": 0.5142857142857142,
188
- "mmlu_eval_accuracy_professional_accounting": 0.3225806451612903,
189
  "mmlu_eval_accuracy_professional_law": 0.3411764705882353,
190
  "mmlu_eval_accuracy_professional_medicine": 0.5806451612903226,
191
  "mmlu_eval_accuracy_professional_psychology": 0.463768115942029,
@@ -195,7 +195,7 @@
195
  "mmlu_eval_accuracy_us_foreign_policy": 0.7272727272727273,
196
  "mmlu_eval_accuracy_virology": 0.4444444444444444,
197
  "mmlu_eval_accuracy_world_religions": 0.6842105263157895,
198
- "mmlu_loss": 1.6262522289242507,
199
  "step": 200
200
  }
201
  ],
 
1
  {
2
+ "best_metric": 0.970292329788208,
3
  "best_model_checkpoint": "experts/expert-14/checkpoint-200",
4
  "epoch": 0.06151480199923107,
5
  "global_step": 200,
 
10
  {
11
  "epoch": 0.0,
12
  "learning_rate": 0.0002,
13
+ "loss": 1.064,
14
  "step": 10
15
  },
16
  {
17
  "epoch": 0.01,
18
  "learning_rate": 0.0002,
19
+ "loss": 1.0358,
20
  "step": 20
21
  },
22
  {
 
28
  {
29
  "epoch": 0.01,
30
  "learning_rate": 0.0002,
31
+ "loss": 0.9861,
32
  "step": 40
33
  },
34
  {
35
  "epoch": 0.02,
36
  "learning_rate": 0.0002,
37
+ "loss": 1.0148,
38
  "step": 50
39
  },
40
  {
41
  "epoch": 0.02,
42
  "learning_rate": 0.0002,
43
+ "loss": 1.0293,
44
  "step": 60
45
  },
46
  {
47
  "epoch": 0.02,
48
  "learning_rate": 0.0002,
49
+ "loss": 0.9945,
50
  "step": 70
51
  },
52
  {
53
  "epoch": 0.02,
54
  "learning_rate": 0.0002,
55
+ "loss": 1.0069,
56
  "step": 80
57
  },
58
  {
59
  "epoch": 0.03,
60
  "learning_rate": 0.0002,
61
+ "loss": 1.0147,
62
  "step": 90
63
  },
64
  {
65
  "epoch": 0.03,
66
  "learning_rate": 0.0002,
67
+ "loss": 1.0056,
68
  "step": 100
69
  },
70
  {
 
76
  {
77
  "epoch": 0.04,
78
  "learning_rate": 0.0002,
79
+ "loss": 1.0214,
80
  "step": 120
81
  },
82
  {
83
  "epoch": 0.04,
84
  "learning_rate": 0.0002,
85
+ "loss": 0.9642,
86
  "step": 130
87
  },
88
  {
89
  "epoch": 0.04,
90
  "learning_rate": 0.0002,
91
+ "loss": 1.0051,
92
  "step": 140
93
  },
94
  {
95
  "epoch": 0.05,
96
  "learning_rate": 0.0002,
97
+ "loss": 1.004,
98
  "step": 150
99
  },
100
  {
101
  "epoch": 0.05,
102
  "learning_rate": 0.0002,
103
+ "loss": 0.9631,
104
  "step": 160
105
  },
106
  {
107
  "epoch": 0.05,
108
  "learning_rate": 0.0002,
109
+ "loss": 0.9674,
110
  "step": 170
111
  },
112
  {
 
118
  {
119
  "epoch": 0.06,
120
  "learning_rate": 0.0002,
121
+ "loss": 1.0023,
122
  "step": 190
123
  },
124
  {
125
  "epoch": 0.06,
126
  "learning_rate": 0.0002,
127
+ "loss": 0.9879,
128
  "step": 200
129
  },
130
  {
131
  "epoch": 0.06,
132
+ "eval_loss": 0.970292329788208,
133
+ "eval_runtime": 67.4255,
134
+ "eval_samples_per_second": 14.831,
135
+ "eval_steps_per_second": 7.416,
136
  "step": 200
137
  },
138
  {
139
  "epoch": 0.06,
140
+ "mmlu_eval_accuracy": 0.49145352624106475,
141
  "mmlu_eval_accuracy_abstract_algebra": 0.2727272727272727,
142
+ "mmlu_eval_accuracy_anatomy": 0.5714285714285714,
143
  "mmlu_eval_accuracy_astronomy": 0.4375,
144
  "mmlu_eval_accuracy_business_ethics": 0.5454545454545454,
145
+ "mmlu_eval_accuracy_clinical_knowledge": 0.5517241379310345,
146
  "mmlu_eval_accuracy_college_biology": 0.3125,
147
  "mmlu_eval_accuracy_college_chemistry": 0.25,
148
  "mmlu_eval_accuracy_college_computer_science": 0.36363636363636365,
149
  "mmlu_eval_accuracy_college_mathematics": 0.18181818181818182,
150
  "mmlu_eval_accuracy_college_medicine": 0.45454545454545453,
151
  "mmlu_eval_accuracy_college_physics": 0.45454545454545453,
152
+ "mmlu_eval_accuracy_computer_security": 0.6363636363636364,
153
  "mmlu_eval_accuracy_conceptual_physics": 0.38461538461538464,
154
  "mmlu_eval_accuracy_econometrics": 0.25,
155
  "mmlu_eval_accuracy_electrical_engineering": 0.3125,
156
  "mmlu_eval_accuracy_elementary_mathematics": 0.3902439024390244,
157
+ "mmlu_eval_accuracy_formal_logic": 0.21428571428571427,
158
+ "mmlu_eval_accuracy_global_facts": 0.5,
159
+ "mmlu_eval_accuracy_high_school_biology": 0.4375,
160
  "mmlu_eval_accuracy_high_school_chemistry": 0.22727272727272727,
161
  "mmlu_eval_accuracy_high_school_computer_science": 0.5555555555555556,
162
  "mmlu_eval_accuracy_high_school_european_history": 0.6111111111111112,
163
  "mmlu_eval_accuracy_high_school_geography": 0.9090909090909091,
164
  "mmlu_eval_accuracy_high_school_government_and_politics": 0.5238095238095238,
165
  "mmlu_eval_accuracy_high_school_macroeconomics": 0.46511627906976744,
166
+ "mmlu_eval_accuracy_high_school_mathematics": 0.20689655172413793,
167
  "mmlu_eval_accuracy_high_school_microeconomics": 0.46153846153846156,
168
  "mmlu_eval_accuracy_high_school_physics": 0.17647058823529413,
169
+ "mmlu_eval_accuracy_high_school_psychology": 0.8333333333333334,
170
+ "mmlu_eval_accuracy_high_school_statistics": 0.391304347826087,
171
+ "mmlu_eval_accuracy_high_school_us_history": 0.6818181818181818,
172
+ "mmlu_eval_accuracy_high_school_world_history": 0.6923076923076923,
173
+ "mmlu_eval_accuracy_human_aging": 0.6086956521739131,
174
  "mmlu_eval_accuracy_human_sexuality": 0.3333333333333333,
175
  "mmlu_eval_accuracy_international_law": 0.8461538461538461,
176
  "mmlu_eval_accuracy_jurisprudence": 0.36363636363636365,
177
  "mmlu_eval_accuracy_logical_fallacies": 0.6666666666666666,
178
  "mmlu_eval_accuracy_machine_learning": 0.18181818181818182,
179
  "mmlu_eval_accuracy_management": 0.6363636363636364,
180
+ "mmlu_eval_accuracy_marketing": 0.84,
181
  "mmlu_eval_accuracy_medical_genetics": 0.9090909090909091,
182
  "mmlu_eval_accuracy_miscellaneous": 0.6511627906976745,
183
+ "mmlu_eval_accuracy_moral_disputes": 0.42105263157894735,
184
+ "mmlu_eval_accuracy_moral_scenarios": 0.22,
185
+ "mmlu_eval_accuracy_nutrition": 0.6666666666666666,
186
+ "mmlu_eval_accuracy_philosophy": 0.47058823529411764,
187
  "mmlu_eval_accuracy_prehistory": 0.5142857142857142,
188
+ "mmlu_eval_accuracy_professional_accounting": 0.3548387096774194,
189
  "mmlu_eval_accuracy_professional_law": 0.3411764705882353,
190
  "mmlu_eval_accuracy_professional_medicine": 0.5806451612903226,
191
  "mmlu_eval_accuracy_professional_psychology": 0.463768115942029,
 
195
  "mmlu_eval_accuracy_us_foreign_policy": 0.7272727272727273,
196
  "mmlu_eval_accuracy_virology": 0.4444444444444444,
197
  "mmlu_eval_accuracy_world_religions": 0.6842105263157895,
198
+ "mmlu_loss": 1.6139244634862357,
199
  "step": 200
200
  }
201
  ],
checkpoint-200/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:132dc0489795dd2e62e3dfed60f8fa6c2da9f513c3d6724e00171983c97e1c5b
3
  size 5819
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c1f3b4e86a18f376afc61102db182028a2b670075c0fd60d0651751e2c0b01a
3
  size 5819
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:132dc0489795dd2e62e3dfed60f8fa6c2da9f513c3d6724e00171983c97e1c5b
3
  size 5819
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c1f3b4e86a18f376afc61102db182028a2b670075c0fd60d0651751e2c0b01a
3
  size 5819