Farouk commited on
Commit
84fe457
·
1 Parent(s): 4eb0664

Training in progress, step 200

Browse files
adapter_config.json CHANGED
@@ -14,13 +14,13 @@
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
17
- "gate_proj",
18
- "v_proj",
19
  "down_proj",
 
20
  "o_proj",
21
  "up_proj",
22
- "q_proj",
23
- "k_proj"
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }
 
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
17
+ "k_proj",
18
+ "q_proj",
19
  "down_proj",
20
+ "v_proj",
21
  "o_proj",
22
  "up_proj",
23
+ "gate_proj"
 
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7f1f11726260e5f15f3bfb21f6572f487cb0ac87d4816d145899ab0a4826eebe
3
  size 319977229
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2920cd03151b12944b05cac37aae55df3d589cf8838396f445986adf54625c8
3
  size 319977229
checkpoint-200/README.md CHANGED
@@ -4,6 +4,17 @@ library_name: peft
4
  ## Training procedure
5
 
6
 
 
 
 
 
 
 
 
 
 
 
 
7
  The following `bitsandbytes` quantization config was used during training:
8
  - load_in_8bit: False
9
  - load_in_4bit: True
@@ -16,5 +27,6 @@ The following `bitsandbytes` quantization config was used during training:
16
  - bnb_4bit_compute_dtype: bfloat16
17
  ### Framework versions
18
 
 
19
 
20
  - PEFT 0.4.0
 
4
  ## Training procedure
5
 
6
 
7
+ The following `bitsandbytes` quantization config was used during training:
8
+ - load_in_8bit: False
9
+ - load_in_4bit: True
10
+ - llm_int8_threshold: 6.0
11
+ - llm_int8_skip_modules: None
12
+ - llm_int8_enable_fp32_cpu_offload: False
13
+ - llm_int8_has_fp16_weight: False
14
+ - bnb_4bit_quant_type: nf4
15
+ - bnb_4bit_use_double_quant: True
16
+ - bnb_4bit_compute_dtype: bfloat16
17
+
18
  The following `bitsandbytes` quantization config was used during training:
19
  - load_in_8bit: False
20
  - load_in_4bit: True
 
27
  - bnb_4bit_compute_dtype: bfloat16
28
  ### Framework versions
29
 
30
+ - PEFT 0.4.0
31
 
32
  - PEFT 0.4.0
checkpoint-200/adapter_config.json CHANGED
@@ -14,13 +14,13 @@
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
17
- "gate_proj",
18
- "v_proj",
19
  "down_proj",
 
20
  "o_proj",
21
  "up_proj",
22
- "q_proj",
23
- "k_proj"
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }
 
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
17
+ "k_proj",
18
+ "q_proj",
19
  "down_proj",
20
+ "v_proj",
21
  "o_proj",
22
  "up_proj",
23
+ "gate_proj"
 
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }
checkpoint-200/adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:263033e4bde69e7abe4e2bb79382c8e341df3d121300b2764dd67d5250bcdf20
3
  size 319977229
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2920cd03151b12944b05cac37aae55df3d589cf8838396f445986adf54625c8
3
  size 319977229
checkpoint-200/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:53621764c0cc9575b8625256911dfdc45779bd8b4271ceace007253b31ec091e
3
  size 1279539525
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d534063120b4b91a06cca5cc95136916b072bee232394f473327d709028c14e
3
  size 1279539525
checkpoint-200/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:309dcbb280d5f4bbe84f8d52da5e6adaee03fc403cd67b4585737b3a40ebb902
3
  size 14511
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b846af86a019f5bd4a9ebc394631e9d7509ac5926fa4e13bc1ffe43eed596ec
3
  size 14511
checkpoint-200/trainer_state.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "best_metric": 0.8084444403648376,
3
  "best_model_checkpoint": "experts/expert-13/checkpoint-200",
4
  "epoch": 0.042034468263976464,
5
  "global_step": 200,
@@ -10,85 +10,85 @@
10
  {
11
  "epoch": 0.0,
12
  "learning_rate": 0.0002,
13
- "loss": 0.9192,
14
  "step": 10
15
  },
16
  {
17
  "epoch": 0.0,
18
  "learning_rate": 0.0002,
19
- "loss": 0.8713,
20
  "step": 20
21
  },
22
  {
23
  "epoch": 0.01,
24
  "learning_rate": 0.0002,
25
- "loss": 0.859,
26
  "step": 30
27
  },
28
  {
29
  "epoch": 0.01,
30
  "learning_rate": 0.0002,
31
- "loss": 0.8968,
32
  "step": 40
33
  },
34
  {
35
  "epoch": 0.01,
36
  "learning_rate": 0.0002,
37
- "loss": 0.8191,
38
  "step": 50
39
  },
40
  {
41
  "epoch": 0.01,
42
  "learning_rate": 0.0002,
43
- "loss": 0.8115,
44
  "step": 60
45
  },
46
  {
47
  "epoch": 0.01,
48
  "learning_rate": 0.0002,
49
- "loss": 0.7799,
50
  "step": 70
51
  },
52
  {
53
  "epoch": 0.02,
54
  "learning_rate": 0.0002,
55
- "loss": 0.8676,
56
  "step": 80
57
  },
58
  {
59
  "epoch": 0.02,
60
  "learning_rate": 0.0002,
61
- "loss": 0.887,
62
  "step": 90
63
  },
64
  {
65
  "epoch": 0.02,
66
  "learning_rate": 0.0002,
67
- "loss": 0.8152,
68
  "step": 100
69
  },
70
  {
71
  "epoch": 0.02,
72
  "learning_rate": 0.0002,
73
- "loss": 0.8251,
74
  "step": 110
75
  },
76
  {
77
  "epoch": 0.03,
78
  "learning_rate": 0.0002,
79
- "loss": 0.8274,
80
  "step": 120
81
  },
82
  {
83
  "epoch": 0.03,
84
  "learning_rate": 0.0002,
85
- "loss": 0.8151,
86
  "step": 130
87
  },
88
  {
89
  "epoch": 0.03,
90
  "learning_rate": 0.0002,
91
- "loss": 0.8547,
92
  "step": 140
93
  },
94
  {
@@ -100,7 +100,7 @@
100
  {
101
  "epoch": 0.03,
102
  "learning_rate": 0.0002,
103
- "loss": 0.8747,
104
  "step": 160
105
  },
106
  {
@@ -112,37 +112,37 @@
112
  {
113
  "epoch": 0.04,
114
  "learning_rate": 0.0002,
115
- "loss": 0.7856,
116
  "step": 180
117
  },
118
  {
119
  "epoch": 0.04,
120
  "learning_rate": 0.0002,
121
- "loss": 0.839,
122
  "step": 190
123
  },
124
  {
125
  "epoch": 0.04,
126
  "learning_rate": 0.0002,
127
- "loss": 0.7512,
128
  "step": 200
129
  },
130
  {
131
  "epoch": 0.04,
132
- "eval_loss": 0.8084444403648376,
133
- "eval_runtime": 91.3593,
134
- "eval_samples_per_second": 10.946,
135
- "eval_steps_per_second": 5.473,
136
  "step": 200
137
  },
138
  {
139
  "epoch": 0.04,
140
- "mmlu_eval_accuracy": 0.4842637889631141,
141
  "mmlu_eval_accuracy_abstract_algebra": 0.2727272727272727,
142
- "mmlu_eval_accuracy_anatomy": 0.7142857142857143,
143
  "mmlu_eval_accuracy_astronomy": 0.4375,
144
- "mmlu_eval_accuracy_business_ethics": 0.5454545454545454,
145
- "mmlu_eval_accuracy_clinical_knowledge": 0.4827586206896552,
146
  "mmlu_eval_accuracy_college_biology": 0.4375,
147
  "mmlu_eval_accuracy_college_chemistry": 0.25,
148
  "mmlu_eval_accuracy_college_computer_science": 0.36363636363636365,
@@ -153,49 +153,49 @@
153
  "mmlu_eval_accuracy_conceptual_physics": 0.34615384615384615,
154
  "mmlu_eval_accuracy_econometrics": 0.25,
155
  "mmlu_eval_accuracy_electrical_engineering": 0.25,
156
- "mmlu_eval_accuracy_elementary_mathematics": 0.34146341463414637,
157
  "mmlu_eval_accuracy_formal_logic": 0.07142857142857142,
158
  "mmlu_eval_accuracy_global_facts": 0.4,
159
  "mmlu_eval_accuracy_high_school_biology": 0.375,
160
  "mmlu_eval_accuracy_high_school_chemistry": 0.2727272727272727,
161
  "mmlu_eval_accuracy_high_school_computer_science": 0.5555555555555556,
162
- "mmlu_eval_accuracy_high_school_european_history": 0.6111111111111112,
163
  "mmlu_eval_accuracy_high_school_geography": 0.9090909090909091,
164
- "mmlu_eval_accuracy_high_school_government_and_politics": 0.5714285714285714,
165
- "mmlu_eval_accuracy_high_school_macroeconomics": 0.4883720930232558,
166
  "mmlu_eval_accuracy_high_school_mathematics": 0.20689655172413793,
167
  "mmlu_eval_accuracy_high_school_microeconomics": 0.4230769230769231,
168
  "mmlu_eval_accuracy_high_school_physics": 0.11764705882352941,
169
  "mmlu_eval_accuracy_high_school_psychology": 0.85,
170
  "mmlu_eval_accuracy_high_school_statistics": 0.30434782608695654,
171
- "mmlu_eval_accuracy_high_school_us_history": 0.5909090909090909,
172
  "mmlu_eval_accuracy_high_school_world_history": 0.7307692307692307,
173
  "mmlu_eval_accuracy_human_aging": 0.6956521739130435,
174
  "mmlu_eval_accuracy_human_sexuality": 0.3333333333333333,
175
  "mmlu_eval_accuracy_international_law": 0.9230769230769231,
176
- "mmlu_eval_accuracy_jurisprudence": 0.5454545454545454,
177
  "mmlu_eval_accuracy_logical_fallacies": 0.6666666666666666,
178
- "mmlu_eval_accuracy_machine_learning": 0.18181818181818182,
179
  "mmlu_eval_accuracy_management": 0.6363636363636364,
180
- "mmlu_eval_accuracy_marketing": 0.92,
181
  "mmlu_eval_accuracy_medical_genetics": 0.9090909090909091,
182
- "mmlu_eval_accuracy_miscellaneous": 0.6976744186046512,
183
  "mmlu_eval_accuracy_moral_disputes": 0.4473684210526316,
184
- "mmlu_eval_accuracy_moral_scenarios": 0.23,
185
  "mmlu_eval_accuracy_nutrition": 0.6060606060606061,
186
  "mmlu_eval_accuracy_philosophy": 0.4411764705882353,
187
  "mmlu_eval_accuracy_prehistory": 0.45714285714285713,
188
- "mmlu_eval_accuracy_professional_accounting": 0.3870967741935484,
189
- "mmlu_eval_accuracy_professional_law": 0.3352941176470588,
190
  "mmlu_eval_accuracy_professional_medicine": 0.6129032258064516,
191
- "mmlu_eval_accuracy_professional_psychology": 0.4782608695652174,
192
  "mmlu_eval_accuracy_public_relations": 0.6666666666666666,
193
  "mmlu_eval_accuracy_security_studies": 0.4444444444444444,
194
  "mmlu_eval_accuracy_sociology": 0.7272727272727273,
195
  "mmlu_eval_accuracy_us_foreign_policy": 0.5454545454545454,
196
  "mmlu_eval_accuracy_virology": 0.4444444444444444,
197
  "mmlu_eval_accuracy_world_religions": 0.7368421052631579,
198
- "mmlu_loss": 1.5038623624619865,
199
  "step": 200
200
  }
201
  ],
 
1
  {
2
+ "best_metric": 0.8083679676055908,
3
  "best_model_checkpoint": "experts/expert-13/checkpoint-200",
4
  "epoch": 0.042034468263976464,
5
  "global_step": 200,
 
10
  {
11
  "epoch": 0.0,
12
  "learning_rate": 0.0002,
13
+ "loss": 0.9187,
14
  "step": 10
15
  },
16
  {
17
  "epoch": 0.0,
18
  "learning_rate": 0.0002,
19
+ "loss": 0.8703,
20
  "step": 20
21
  },
22
  {
23
  "epoch": 0.01,
24
  "learning_rate": 0.0002,
25
+ "loss": 0.8582,
26
  "step": 30
27
  },
28
  {
29
  "epoch": 0.01,
30
  "learning_rate": 0.0002,
31
+ "loss": 0.8973,
32
  "step": 40
33
  },
34
  {
35
  "epoch": 0.01,
36
  "learning_rate": 0.0002,
37
+ "loss": 0.8188,
38
  "step": 50
39
  },
40
  {
41
  "epoch": 0.01,
42
  "learning_rate": 0.0002,
43
+ "loss": 0.8114,
44
  "step": 60
45
  },
46
  {
47
  "epoch": 0.01,
48
  "learning_rate": 0.0002,
49
+ "loss": 0.7796,
50
  "step": 70
51
  },
52
  {
53
  "epoch": 0.02,
54
  "learning_rate": 0.0002,
55
+ "loss": 0.8671,
56
  "step": 80
57
  },
58
  {
59
  "epoch": 0.02,
60
  "learning_rate": 0.0002,
61
+ "loss": 0.8863,
62
  "step": 90
63
  },
64
  {
65
  "epoch": 0.02,
66
  "learning_rate": 0.0002,
67
+ "loss": 0.815,
68
  "step": 100
69
  },
70
  {
71
  "epoch": 0.02,
72
  "learning_rate": 0.0002,
73
+ "loss": 0.8255,
74
  "step": 110
75
  },
76
  {
77
  "epoch": 0.03,
78
  "learning_rate": 0.0002,
79
+ "loss": 0.8272,
80
  "step": 120
81
  },
82
  {
83
  "epoch": 0.03,
84
  "learning_rate": 0.0002,
85
+ "loss": 0.8148,
86
  "step": 130
87
  },
88
  {
89
  "epoch": 0.03,
90
  "learning_rate": 0.0002,
91
+ "loss": 0.8542,
92
  "step": 140
93
  },
94
  {
 
100
  {
101
  "epoch": 0.03,
102
  "learning_rate": 0.0002,
103
+ "loss": 0.8743,
104
  "step": 160
105
  },
106
  {
 
112
  {
113
  "epoch": 0.04,
114
  "learning_rate": 0.0002,
115
+ "loss": 0.7863,
116
  "step": 180
117
  },
118
  {
119
  "epoch": 0.04,
120
  "learning_rate": 0.0002,
121
+ "loss": 0.8397,
122
  "step": 190
123
  },
124
  {
125
  "epoch": 0.04,
126
  "learning_rate": 0.0002,
127
+ "loss": 0.7508,
128
  "step": 200
129
  },
130
  {
131
  "epoch": 0.04,
132
+ "eval_loss": 0.8083679676055908,
133
+ "eval_runtime": 68.466,
134
+ "eval_samples_per_second": 14.606,
135
+ "eval_steps_per_second": 7.303,
136
  "step": 200
137
  },
138
  {
139
  "epoch": 0.04,
140
+ "mmlu_eval_accuracy": 0.4843923123312093,
141
  "mmlu_eval_accuracy_abstract_algebra": 0.2727272727272727,
142
+ "mmlu_eval_accuracy_anatomy": 0.6428571428571429,
143
  "mmlu_eval_accuracy_astronomy": 0.4375,
144
+ "mmlu_eval_accuracy_business_ethics": 0.6363636363636364,
145
+ "mmlu_eval_accuracy_clinical_knowledge": 0.5172413793103449,
146
  "mmlu_eval_accuracy_college_biology": 0.4375,
147
  "mmlu_eval_accuracy_college_chemistry": 0.25,
148
  "mmlu_eval_accuracy_college_computer_science": 0.36363636363636365,
 
153
  "mmlu_eval_accuracy_conceptual_physics": 0.34615384615384615,
154
  "mmlu_eval_accuracy_econometrics": 0.25,
155
  "mmlu_eval_accuracy_electrical_engineering": 0.25,
156
+ "mmlu_eval_accuracy_elementary_mathematics": 0.36585365853658536,
157
  "mmlu_eval_accuracy_formal_logic": 0.07142857142857142,
158
  "mmlu_eval_accuracy_global_facts": 0.4,
159
  "mmlu_eval_accuracy_high_school_biology": 0.375,
160
  "mmlu_eval_accuracy_high_school_chemistry": 0.2727272727272727,
161
  "mmlu_eval_accuracy_high_school_computer_science": 0.5555555555555556,
162
+ "mmlu_eval_accuracy_high_school_european_history": 0.5555555555555556,
163
  "mmlu_eval_accuracy_high_school_geography": 0.9090909090909091,
164
+ "mmlu_eval_accuracy_high_school_government_and_politics": 0.5238095238095238,
165
+ "mmlu_eval_accuracy_high_school_macroeconomics": 0.4186046511627907,
166
  "mmlu_eval_accuracy_high_school_mathematics": 0.20689655172413793,
167
  "mmlu_eval_accuracy_high_school_microeconomics": 0.4230769230769231,
168
  "mmlu_eval_accuracy_high_school_physics": 0.11764705882352941,
169
  "mmlu_eval_accuracy_high_school_psychology": 0.85,
170
  "mmlu_eval_accuracy_high_school_statistics": 0.30434782608695654,
171
+ "mmlu_eval_accuracy_high_school_us_history": 0.6818181818181818,
172
  "mmlu_eval_accuracy_high_school_world_history": 0.7307692307692307,
173
  "mmlu_eval_accuracy_human_aging": 0.6956521739130435,
174
  "mmlu_eval_accuracy_human_sexuality": 0.3333333333333333,
175
  "mmlu_eval_accuracy_international_law": 0.9230769230769231,
176
+ "mmlu_eval_accuracy_jurisprudence": 0.45454545454545453,
177
  "mmlu_eval_accuracy_logical_fallacies": 0.6666666666666666,
178
+ "mmlu_eval_accuracy_machine_learning": 0.2727272727272727,
179
  "mmlu_eval_accuracy_management": 0.6363636363636364,
180
+ "mmlu_eval_accuracy_marketing": 0.88,
181
  "mmlu_eval_accuracy_medical_genetics": 0.9090909090909091,
182
+ "mmlu_eval_accuracy_miscellaneous": 0.686046511627907,
183
  "mmlu_eval_accuracy_moral_disputes": 0.4473684210526316,
184
+ "mmlu_eval_accuracy_moral_scenarios": 0.24,
185
  "mmlu_eval_accuracy_nutrition": 0.6060606060606061,
186
  "mmlu_eval_accuracy_philosophy": 0.4411764705882353,
187
  "mmlu_eval_accuracy_prehistory": 0.45714285714285713,
188
+ "mmlu_eval_accuracy_professional_accounting": 0.41935483870967744,
189
+ "mmlu_eval_accuracy_professional_law": 0.3411764705882353,
190
  "mmlu_eval_accuracy_professional_medicine": 0.6129032258064516,
191
+ "mmlu_eval_accuracy_professional_psychology": 0.4927536231884058,
192
  "mmlu_eval_accuracy_public_relations": 0.6666666666666666,
193
  "mmlu_eval_accuracy_security_studies": 0.4444444444444444,
194
  "mmlu_eval_accuracy_sociology": 0.7272727272727273,
195
  "mmlu_eval_accuracy_us_foreign_policy": 0.5454545454545454,
196
  "mmlu_eval_accuracy_virology": 0.4444444444444444,
197
  "mmlu_eval_accuracy_world_religions": 0.7368421052631579,
198
+ "mmlu_loss": 1.48628691881506,
199
  "step": 200
200
  }
201
  ],
checkpoint-200/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c88cdffadce27d1fa15d3da3b2e83029e51e413f14a49ea322f0a4eee84d961
3
  size 5819
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe7cdc1358006edc28e72abc36fb4c6703372887fc1df961096270e51aa9d994
3
  size 5819
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c88cdffadce27d1fa15d3da3b2e83029e51e413f14a49ea322f0a4eee84d961
3
  size 5819
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe7cdc1358006edc28e72abc36fb4c6703372887fc1df961096270e51aa9d994
3
  size 5819