thieu86 commited on
Commit
1b43143
·
verified ·
1 Parent(s): 09db226

Run 5. Outer Step 16. Inner Step 2.

Browse files
Files changed (4) hide show
  1. config.json +68 -81
  2. generation_config.json +7 -0
  3. inner_optimizer.pt +2 -2
  4. model.safetensors +2 -2
config.json CHANGED
@@ -1,9 +1,8 @@
1
  {
2
- "_name_or_path": "distributed/optimized-gpt2-1b",
3
- "activation_function": "gelu_new",
4
  "all_reduce_scores": {
5
  "0": "NON_PARTICIPATING",
6
- "1": "NON_PARTICIPATING",
7
  "10": "NON_PARTICIPATING",
8
  "100": "NON_PARTICIPATING",
9
  "101": "NON_PARTICIPATING",
@@ -12,8 +11,8 @@
12
  "104": "NON_PARTICIPATING",
13
  "105": "NON_PARTICIPATING",
14
  "106": "NON_PARTICIPATING",
15
- "107": "SUCCESS",
16
- "108": "SUCCESS",
17
  "109": "NON_PARTICIPATING",
18
  "11": "SUCCESS",
19
  "110": "NON_PARTICIPATING",
@@ -24,13 +23,13 @@
24
  "115": "NON_PARTICIPATING",
25
  "116": "NON_PARTICIPATING",
26
  "117": "NON_PARTICIPATING",
27
- "118": "SUCCESS",
28
  "119": "NON_PARTICIPATING",
29
  "12": "SUCCESS",
30
- "120": "SUCCESS",
31
- "121": "NON_PARTICIPATING",
32
- "122": "SUCCESS",
33
- "123": "SUCCESS",
34
  "124": "NON_PARTICIPATING",
35
  "125": "NON_PARTICIPATING",
36
  "126": "NON_PARTICIPATING",
@@ -50,7 +49,7 @@
50
  "139": "NON_PARTICIPATING",
51
  "14": "SUCCESS",
52
  "140": "NON_PARTICIPATING",
53
- "141": "NON_PARTICIPATING",
54
  "142": "NON_PARTICIPATING",
55
  "143": "SUCCESS",
56
  "144": "NON_PARTICIPATING",
@@ -63,10 +62,10 @@
63
  "150": "NON_PARTICIPATING",
64
  "151": "NON_PARTICIPATING",
65
  "152": "NON_PARTICIPATING",
66
- "153": "SUCCESS",
67
- "154": "NON_PARTICIPATING",
68
  "155": "NON_PARTICIPATING",
69
- "156": "SUCCESS",
70
  "157": "NON_PARTICIPATING",
71
  "158": "NON_PARTICIPATING",
72
  "159": "NON_PARTICIPATING",
@@ -75,10 +74,10 @@
75
  "161": "NON_PARTICIPATING",
76
  "162": "NON_PARTICIPATING",
77
  "163": "NON_PARTICIPATING",
78
- "164": "SUCCESS",
79
- "165": "SUCCESS",
80
  "166": "NON_PARTICIPATING",
81
- "167": "NON_PARTICIPATING",
82
  "168": "NON_PARTICIPATING",
83
  "169": "NON_PARTICIPATING",
84
  "17": "SUCCESS",
@@ -89,7 +88,7 @@
89
  "174": "NON_PARTICIPATING",
90
  "175": "NON_PARTICIPATING",
91
  "176": "NON_PARTICIPATING",
92
- "177": "SUCCESS",
93
  "178": "NON_PARTICIPATING",
94
  "179": "NON_PARTICIPATING",
95
  "18": "SUCCESS",
@@ -100,15 +99,15 @@
100
  "184": "NON_PARTICIPATING",
101
  "185": "NON_PARTICIPATING",
102
  "186": "NON_PARTICIPATING",
103
- "187": "SUCCESS",
104
  "188": "NON_PARTICIPATING",
105
  "189": "NON_PARTICIPATING",
106
  "19": "SUCCESS",
107
  "190": "NON_PARTICIPATING",
108
  "191": "NON_PARTICIPATING",
109
  "192": "NON_PARTICIPATING",
110
- "193": "SUCCESS",
111
- "194": "NON_PARTICIPATING",
112
  "195": "NON_PARTICIPATING",
113
  "196": "NON_PARTICIPATING",
114
  "197": "NON_PARTICIPATING",
@@ -118,7 +117,7 @@
118
  "20": "SUCCESS",
119
  "200": "NON_PARTICIPATING",
120
  "201": "NON_PARTICIPATING",
121
- "202": "SUCCESS",
122
  "203": "NON_PARTICIPATING",
123
  "204": "NON_PARTICIPATING",
124
  "205": "NON_PARTICIPATING",
@@ -127,7 +126,7 @@
127
  "208": "NON_PARTICIPATING",
128
  "209": "SUCCESS",
129
  "21": "SUCCESS",
130
- "210": "SUCCESS",
131
  "211": "NON_PARTICIPATING",
132
  "212": "NON_PARTICIPATING",
133
  "213": "NON_PARTICIPATING",
@@ -146,7 +145,7 @@
146
  "225": "NON_PARTICIPATING",
147
  "226": "NON_PARTICIPATING",
148
  "227": "NON_PARTICIPATING",
149
- "228": "SUCCESS",
150
  "229": "NON_PARTICIPATING",
151
  "23": "SUCCESS",
152
  "230": "NON_PARTICIPATING",
@@ -164,10 +163,10 @@
164
  "241": "NON_PARTICIPATING",
165
  "242": "NON_PARTICIPATING",
166
  "243": "NON_PARTICIPATING",
167
- "244": "SUCCESS",
168
  "245": "NON_PARTICIPATING",
169
  "246": "NON_PARTICIPATING",
170
- "247": "SUCCESS",
171
  "248": "NON_PARTICIPATING",
172
  "249": "NON_PARTICIPATING",
173
  "25": "SUCCESS",
@@ -181,9 +180,9 @@
181
  "27": "SUCCESS",
182
  "28": "SUCCESS",
183
  "29": "SUCCESS",
184
- "3": "NON_PARTICIPATING",
185
- "30": "NON_PARTICIPATING",
186
- "31": "NON_PARTICIPATING",
187
  "32": "SUCCESS",
188
  "33": "NON_PARTICIPATING",
189
  "34": "SUCCESS",
@@ -199,26 +198,26 @@
199
  "43": "SUCCESS",
200
  "44": "SUCCESS",
201
  "45": "SUCCESS",
202
- "46": "NON_PARTICIPATING",
203
- "47": "NON_PARTICIPATING",
204
  "48": "NON_PARTICIPATING",
205
- "49": "NON_PARTICIPATING",
206
  "5": "SUCCESS",
207
  "50": "SUCCESS",
208
- "51": "NON_PARTICIPATING",
209
  "52": "NON_PARTICIPATING",
210
  "53": "SUCCESS",
211
  "54": "SUCCESS",
212
  "55": "SUCCESS",
213
  "56": "SUCCESS",
214
- "57": "NON_PARTICIPATING",
215
  "58": "NON_PARTICIPATING",
216
  "59": "SUCCESS",
217
- "6": "NON_PARTICIPATING",
218
  "60": "SUCCESS",
219
- "61": "NON_PARTICIPATING",
220
  "62": "SUCCESS",
221
- "63": "SUCCESS",
222
  "64": "SUCCESS",
223
  "65": "SUCCESS",
224
  "66": "SUCCESS",
@@ -227,23 +226,23 @@
227
  "69": "SUCCESS",
228
  "7": "SUCCESS",
229
  "70": "SUCCESS",
230
- "71": "NON_PARTICIPATING",
231
  "72": "SUCCESS",
232
- "73": "NON_PARTICIPATING",
233
  "74": "SUCCESS",
234
  "75": "SUCCESS",
235
  "76": "SUCCESS",
236
  "77": "SUCCESS",
237
  "78": "SUCCESS",
238
- "79": "NON_PARTICIPATING",
239
- "8": "NON_PARTICIPATING",
240
  "80": "SUCCESS",
241
  "81": "SUCCESS",
242
  "82": "SUCCESS",
243
  "83": "SUCCESS",
244
  "84": "SUCCESS",
245
- "85": "NON_PARTICIPATING",
246
- "86": "NON_PARTICIPATING",
247
  "87": "SUCCESS",
248
  "88": "SUCCESS",
249
  "89": "SUCCESS",
@@ -251,7 +250,7 @@
251
  "90": "SUCCESS",
252
  "91": "SUCCESS",
253
  "92": "SUCCESS",
254
- "93": "NON_PARTICIPATING",
255
  "94": "SUCCESS",
256
  "95": "SUCCESS",
257
  "96": "SUCCESS",
@@ -260,47 +259,35 @@
260
  "99": "SUCCESS"
261
  },
262
  "architectures": [
263
- "GPTOptim"
264
  ],
265
- "attn_pdrop": 0.1,
266
- "auto_map": {
267
- "AutoConfig": "distributed/optimized-gpt2-500m--configuration_gpt_optimized.GPTOptimConfig",
268
- "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
269
- },
270
  "block_list": [
271
- 5619541,
272
- 5619545,
273
- 5619548,
274
- 5619552,
275
- 5619555,
276
- 5619558
277
  ],
278
- "block_size": 1024,
279
- "bos_token_id": 50256,
280
- "embd_pdrop": 0.1,
281
- "eos_token_id": 50256,
282
  "initializer_range": 0.02,
283
- "inner_step": 6,
284
- "inner_steps": 0,
285
- "last_allreduce_block": 5619221,
286
- "layer_norm_epsilon": 1e-05,
287
- "model_type": "gpt_optimized",
288
- "n_embd": 1280,
289
- "n_head": 32,
290
- "n_inner": null,
291
- "n_layer": 48,
292
- "n_positions": 1024,
293
- "reorder_and_upcast_attn": false,
294
- "resid_pdrop": 0.1,
295
- "scale_attn_by_inverse_layer_idx": false,
296
- "scale_attn_weights": true,
297
- "summary_activation": null,
298
- "summary_first_dropout": 0.1,
299
- "summary_proj_to_labels": true,
300
- "summary_type": "cls_index",
301
- "summary_use_proj": true,
302
  "torch_dtype": "float32",
303
  "transformers_version": "4.39.3",
304
- "use_cache": true,
305
- "vocab_size": 50257
306
  }
 
1
  {
2
+ "_name_or_path": "distributed/llama-1b",
 
3
  "all_reduce_scores": {
4
  "0": "NON_PARTICIPATING",
5
+ "1": "SUCCESS",
6
  "10": "NON_PARTICIPATING",
7
  "100": "NON_PARTICIPATING",
8
  "101": "NON_PARTICIPATING",
 
11
  "104": "NON_PARTICIPATING",
12
  "105": "NON_PARTICIPATING",
13
  "106": "NON_PARTICIPATING",
14
+ "107": "NON_PARTICIPATING",
15
+ "108": "NON_PARTICIPATING",
16
  "109": "NON_PARTICIPATING",
17
  "11": "SUCCESS",
18
  "110": "NON_PARTICIPATING",
 
23
  "115": "NON_PARTICIPATING",
24
  "116": "NON_PARTICIPATING",
25
  "117": "NON_PARTICIPATING",
26
+ "118": "NON_PARTICIPATING",
27
  "119": "NON_PARTICIPATING",
28
  "12": "SUCCESS",
29
+ "120": "NON_PARTICIPATING",
30
+ "121": "SUCCESS",
31
+ "122": "NON_PARTICIPATING",
32
+ "123": "NON_PARTICIPATING",
33
  "124": "NON_PARTICIPATING",
34
  "125": "NON_PARTICIPATING",
35
  "126": "NON_PARTICIPATING",
 
49
  "139": "NON_PARTICIPATING",
50
  "14": "SUCCESS",
51
  "140": "NON_PARTICIPATING",
52
+ "141": "SUCCESS",
53
  "142": "NON_PARTICIPATING",
54
  "143": "SUCCESS",
55
  "144": "NON_PARTICIPATING",
 
62
  "150": "NON_PARTICIPATING",
63
  "151": "NON_PARTICIPATING",
64
  "152": "NON_PARTICIPATING",
65
+ "153": "NON_PARTICIPATING",
66
+ "154": "SUCCESS",
67
  "155": "NON_PARTICIPATING",
68
+ "156": "NON_PARTICIPATING",
69
  "157": "NON_PARTICIPATING",
70
  "158": "NON_PARTICIPATING",
71
  "159": "NON_PARTICIPATING",
 
74
  "161": "NON_PARTICIPATING",
75
  "162": "NON_PARTICIPATING",
76
  "163": "NON_PARTICIPATING",
77
+ "164": "NON_PARTICIPATING",
78
+ "165": "NON_PARTICIPATING",
79
  "166": "NON_PARTICIPATING",
80
+ "167": "SUCCESS",
81
  "168": "NON_PARTICIPATING",
82
  "169": "NON_PARTICIPATING",
83
  "17": "SUCCESS",
 
88
  "174": "NON_PARTICIPATING",
89
  "175": "NON_PARTICIPATING",
90
  "176": "NON_PARTICIPATING",
91
+ "177": "NON_PARTICIPATING",
92
  "178": "NON_PARTICIPATING",
93
  "179": "NON_PARTICIPATING",
94
  "18": "SUCCESS",
 
99
  "184": "NON_PARTICIPATING",
100
  "185": "NON_PARTICIPATING",
101
  "186": "NON_PARTICIPATING",
102
+ "187": "NON_PARTICIPATING",
103
  "188": "NON_PARTICIPATING",
104
  "189": "NON_PARTICIPATING",
105
  "19": "SUCCESS",
106
  "190": "NON_PARTICIPATING",
107
  "191": "NON_PARTICIPATING",
108
  "192": "NON_PARTICIPATING",
109
+ "193": "NON_PARTICIPATING",
110
+ "194": "SUCCESS",
111
  "195": "NON_PARTICIPATING",
112
  "196": "NON_PARTICIPATING",
113
  "197": "NON_PARTICIPATING",
 
117
  "20": "SUCCESS",
118
  "200": "NON_PARTICIPATING",
119
  "201": "NON_PARTICIPATING",
120
+ "202": "NON_PARTICIPATING",
121
  "203": "NON_PARTICIPATING",
122
  "204": "NON_PARTICIPATING",
123
  "205": "NON_PARTICIPATING",
 
126
  "208": "NON_PARTICIPATING",
127
  "209": "SUCCESS",
128
  "21": "SUCCESS",
129
+ "210": "NON_PARTICIPATING",
130
  "211": "NON_PARTICIPATING",
131
  "212": "NON_PARTICIPATING",
132
  "213": "NON_PARTICIPATING",
 
145
  "225": "NON_PARTICIPATING",
146
  "226": "NON_PARTICIPATING",
147
  "227": "NON_PARTICIPATING",
148
+ "228": "NON_PARTICIPATING",
149
  "229": "NON_PARTICIPATING",
150
  "23": "SUCCESS",
151
  "230": "NON_PARTICIPATING",
 
163
  "241": "NON_PARTICIPATING",
164
  "242": "NON_PARTICIPATING",
165
  "243": "NON_PARTICIPATING",
166
+ "244": "NON_PARTICIPATING",
167
  "245": "NON_PARTICIPATING",
168
  "246": "NON_PARTICIPATING",
169
+ "247": "NON_PARTICIPATING",
170
  "248": "NON_PARTICIPATING",
171
  "249": "NON_PARTICIPATING",
172
  "25": "SUCCESS",
 
180
  "27": "SUCCESS",
181
  "28": "SUCCESS",
182
  "29": "SUCCESS",
183
+ "3": "SUCCESS",
184
+ "30": "SUCCESS",
185
+ "31": "SUCCESS",
186
  "32": "SUCCESS",
187
  "33": "NON_PARTICIPATING",
188
  "34": "SUCCESS",
 
198
  "43": "SUCCESS",
199
  "44": "SUCCESS",
200
  "45": "SUCCESS",
201
+ "46": "SUCCESS",
202
+ "47": "SUCCESS",
203
  "48": "NON_PARTICIPATING",
204
+ "49": "SUCCESS",
205
  "5": "SUCCESS",
206
  "50": "SUCCESS",
207
+ "51": "SUCCESS",
208
  "52": "NON_PARTICIPATING",
209
  "53": "SUCCESS",
210
  "54": "SUCCESS",
211
  "55": "SUCCESS",
212
  "56": "SUCCESS",
213
+ "57": "SUCCESS",
214
  "58": "NON_PARTICIPATING",
215
  "59": "SUCCESS",
216
+ "6": "SUCCESS",
217
  "60": "SUCCESS",
218
+ "61": "SUCCESS",
219
  "62": "SUCCESS",
220
+ "63": "FAIL",
221
  "64": "SUCCESS",
222
  "65": "SUCCESS",
223
  "66": "SUCCESS",
 
226
  "69": "SUCCESS",
227
  "7": "SUCCESS",
228
  "70": "SUCCESS",
229
+ "71": "SUCCESS",
230
  "72": "SUCCESS",
231
+ "73": "SUCCESS",
232
  "74": "SUCCESS",
233
  "75": "SUCCESS",
234
  "76": "SUCCESS",
235
  "77": "SUCCESS",
236
  "78": "SUCCESS",
237
+ "79": "SUCCESS",
238
+ "8": "SUCCESS",
239
  "80": "SUCCESS",
240
  "81": "SUCCESS",
241
  "82": "SUCCESS",
242
  "83": "SUCCESS",
243
  "84": "SUCCESS",
244
+ "85": "SUCCESS",
245
+ "86": "SUCCESS",
246
  "87": "SUCCESS",
247
  "88": "SUCCESS",
248
  "89": "SUCCESS",
 
250
  "90": "SUCCESS",
251
  "91": "SUCCESS",
252
  "92": "SUCCESS",
253
+ "93": "SUCCESS",
254
  "94": "SUCCESS",
255
  "95": "SUCCESS",
256
  "96": "SUCCESS",
 
259
  "99": "SUCCESS"
260
  },
261
  "architectures": [
262
+ "LlamaForCausalLM"
263
  ],
264
+ "attention_bias": false,
265
+ "attention_dropout": 0.0,
 
 
 
266
  "block_list": [
267
+ 5722373,
268
+ 5722401
 
 
 
 
269
  ],
270
+ "bos_token_id": 1,
271
+ "eos_token_id": 2,
272
+ "hidden_act": "silu",
273
+ "hidden_size": 2048,
274
  "initializer_range": 0.02,
275
+ "inner_step": 2,
276
+ "intermediate_size": 5632,
277
+ "last_allreduce_block": 5721007,
278
+ "max_position_embeddings": 2048,
279
+ "mlp_bias": false,
280
+ "model_type": "llama",
281
+ "num_attention_heads": 32,
282
+ "num_hidden_layers": 22,
283
+ "num_key_value_heads": 4,
284
+ "pretraining_tp": 1,
285
+ "rms_norm_eps": 1e-05,
286
+ "rope_scaling": null,
287
+ "rope_theta": 10000.0,
288
+ "tie_word_embeddings": false,
 
 
 
 
 
289
  "torch_dtype": "float32",
290
  "transformers_version": "4.39.3",
291
+ "use_cache": false,
292
+ "vocab_size": 32000
293
  }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.39.3",
6
+ "use_cache": false
7
+ }
inner_optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:365e22bbdb7d38eaaf09b0a256cfc063b40b30671344b346ab19833297f93c9a
3
- size 8081782026
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f22f00ee4062d956d35a0d1116b19519447ceff6dcea7c08be30f07f95dc247
3
+ size 8800564076
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:83cc3751e8424fb9b1e72f5920421da0f005e65b8cfdafc7459e26efed937f54
3
- size 4040701744
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da7a37a9b753366ecf8097bb34c6e8c5584835712dcbd8d4b1ea80da6121a52e
3
+ size 4400216536