Strip Forge-emitted dead keys (quantization_config, use_bidirectional_attention) from config.json
Browse files- config.json +0 -10
config.json
CHANGED
|
@@ -137,7 +137,6 @@
|
|
| 137 |
"sparse",
|
| 138 |
"sparse"
|
| 139 |
],
|
| 140 |
-
"use_bidirectional_attention": false,
|
| 141 |
"moe_routed_scaling_factor": 2.5,
|
| 142 |
"num_attention_heads_per_layer": [
|
| 143 |
48,
|
|
@@ -189,14 +188,5 @@
|
|
| 189 |
"output"
|
| 190 |
],
|
| 191 |
"recompute_fake_quantize": false
|
| 192 |
-
},
|
| 193 |
-
"quantization_config": {
|
| 194 |
-
"mode": null,
|
| 195 |
-
"group_size": 32,
|
| 196 |
-
"eps": 1e-05,
|
| 197 |
-
"filter_fqns": [
|
| 198 |
-
"output"
|
| 199 |
-
],
|
| 200 |
-
"recompute_fake_quantize": false
|
| 201 |
}
|
| 202 |
}
|
|
|
|
| 137 |
"sparse",
|
| 138 |
"sparse"
|
| 139 |
],
|
|
|
|
| 140 |
"moe_routed_scaling_factor": 2.5,
|
| 141 |
"num_attention_heads_per_layer": [
|
| 142 |
48,
|
|
|
|
| 188 |
"output"
|
| 189 |
],
|
| 190 |
"recompute_fake_quantize": false
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 191 |
}
|
| 192 |
}
|