| { | |
| "bits": 4, | |
| "data_type": "int", | |
| "group_size": 128, | |
| "sym": true, | |
| "autoround_version": "0.13.0", | |
| "block_name_to_quantize": "thinker.model.layers", | |
| "quant_method": "auto-round", | |
| "packing_format": "auto_round:auto_gptq", | |
| "extra_config": { | |
| ".*thinker\\.model\\.layers\\.\\d+\\.mlp\\.gate.*": { | |
| "bits": 16, | |
| "data_type": "float" | |
| } | |
| } | |
| } |