aceini commited on
Commit
9f68dc1
·
verified ·
1 Parent(s): 34aaf67

Revert: model_type back to qwen3_5_text (Qwen3_5ForCausalLM compatible)

Browse files
Files changed (1) hide show
  1. config.json +2 -2
config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "architectures": [
3
- "Qwen3_5ForConditionalGeneration"
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
@@ -56,7 +56,7 @@
56
  "mamba_ssm_dtype": "float32",
57
  "max_position_embeddings": 262144,
58
  "mlp_only_layers": [],
59
- "model_type": "qwen3_5",
60
  "mtp_num_hidden_layers": 1,
61
  "mtp_use_dedicated_embeddings": false,
62
  "num_attention_heads": 16,
 
1
  {
2
  "architectures": [
3
+ "Qwen3_5ForCausalLM"
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
 
56
  "mamba_ssm_dtype": "float32",
57
  "max_position_embeddings": 262144,
58
  "mlp_only_layers": [],
59
+ "model_type": "qwen3_5_text",
60
  "mtp_num_hidden_layers": 1,
61
  "mtp_use_dedicated_embeddings": false,
62
  "num_attention_heads": 16,