kernelpool commited on
Commit
8e4d56f
·
verified ·
1 Parent(s): dea09fb

Add files using upload-large-folder tool

Browse files
Files changed (41) hide show
  1. README.md +36 -0
  2. chat_template.jinja +195 -0
  3. config.json +688 -0
  4. generation_config.json +10 -0
  5. model-00001-of-00034.safetensors +3 -0
  6. model-00002-of-00034.safetensors +3 -0
  7. model-00003-of-00034.safetensors +3 -0
  8. model-00004-of-00034.safetensors +3 -0
  9. model-00005-of-00034.safetensors +3 -0
  10. model-00006-of-00034.safetensors +3 -0
  11. model-00007-of-00034.safetensors +3 -0
  12. model-00008-of-00034.safetensors +3 -0
  13. model-00009-of-00034.safetensors +3 -0
  14. model-00010-of-00034.safetensors +3 -0
  15. model-00011-of-00034.safetensors +3 -0
  16. model-00012-of-00034.safetensors +3 -0
  17. model-00013-of-00034.safetensors +3 -0
  18. model-00014-of-00034.safetensors +3 -0
  19. model-00015-of-00034.safetensors +3 -0
  20. model-00016-of-00034.safetensors +3 -0
  21. model-00017-of-00034.safetensors +3 -0
  22. model-00018-of-00034.safetensors +3 -0
  23. model-00019-of-00034.safetensors +3 -0
  24. model-00020-of-00034.safetensors +3 -0
  25. model-00021-of-00034.safetensors +3 -0
  26. model-00022-of-00034.safetensors +3 -0
  27. model-00023-of-00034.safetensors +3 -0
  28. model-00024-of-00034.safetensors +3 -0
  29. model-00025-of-00034.safetensors +3 -0
  30. model-00026-of-00034.safetensors +3 -0
  31. model-00027-of-00034.safetensors +3 -0
  32. model-00028-of-00034.safetensors +3 -0
  33. model-00029-of-00034.safetensors +3 -0
  34. model-00030-of-00034.safetensors +3 -0
  35. model-00031-of-00034.safetensors +3 -0
  36. model-00032-of-00034.safetensors +3 -0
  37. model-00033-of-00034.safetensors +3 -0
  38. model-00034-of-00034.safetensors +3 -0
  39. model.safetensors.index.json +0 -0
  40. tokenizer.json +0 -0
  41. tokenizer_config.json +11 -0
README.md ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ library_name: mlx
4
+ tags:
5
+ - mlx
6
+ base_model: tencent/Hy3-preview
7
+ pipeline_tag: text-generation
8
+ ---
9
+
10
+ # mlx-community/Hy3-preview-4bit
11
+
12
+ This model [mlx-community/Hy3-preview-4bit](https://huggingface.co/mlx-community/Hy3-preview-4bit) was
13
+ converted to MLX format from [tencent/Hy3-preview](https://huggingface.co/tencent/Hy3-preview)
14
+ using mlx-lm version **0.31.3**.
15
+
16
+ ## Use with mlx
17
+
18
+ ```bash
19
+ pip install mlx-lm
20
+ ```
21
+
22
+ ```python
23
+ from mlx_lm import load, generate
24
+
25
+ model, tokenizer = load("mlx-community/Hy3-preview-4bit")
26
+
27
+ prompt = "hello"
28
+
29
+ if tokenizer.chat_template is not None:
30
+ messages = [{"role": "user", "content": prompt}]
31
+ prompt = tokenizer.apply_chat_template(
32
+ messages, add_generation_prompt=True, return_dict=False,
33
+ )
34
+
35
+ response = generate(model, tokenizer, prompt=prompt, verbose=True)
36
+ ```
chat_template.jinja ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {#- ----------‑‑‑ special token variables ‑‑‑---------- -#}
2
+ {%- set bos_token = '<|hy_begin▁of▁sentence|>' %}
3
+ {%- set pad_token = '<|hy_▁pad▁|>' %}
4
+ {%- set user_token = '<|hy_User|>' %}
5
+ {%- set assistant_token = '<|hy_Assistant|>' %}
6
+ {%- set eos_token = '<|hy_eos|>' %}
7
+ {%- set think_begin_token = '<think>' %}
8
+ {%- set think_end_token = '</think>' %}
9
+ {%- set toolcalls_begin_token = '<tool_calls>' %}
10
+ {%- set toolcalls_end_token = '</tool_calls>' %}
11
+ {%- set toolcall_begin_token = '<tool_call>' %}
12
+ {%- set toolcall_end_token = '</tool_call>' %}
13
+ {%- set toolsep_token = '<tool_sep>' %}
14
+ {%- set argkey_begin_token = '<arg_key>' %}
15
+ {%- set argkey_end_token = '</arg_key>' %}
16
+ {%- set argvalue_begin_token = '<arg_value>' %}
17
+ {%- set argvalue_end_token = '</arg_value>' %}
18
+ {%- set toolresponses_begin_token = '<tool_responses>' %}
19
+ {%- set toolresponses_end_token = '</tool_responses>' %}
20
+ {%- set toolresponse_begin_token = '<tool_response>' %}
21
+ {%- set toolresponse_end_token = '</tool_response>' %}
22
+ {%- set reasoning_mode_token = '<|reasoning_mode|>' %}
23
+ {#- ----------‑‑‑ hyperparameters variables ‑‑‑---------- -#}
24
+ {%- if not add_generation_prompt is defined %}
25
+ {%- set add_generation_prompt = false %}
26
+ {%- endif %}
27
+ {%- if not interleaved_thinking is defined %}
28
+ {%- set interleaved_thinking = false %}
29
+ {%- endif %}
30
+ {%- if not tools %}
31
+ {%- set interleaved_thinking = false %}
32
+ {%- endif %}
33
+ {%- if not is_training is defined %}
34
+ {%- set is_training = false %}
35
+ {%- endif %}
36
+ {%- if not reasoning_effort is defined or reasoning_effort not in ['high', 'low', 'no_think'] %}
37
+ {%- set reasoning_effort = 'no_think' %}
38
+ {%- endif %}
39
+
40
+ {%- macro visible_text(content) -%}
41
+ {%- if content is string -%}
42
+ {{- content }}
43
+ {%- elif content is iterable and content is not mapping -%}
44
+ {%- for item in content -%}
45
+ {%- if item is mapping and item.type == 'text' -%}
46
+ {{- item.text }}
47
+ {%- elif item is string -%}
48
+ {{- item }}
49
+ {%- endif -%}
50
+ {%- endfor -%}
51
+ {%- elif content is none -%}
52
+ {{- '' }}
53
+ {%- else -%}
54
+ {{- content }}
55
+ {%- endif -%}
56
+ {%- endmacro -%}
57
+
58
+ {%- set ns = namespace(last_user_index=-1) %}
59
+ {%- set sp_ns = namespace(system_prompt='', is_first_sp=true) %}
60
+ {%- for message in messages %}
61
+ {%- if message['role'] == 'system' %}
62
+ {%- set sp_ns.system_prompt = sp_ns.system_prompt + visible_text(message['content']) %}
63
+ {%- endif %}
64
+ {%- if message['role'] == 'user' %}
65
+ {%- set ns.last_user_index = loop.index0 %}
66
+ {%- endif %}
67
+ {%- endfor %}
68
+ {%- if reasoning_effort is defined and reasoning_effort is string and reasoning_effort != '' and not tools %}
69
+ {%- set sp_ns.system_prompt = sp_ns.system_prompt + reasoning_mode_token + 'reasoning_effort:' + reasoning_effort %}
70
+ {%- endif %}
71
+ {{- bos_token }}
72
+ {{- sp_ns.system_prompt }}
73
+ {%- if tools %}
74
+ {%- if sp_ns.system_prompt != '' %}
75
+ {{- '\n\n# Tools\n\nYou may call one or more functions to assist with the user query.' }}
76
+ {%- else %}
77
+ {{- '# Tools\n\nYou may call one or more functions to assist with the user query.' }}
78
+ {%- endif %}
79
+ {{- '\n\nYou are provided with function signatures within <tools></tools> XML tags:' }}
80
+ {{- '\n<tools>\n' }}
81
+ {%- for tool in tools %}
82
+ {%- if loop.index0 > 0 %}
83
+ {{- '\n' }}
84
+ {%- endif %}
85
+ {{- tool | tojson }}
86
+ {%- endfor %}
87
+ {{- '\n</tools>\n\n' }}
88
+ {{- 'For function call returns, you should first print ' + toolcalls_begin_token + '\n' }}
89
+ {{- 'For each function call, you should return object like:\n' }}
90
+ {{- toolcall_begin_token + '{function-name}' + toolsep_token + '\n' }}
91
+ {{- argkey_begin_token + '{arg-key-1}' + argkey_end_token + '\n' }}
92
+ {{- argvalue_begin_token + '{arg-value-1}' + argvalue_end_token + '\n' }}
93
+ {{- argkey_begin_token + '{arg-key-2}' + argkey_end_token + '\n' }}
94
+ {{- argvalue_begin_token + '{arg-value-2}' + argvalue_end_token + '\n' }}
95
+ {{- '...\n' }}
96
+ {{- toolcall_end_token + '\n' }}
97
+ {%- if reasoning_effort is defined and reasoning_effort is string and reasoning_effort != '' %}
98
+ {{- 'At the end of function call returns, you should print ' + toolcalls_end_token + reasoning_mode_token + 'reasoning_effort:' + reasoning_effort }}
99
+ {%- else %}
100
+ {{- 'At the end of function call returns, you should print ' + toolcalls_end_token }}
101
+ {%- endif %}
102
+ {%- endif %}
103
+
104
+ {%- set prev_ns = namespace(is_tool=false, is_tool_first=true) %}
105
+ {%- set last_ns = namespace(last_is_assistant=false) %}
106
+ {%- for message in messages %}
107
+ {%- if message['role'] == 'user' %}
108
+ {%- if prev_ns.is_tool %}
109
+ {{- toolresponses_end_token }}
110
+ {%- endif %}
111
+ {{- user_token + visible_text(message['content']) }}
112
+ {%- set prev_ns.is_tool = false %}
113
+ {%- endif %}
114
+ {%- if message['role'] == 'assistant' %}
115
+ {%- if 'reasoning_content' in message and message['reasoning_content'] is string %}
116
+ {%- set rc = message['reasoning_content'] %}
117
+ {%- elif 'reasoning' in message and message['reasoning'] is string %}
118
+ {%- set rc = message['reasoning'] %}
119
+ {%- else %}
120
+ {%- set rc = none %}
121
+ {%- endif %}
122
+ {%- if is_training %}
123
+ {%- if rc is not none %}
124
+ {%- set content = think_begin_token + rc + think_end_token + visible_text(message['content']) %}
125
+ {%- else %}
126
+ {%- set content = think_begin_token + think_end_token + visible_text(message['content']) %}
127
+ {%- endif %}
128
+ {%- else %}
129
+ {%- if interleaved_thinking %}
130
+ {%- if loop.index0 > ns.last_user_index and rc is not none %}
131
+ {%- set content = think_begin_token + rc + think_end_token + visible_text(message['content']) %}
132
+ {%- else %}
133
+ {%- set content = think_begin_token + think_end_token + visible_text(message['content']) %}
134
+ {%- endif %}
135
+ {%- else %}
136
+ {%- set content = think_begin_token + think_end_token + visible_text(message['content']) %}
137
+ {%- endif %}
138
+ {%- endif %}
139
+ {%- if prev_ns.is_tool %}
140
+ {{- toolresponses_end_token }}
141
+ {%- endif %}
142
+ {{- assistant_token }}
143
+ {%- if message['tool_calls'] is defined and message['tool_calls'] %}
144
+ {%- set prev_ns.is_tool_first = true %}
145
+ {{- content }}
146
+ {{- toolcalls_begin_token + '\n' }}
147
+ {%- for tool in message['tool_calls'] %}
148
+ {%- set arguments = tool['function']['arguments'] %}
149
+ {{- toolcall_begin_token + tool['function']['name'] + toolsep_token + '\n' }}
150
+ {%- for key, value in arguments.items() %}
151
+ {{- argkey_begin_token + key + argkey_end_token + '\n' }}
152
+ {%- if value is not string %}
153
+ {%- set value = value | tojson(ensure_ascii=False) %}
154
+ {%- endif %}
155
+ {{- argvalue_begin_token + value + argvalue_end_token + '\n' }}
156
+ {%- endfor %}
157
+ {{- toolcall_end_token + '\n' }}
158
+ {%- endfor %}
159
+ {{- toolcalls_end_token + eos_token }}
160
+ {%- else %}
161
+ {%- if not loop.last or is_training %}
162
+ {{- content + eos_token }}
163
+ {%- else %}
164
+ {{- content }}
165
+ {%- endif %}
166
+ {%- endif %}
167
+ {%- set prev_ns.is_tool = false %}
168
+ {%- endif %}
169
+ {%- if message['role'] == 'tool' %}
170
+ {%- set prev_ns.is_tool = true %}
171
+ {%- if prev_ns.is_tool_first %}
172
+ {{- toolresponses_begin_token + '\n' }}
173
+ {%- set prev_ns.is_tool_first = false %}
174
+ {%- endif %}
175
+ {{- toolresponse_begin_token + '\n' + visible_text(message['content']) + '\n' + toolresponse_end_token + '\n' }}
176
+ {%- endif %}
177
+ {%- if loop.last and message['role'] == 'assistant' %}
178
+ {%- set last_ns.last_is_assistant = true %}
179
+ {%- endif %}
180
+
181
+ {%- endfor %}
182
+ {%- if prev_ns.is_tool %}
183
+ {{- toolresponses_end_token }}
184
+ {%- endif %}
185
+ {%- if add_generation_prompt %}
186
+ {%- if not last_ns.last_is_assistant %}
187
+ {%- if reasoning_effort is defined and reasoning_effort in ['low', 'high'] %}
188
+ {{- assistant_token + think_begin_token }}
189
+ {%- elif reasoning_effort is defined and reasoning_effort == 'no_think' %}
190
+ {{- assistant_token + think_begin_token + think_end_token }}
191
+ {%- else %}
192
+ {{- assistant_token }}
193
+ {%- endif %}
194
+ {%- endif %}
195
+ {%- endif %}
config.json ADDED
@@ -0,0 +1,688 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "HYV3ForCausalLM"
4
+ ],
5
+ "bos_token_id": 120000,
6
+ "enable_attention_fp32_softmax": false,
7
+ "enable_lm_head_fp32": true,
8
+ "enable_moe_fp32_combine": false,
9
+ "eod_token_id": 120026,
10
+ "eos_token_id": 120025,
11
+ "expert_hidden_dim": 1536,
12
+ "first_k_dense_replace": 1,
13
+ "head_dim": 128,
14
+ "hidden_act": "silu",
15
+ "hidden_size": 4096,
16
+ "initializer_range": 0.006,
17
+ "intermediate_size": 13312,
18
+ "max_position_embeddings": 262144,
19
+ "model_type": "hy_v3",
20
+ "moe_intermediate_size": 1536,
21
+ "moe_router_enable_expert_bias": true,
22
+ "moe_router_use_sigmoid": true,
23
+ "num_attention_heads": 64,
24
+ "num_experts": 192,
25
+ "num_experts_per_tok": 8,
26
+ "num_hidden_layers": 80,
27
+ "num_key_value_heads": 8,
28
+ "num_nextn_predict_layers": 1,
29
+ "num_shared_experts": 1,
30
+ "output_router_logits": true,
31
+ "pad_token_id": 120002,
32
+ "qk_norm": true,
33
+ "quantization": {
34
+ "group_size": 64,
35
+ "bits": 4,
36
+ "mode": "affine",
37
+ "model.layers.1.mlp.router.gate": {
38
+ "group_size": 64,
39
+ "bits": 8
40
+ },
41
+ "model.layers.2.mlp.router.gate": {
42
+ "group_size": 64,
43
+ "bits": 8
44
+ },
45
+ "model.layers.3.mlp.router.gate": {
46
+ "group_size": 64,
47
+ "bits": 8
48
+ },
49
+ "model.layers.4.mlp.router.gate": {
50
+ "group_size": 64,
51
+ "bits": 8
52
+ },
53
+ "model.layers.5.mlp.router.gate": {
54
+ "group_size": 64,
55
+ "bits": 8
56
+ },
57
+ "model.layers.6.mlp.router.gate": {
58
+ "group_size": 64,
59
+ "bits": 8
60
+ },
61
+ "model.layers.7.mlp.router.gate": {
62
+ "group_size": 64,
63
+ "bits": 8
64
+ },
65
+ "model.layers.8.mlp.router.gate": {
66
+ "group_size": 64,
67
+ "bits": 8
68
+ },
69
+ "model.layers.9.mlp.router.gate": {
70
+ "group_size": 64,
71
+ "bits": 8
72
+ },
73
+ "model.layers.10.mlp.router.gate": {
74
+ "group_size": 64,
75
+ "bits": 8
76
+ },
77
+ "model.layers.11.mlp.router.gate": {
78
+ "group_size": 64,
79
+ "bits": 8
80
+ },
81
+ "model.layers.12.mlp.router.gate": {
82
+ "group_size": 64,
83
+ "bits": 8
84
+ },
85
+ "model.layers.13.mlp.router.gate": {
86
+ "group_size": 64,
87
+ "bits": 8
88
+ },
89
+ "model.layers.14.mlp.router.gate": {
90
+ "group_size": 64,
91
+ "bits": 8
92
+ },
93
+ "model.layers.15.mlp.router.gate": {
94
+ "group_size": 64,
95
+ "bits": 8
96
+ },
97
+ "model.layers.16.mlp.router.gate": {
98
+ "group_size": 64,
99
+ "bits": 8
100
+ },
101
+ "model.layers.17.mlp.router.gate": {
102
+ "group_size": 64,
103
+ "bits": 8
104
+ },
105
+ "model.layers.18.mlp.router.gate": {
106
+ "group_size": 64,
107
+ "bits": 8
108
+ },
109
+ "model.layers.19.mlp.router.gate": {
110
+ "group_size": 64,
111
+ "bits": 8
112
+ },
113
+ "model.layers.20.mlp.router.gate": {
114
+ "group_size": 64,
115
+ "bits": 8
116
+ },
117
+ "model.layers.21.mlp.router.gate": {
118
+ "group_size": 64,
119
+ "bits": 8
120
+ },
121
+ "model.layers.22.mlp.router.gate": {
122
+ "group_size": 64,
123
+ "bits": 8
124
+ },
125
+ "model.layers.23.mlp.router.gate": {
126
+ "group_size": 64,
127
+ "bits": 8
128
+ },
129
+ "model.layers.24.mlp.router.gate": {
130
+ "group_size": 64,
131
+ "bits": 8
132
+ },
133
+ "model.layers.25.mlp.router.gate": {
134
+ "group_size": 64,
135
+ "bits": 8
136
+ },
137
+ "model.layers.26.mlp.router.gate": {
138
+ "group_size": 64,
139
+ "bits": 8
140
+ },
141
+ "model.layers.27.mlp.router.gate": {
142
+ "group_size": 64,
143
+ "bits": 8
144
+ },
145
+ "model.layers.28.mlp.router.gate": {
146
+ "group_size": 64,
147
+ "bits": 8
148
+ },
149
+ "model.layers.29.mlp.router.gate": {
150
+ "group_size": 64,
151
+ "bits": 8
152
+ },
153
+ "model.layers.30.mlp.router.gate": {
154
+ "group_size": 64,
155
+ "bits": 8
156
+ },
157
+ "model.layers.31.mlp.router.gate": {
158
+ "group_size": 64,
159
+ "bits": 8
160
+ },
161
+ "model.layers.32.mlp.router.gate": {
162
+ "group_size": 64,
163
+ "bits": 8
164
+ },
165
+ "model.layers.33.mlp.router.gate": {
166
+ "group_size": 64,
167
+ "bits": 8
168
+ },
169
+ "model.layers.34.mlp.router.gate": {
170
+ "group_size": 64,
171
+ "bits": 8
172
+ },
173
+ "model.layers.35.mlp.router.gate": {
174
+ "group_size": 64,
175
+ "bits": 8
176
+ },
177
+ "model.layers.36.mlp.router.gate": {
178
+ "group_size": 64,
179
+ "bits": 8
180
+ },
181
+ "model.layers.37.mlp.router.gate": {
182
+ "group_size": 64,
183
+ "bits": 8
184
+ },
185
+ "model.layers.38.mlp.router.gate": {
186
+ "group_size": 64,
187
+ "bits": 8
188
+ },
189
+ "model.layers.39.mlp.router.gate": {
190
+ "group_size": 64,
191
+ "bits": 8
192
+ },
193
+ "model.layers.40.mlp.router.gate": {
194
+ "group_size": 64,
195
+ "bits": 8
196
+ },
197
+ "model.layers.41.mlp.router.gate": {
198
+ "group_size": 64,
199
+ "bits": 8
200
+ },
201
+ "model.layers.42.mlp.router.gate": {
202
+ "group_size": 64,
203
+ "bits": 8
204
+ },
205
+ "model.layers.43.mlp.router.gate": {
206
+ "group_size": 64,
207
+ "bits": 8
208
+ },
209
+ "model.layers.44.mlp.router.gate": {
210
+ "group_size": 64,
211
+ "bits": 8
212
+ },
213
+ "model.layers.45.mlp.router.gate": {
214
+ "group_size": 64,
215
+ "bits": 8
216
+ },
217
+ "model.layers.46.mlp.router.gate": {
218
+ "group_size": 64,
219
+ "bits": 8
220
+ },
221
+ "model.layers.47.mlp.router.gate": {
222
+ "group_size": 64,
223
+ "bits": 8
224
+ },
225
+ "model.layers.48.mlp.router.gate": {
226
+ "group_size": 64,
227
+ "bits": 8
228
+ },
229
+ "model.layers.49.mlp.router.gate": {
230
+ "group_size": 64,
231
+ "bits": 8
232
+ },
233
+ "model.layers.50.mlp.router.gate": {
234
+ "group_size": 64,
235
+ "bits": 8
236
+ },
237
+ "model.layers.51.mlp.router.gate": {
238
+ "group_size": 64,
239
+ "bits": 8
240
+ },
241
+ "model.layers.52.mlp.router.gate": {
242
+ "group_size": 64,
243
+ "bits": 8
244
+ },
245
+ "model.layers.53.mlp.router.gate": {
246
+ "group_size": 64,
247
+ "bits": 8
248
+ },
249
+ "model.layers.54.mlp.router.gate": {
250
+ "group_size": 64,
251
+ "bits": 8
252
+ },
253
+ "model.layers.55.mlp.router.gate": {
254
+ "group_size": 64,
255
+ "bits": 8
256
+ },
257
+ "model.layers.56.mlp.router.gate": {
258
+ "group_size": 64,
259
+ "bits": 8
260
+ },
261
+ "model.layers.57.mlp.router.gate": {
262
+ "group_size": 64,
263
+ "bits": 8
264
+ },
265
+ "model.layers.58.mlp.router.gate": {
266
+ "group_size": 64,
267
+ "bits": 8
268
+ },
269
+ "model.layers.59.mlp.router.gate": {
270
+ "group_size": 64,
271
+ "bits": 8
272
+ },
273
+ "model.layers.60.mlp.router.gate": {
274
+ "group_size": 64,
275
+ "bits": 8
276
+ },
277
+ "model.layers.61.mlp.router.gate": {
278
+ "group_size": 64,
279
+ "bits": 8
280
+ },
281
+ "model.layers.62.mlp.router.gate": {
282
+ "group_size": 64,
283
+ "bits": 8
284
+ },
285
+ "model.layers.63.mlp.router.gate": {
286
+ "group_size": 64,
287
+ "bits": 8
288
+ },
289
+ "model.layers.64.mlp.router.gate": {
290
+ "group_size": 64,
291
+ "bits": 8
292
+ },
293
+ "model.layers.65.mlp.router.gate": {
294
+ "group_size": 64,
295
+ "bits": 8
296
+ },
297
+ "model.layers.66.mlp.router.gate": {
298
+ "group_size": 64,
299
+ "bits": 8
300
+ },
301
+ "model.layers.67.mlp.router.gate": {
302
+ "group_size": 64,
303
+ "bits": 8
304
+ },
305
+ "model.layers.68.mlp.router.gate": {
306
+ "group_size": 64,
307
+ "bits": 8
308
+ },
309
+ "model.layers.69.mlp.router.gate": {
310
+ "group_size": 64,
311
+ "bits": 8
312
+ },
313
+ "model.layers.70.mlp.router.gate": {
314
+ "group_size": 64,
315
+ "bits": 8
316
+ },
317
+ "model.layers.71.mlp.router.gate": {
318
+ "group_size": 64,
319
+ "bits": 8
320
+ },
321
+ "model.layers.72.mlp.router.gate": {
322
+ "group_size": 64,
323
+ "bits": 8
324
+ },
325
+ "model.layers.73.mlp.router.gate": {
326
+ "group_size": 64,
327
+ "bits": 8
328
+ },
329
+ "model.layers.74.mlp.router.gate": {
330
+ "group_size": 64,
331
+ "bits": 8
332
+ },
333
+ "model.layers.75.mlp.router.gate": {
334
+ "group_size": 64,
335
+ "bits": 8
336
+ },
337
+ "model.layers.76.mlp.router.gate": {
338
+ "group_size": 64,
339
+ "bits": 8
340
+ },
341
+ "model.layers.77.mlp.router.gate": {
342
+ "group_size": 64,
343
+ "bits": 8
344
+ },
345
+ "model.layers.78.mlp.router.gate": {
346
+ "group_size": 64,
347
+ "bits": 8
348
+ },
349
+ "model.layers.79.mlp.router.gate": {
350
+ "group_size": 64,
351
+ "bits": 8
352
+ }
353
+ },
354
+ "quantization_config": {
355
+ "group_size": 64,
356
+ "bits": 4,
357
+ "mode": "affine",
358
+ "model.layers.1.mlp.router.gate": {
359
+ "group_size": 64,
360
+ "bits": 8
361
+ },
362
+ "model.layers.2.mlp.router.gate": {
363
+ "group_size": 64,
364
+ "bits": 8
365
+ },
366
+ "model.layers.3.mlp.router.gate": {
367
+ "group_size": 64,
368
+ "bits": 8
369
+ },
370
+ "model.layers.4.mlp.router.gate": {
371
+ "group_size": 64,
372
+ "bits": 8
373
+ },
374
+ "model.layers.5.mlp.router.gate": {
375
+ "group_size": 64,
376
+ "bits": 8
377
+ },
378
+ "model.layers.6.mlp.router.gate": {
379
+ "group_size": 64,
380
+ "bits": 8
381
+ },
382
+ "model.layers.7.mlp.router.gate": {
383
+ "group_size": 64,
384
+ "bits": 8
385
+ },
386
+ "model.layers.8.mlp.router.gate": {
387
+ "group_size": 64,
388
+ "bits": 8
389
+ },
390
+ "model.layers.9.mlp.router.gate": {
391
+ "group_size": 64,
392
+ "bits": 8
393
+ },
394
+ "model.layers.10.mlp.router.gate": {
395
+ "group_size": 64,
396
+ "bits": 8
397
+ },
398
+ "model.layers.11.mlp.router.gate": {
399
+ "group_size": 64,
400
+ "bits": 8
401
+ },
402
+ "model.layers.12.mlp.router.gate": {
403
+ "group_size": 64,
404
+ "bits": 8
405
+ },
406
+ "model.layers.13.mlp.router.gate": {
407
+ "group_size": 64,
408
+ "bits": 8
409
+ },
410
+ "model.layers.14.mlp.router.gate": {
411
+ "group_size": 64,
412
+ "bits": 8
413
+ },
414
+ "model.layers.15.mlp.router.gate": {
415
+ "group_size": 64,
416
+ "bits": 8
417
+ },
418
+ "model.layers.16.mlp.router.gate": {
419
+ "group_size": 64,
420
+ "bits": 8
421
+ },
422
+ "model.layers.17.mlp.router.gate": {
423
+ "group_size": 64,
424
+ "bits": 8
425
+ },
426
+ "model.layers.18.mlp.router.gate": {
427
+ "group_size": 64,
428
+ "bits": 8
429
+ },
430
+ "model.layers.19.mlp.router.gate": {
431
+ "group_size": 64,
432
+ "bits": 8
433
+ },
434
+ "model.layers.20.mlp.router.gate": {
435
+ "group_size": 64,
436
+ "bits": 8
437
+ },
438
+ "model.layers.21.mlp.router.gate": {
439
+ "group_size": 64,
440
+ "bits": 8
441
+ },
442
+ "model.layers.22.mlp.router.gate": {
443
+ "group_size": 64,
444
+ "bits": 8
445
+ },
446
+ "model.layers.23.mlp.router.gate": {
447
+ "group_size": 64,
448
+ "bits": 8
449
+ },
450
+ "model.layers.24.mlp.router.gate": {
451
+ "group_size": 64,
452
+ "bits": 8
453
+ },
454
+ "model.layers.25.mlp.router.gate": {
455
+ "group_size": 64,
456
+ "bits": 8
457
+ },
458
+ "model.layers.26.mlp.router.gate": {
459
+ "group_size": 64,
460
+ "bits": 8
461
+ },
462
+ "model.layers.27.mlp.router.gate": {
463
+ "group_size": 64,
464
+ "bits": 8
465
+ },
466
+ "model.layers.28.mlp.router.gate": {
467
+ "group_size": 64,
468
+ "bits": 8
469
+ },
470
+ "model.layers.29.mlp.router.gate": {
471
+ "group_size": 64,
472
+ "bits": 8
473
+ },
474
+ "model.layers.30.mlp.router.gate": {
475
+ "group_size": 64,
476
+ "bits": 8
477
+ },
478
+ "model.layers.31.mlp.router.gate": {
479
+ "group_size": 64,
480
+ "bits": 8
481
+ },
482
+ "model.layers.32.mlp.router.gate": {
483
+ "group_size": 64,
484
+ "bits": 8
485
+ },
486
+ "model.layers.33.mlp.router.gate": {
487
+ "group_size": 64,
488
+ "bits": 8
489
+ },
490
+ "model.layers.34.mlp.router.gate": {
491
+ "group_size": 64,
492
+ "bits": 8
493
+ },
494
+ "model.layers.35.mlp.router.gate": {
495
+ "group_size": 64,
496
+ "bits": 8
497
+ },
498
+ "model.layers.36.mlp.router.gate": {
499
+ "group_size": 64,
500
+ "bits": 8
501
+ },
502
+ "model.layers.37.mlp.router.gate": {
503
+ "group_size": 64,
504
+ "bits": 8
505
+ },
506
+ "model.layers.38.mlp.router.gate": {
507
+ "group_size": 64,
508
+ "bits": 8
509
+ },
510
+ "model.layers.39.mlp.router.gate": {
511
+ "group_size": 64,
512
+ "bits": 8
513
+ },
514
+ "model.layers.40.mlp.router.gate": {
515
+ "group_size": 64,
516
+ "bits": 8
517
+ },
518
+ "model.layers.41.mlp.router.gate": {
519
+ "group_size": 64,
520
+ "bits": 8
521
+ },
522
+ "model.layers.42.mlp.router.gate": {
523
+ "group_size": 64,
524
+ "bits": 8
525
+ },
526
+ "model.layers.43.mlp.router.gate": {
527
+ "group_size": 64,
528
+ "bits": 8
529
+ },
530
+ "model.layers.44.mlp.router.gate": {
531
+ "group_size": 64,
532
+ "bits": 8
533
+ },
534
+ "model.layers.45.mlp.router.gate": {
535
+ "group_size": 64,
536
+ "bits": 8
537
+ },
538
+ "model.layers.46.mlp.router.gate": {
539
+ "group_size": 64,
540
+ "bits": 8
541
+ },
542
+ "model.layers.47.mlp.router.gate": {
543
+ "group_size": 64,
544
+ "bits": 8
545
+ },
546
+ "model.layers.48.mlp.router.gate": {
547
+ "group_size": 64,
548
+ "bits": 8
549
+ },
550
+ "model.layers.49.mlp.router.gate": {
551
+ "group_size": 64,
552
+ "bits": 8
553
+ },
554
+ "model.layers.50.mlp.router.gate": {
555
+ "group_size": 64,
556
+ "bits": 8
557
+ },
558
+ "model.layers.51.mlp.router.gate": {
559
+ "group_size": 64,
560
+ "bits": 8
561
+ },
562
+ "model.layers.52.mlp.router.gate": {
563
+ "group_size": 64,
564
+ "bits": 8
565
+ },
566
+ "model.layers.53.mlp.router.gate": {
567
+ "group_size": 64,
568
+ "bits": 8
569
+ },
570
+ "model.layers.54.mlp.router.gate": {
571
+ "group_size": 64,
572
+ "bits": 8
573
+ },
574
+ "model.layers.55.mlp.router.gate": {
575
+ "group_size": 64,
576
+ "bits": 8
577
+ },
578
+ "model.layers.56.mlp.router.gate": {
579
+ "group_size": 64,
580
+ "bits": 8
581
+ },
582
+ "model.layers.57.mlp.router.gate": {
583
+ "group_size": 64,
584
+ "bits": 8
585
+ },
586
+ "model.layers.58.mlp.router.gate": {
587
+ "group_size": 64,
588
+ "bits": 8
589
+ },
590
+ "model.layers.59.mlp.router.gate": {
591
+ "group_size": 64,
592
+ "bits": 8
593
+ },
594
+ "model.layers.60.mlp.router.gate": {
595
+ "group_size": 64,
596
+ "bits": 8
597
+ },
598
+ "model.layers.61.mlp.router.gate": {
599
+ "group_size": 64,
600
+ "bits": 8
601
+ },
602
+ "model.layers.62.mlp.router.gate": {
603
+ "group_size": 64,
604
+ "bits": 8
605
+ },
606
+ "model.layers.63.mlp.router.gate": {
607
+ "group_size": 64,
608
+ "bits": 8
609
+ },
610
+ "model.layers.64.mlp.router.gate": {
611
+ "group_size": 64,
612
+ "bits": 8
613
+ },
614
+ "model.layers.65.mlp.router.gate": {
615
+ "group_size": 64,
616
+ "bits": 8
617
+ },
618
+ "model.layers.66.mlp.router.gate": {
619
+ "group_size": 64,
620
+ "bits": 8
621
+ },
622
+ "model.layers.67.mlp.router.gate": {
623
+ "group_size": 64,
624
+ "bits": 8
625
+ },
626
+ "model.layers.68.mlp.router.gate": {
627
+ "group_size": 64,
628
+ "bits": 8
629
+ },
630
+ "model.layers.69.mlp.router.gate": {
631
+ "group_size": 64,
632
+ "bits": 8
633
+ },
634
+ "model.layers.70.mlp.router.gate": {
635
+ "group_size": 64,
636
+ "bits": 8
637
+ },
638
+ "model.layers.71.mlp.router.gate": {
639
+ "group_size": 64,
640
+ "bits": 8
641
+ },
642
+ "model.layers.72.mlp.router.gate": {
643
+ "group_size": 64,
644
+ "bits": 8
645
+ },
646
+ "model.layers.73.mlp.router.gate": {
647
+ "group_size": 64,
648
+ "bits": 8
649
+ },
650
+ "model.layers.74.mlp.router.gate": {
651
+ "group_size": 64,
652
+ "bits": 8
653
+ },
654
+ "model.layers.75.mlp.router.gate": {
655
+ "group_size": 64,
656
+ "bits": 8
657
+ },
658
+ "model.layers.76.mlp.router.gate": {
659
+ "group_size": 64,
660
+ "bits": 8
661
+ },
662
+ "model.layers.77.mlp.router.gate": {
663
+ "group_size": 64,
664
+ "bits": 8
665
+ },
666
+ "model.layers.78.mlp.router.gate": {
667
+ "group_size": 64,
668
+ "bits": 8
669
+ },
670
+ "model.layers.79.mlp.router.gate": {
671
+ "group_size": 64,
672
+ "bits": 8
673
+ }
674
+ },
675
+ "rms_norm_eps": 1e-05,
676
+ "rope_parameters": {
677
+ "rope_theta": 11158840.0,
678
+ "rope_type": "default"
679
+ },
680
+ "route_norm": true,
681
+ "router_scaling_factor": 2.826,
682
+ "sep_token_id": 120007,
683
+ "tie_word_embeddings": false,
684
+ "transformers_version": "5.6.0",
685
+ "use_cache": true,
686
+ "use_grouped_mm": false,
687
+ "vocab_size": 120832
688
+ }
generation_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 120000,
3
+ "do_sample": true,
4
+ "eos_token_id": 120025,
5
+ "pad_token_id": 120002,
6
+ "temperature": 0.9,
7
+ "top_k": -1,
8
+ "top_p": 1,
9
+ "transformers_version": "5.6.0"
10
+ }
model-00001-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3cdf1fc52c99ca762739a40bdad9f4fc80a07a329c9f9933bc527f1284284a7
3
+ size 5319591251
model-00002-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d586afc124446c37e4c7fe8b3bdc93be71f897df00601724bd6a5042e4b13d2
3
+ size 4864224955
model-00003-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da2cd7a70e59c475de61404e50468267613877ad69787548125eba43667d53a7
3
+ size 4918165689
model-00004-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb7c9ab0944202f19a9a74691a673de0f8c2753bb33e294fc127c4bd9ec654c2
3
+ size 4864224960
model-00005-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fdcee8de10f7718388551c783e7ad4c89ff4d9fe622c21f00e5cae604231acf7
3
+ size 4864225040
model-00006-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf2df009d84fd2e17d2c2ec2b01b7d821b557a59232ac24f4f4cbe7485586eaa
3
+ size 4918165735
model-00007-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c43df9f1d7d72874b73e779def5c732376c0278476cb70b4092d97b3aa2bbc5
3
+ size 4864224956
model-00008-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfabaff58aea50089b3f383ae77a38fed049e9ada0933212cec6b7f718045e0d
3
+ size 4864225074
model-00009-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed23b3cd6f679e3ca5c904e31b13abc8235a19061f1a75c54c7d8e3f2ef62ea7
3
+ size 4918165767
model-00010-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dedeeec81e53c3f3851044024ae6ccd1121a54675a0c1eb9101ff3cacba1e41b
3
+ size 4864224972
model-00011-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:004e03a4a2f0c23628708ddd812ad21055300aa6014f5819674bda8b9ad87502
3
+ size 4864225022
model-00012-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:840cb1ebafcd0eb20f240346cd588c55ba0c600349935742f710615ab4a9e6ce
3
+ size 4918165825
model-00013-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be83d61db071d933a6bd7bb25b9d3f75c456e78200e9a774d1d72fb5db7e4682
3
+ size 4864225016
model-00014-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c1755fb4ad8e951eea9b3e59ed973d7e78044155010aa22ca132b1d5806e992
3
+ size 4864225000
model-00015-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28dab91860d804497cc4af970b2beb4eb1c0c4f4750585f477c8a586ba64e0d6
3
+ size 4918165779
model-00016-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24407218c575f257f9be3c0090888239d93e12ad14a12bc3dbf18f0966dc90cd
3
+ size 4864225006
model-00017-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0d7e24cdccf05c401256d3ae89b10854a73a39bd81f455307439f0b6e6ff5fc
3
+ size 4864225042
model-00018-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6038a0a4501e41bcc4814860107cd5c46fcc89b7d934a396205aff72ebf957c7
3
+ size 4918165767
model-00019-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:787bf8c5455cf94ae002c6a17a7d8117ee18ff07e5f67081eb3ce9a7f37016fd
3
+ size 4864224990
model-00020-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fad9a1a94f5efb9c749e1df367f2541a6b61f4dc0b168c7f36b33ef1808e34f3
3
+ size 4864225000
model-00021-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de2839beab09cb2a68b22ce6cfe7ed9ec1efb87660c91e99602368ad763918c8
3
+ size 4918165807
model-00022-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:467c8ffaaef5171c1dafda160eaa5829608569422cbb9e78413e0095fa4040ad
3
+ size 4864225044
model-00023-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdc6265db462d41c719bf56830e9a5c369d24c7d99d6e5c41d44b9c95673390a
3
+ size 4864225034
model-00024-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2c205270c049bf31fda52d51e2f669e5fc3da466d4fe7548084768e92c5fd2e
3
+ size 4918165781
model-00025-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32578d17cb599c1853d1e3301da7fde75e5c7cd043be48250c8b9cd2ec99c76e
3
+ size 4864225022
model-00026-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6868c14bcde362f17c9be5ae8817e3603da7da6dd8921e514ddcba31c47b1759
3
+ size 4864225082
model-00027-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:509d647b23611c2090a8c2d68c4877f06135785f0e09f506d076cb33dd0be0dd
3
+ size 4918165785
model-00028-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ac7c5d16fd88bfdc27ac487665872082c8d2816618c379c2734d70bb22c341e
3
+ size 4864225046
model-00029-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e4b69590c19a5c21c59fd4472fd9f62000e8d831de5aac2afefe5086d2a892c
3
+ size 4864225076
model-00030-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a75c759fcf65881aada483183223aad979964f5e69efef84da7add01c921b7b
3
+ size 4918165811
model-00031-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bf8e4aab4b0c6f76baa533c9386cf33950f1737c3f0d181d63f97bf1e25a6e1
3
+ size 4864225050
model-00032-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bd6e8f3ac93eace8e32b60cdb405398d4a2d2d5a61e356d32f8c149b1138080
3
+ size 4864225032
model-00033-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eaacd187f4e2db10328093ffddedd791bb068d1ede22f614efe8aef0659e320a
3
+ size 4918165797
model-00034-of-00034.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1341217be54b7b5ccb885ad5e45f0faa692cb526a41cb46b5ea5f4bf4c3420b
3
+ size 4420683411
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "bos_token": "<|hy_begin▁of▁sentence|>",
4
+ "clean_up_tokenization_spaces": false,
5
+ "eos_token": "<|hy_eos|>",
6
+ "is_local": true,
7
+ "model_max_length": 1000000000000000019884624838656,
8
+ "pad_token": "<|hy_▁pad▁|>",
9
+ "tokenizer_class": "TokenizersBackend",
10
+ "tool_parser_type": "hy_v3"
11
+ }