kernelpool commited on
Commit
af0cbd6
·
verified ·
1 Parent(s): 0458a04

Add files using upload-large-folder tool

Browse files
Files changed (50) hide show
  1. README.md +36 -0
  2. chat_template.jinja +195 -0
  3. config.json +688 -0
  4. generation_config.json +10 -0
  5. model-00001-of-00048.safetensors +3 -0
  6. model-00002-of-00048.safetensors +3 -0
  7. model-00003-of-00048.safetensors +3 -0
  8. model-00004-of-00048.safetensors +3 -0
  9. model-00005-of-00048.safetensors +3 -0
  10. model-00006-of-00048.safetensors +3 -0
  11. model-00007-of-00048.safetensors +3 -0
  12. model-00008-of-00048.safetensors +3 -0
  13. model-00009-of-00048.safetensors +3 -0
  14. model-00010-of-00048.safetensors +3 -0
  15. model-00011-of-00048.safetensors +3 -0
  16. model-00012-of-00048.safetensors +3 -0
  17. model-00013-of-00048.safetensors +3 -0
  18. model-00014-of-00048.safetensors +3 -0
  19. model-00015-of-00048.safetensors +3 -0
  20. model-00016-of-00048.safetensors +3 -0
  21. model-00017-of-00048.safetensors +3 -0
  22. model-00018-of-00048.safetensors +3 -0
  23. model-00020-of-00048.safetensors +3 -0
  24. model-00021-of-00048.safetensors +3 -0
  25. model-00022-of-00048.safetensors +3 -0
  26. model-00023-of-00048.safetensors +3 -0
  27. model-00026-of-00048.safetensors +3 -0
  28. model-00027-of-00048.safetensors +3 -0
  29. model-00028-of-00048.safetensors +3 -0
  30. model-00029-of-00048.safetensors +3 -0
  31. model-00030-of-00048.safetensors +3 -0
  32. model-00031-of-00048.safetensors +3 -0
  33. model-00032-of-00048.safetensors +3 -0
  34. model-00033-of-00048.safetensors +3 -0
  35. model-00034-of-00048.safetensors +3 -0
  36. model-00035-of-00048.safetensors +3 -0
  37. model-00036-of-00048.safetensors +3 -0
  38. model-00039-of-00048.safetensors +3 -0
  39. model-00040-of-00048.safetensors +3 -0
  40. model-00041-of-00048.safetensors +3 -0
  41. model-00042-of-00048.safetensors +3 -0
  42. model-00043-of-00048.safetensors +3 -0
  43. model-00044-of-00048.safetensors +3 -0
  44. model-00045-of-00048.safetensors +3 -0
  45. model-00046-of-00048.safetensors +3 -0
  46. model-00047-of-00048.safetensors +3 -0
  47. model-00048-of-00048.safetensors +3 -0
  48. model.safetensors.index.json +0 -0
  49. tokenizer.json +0 -0
  50. tokenizer_config.json +11 -0
README.md ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ library_name: mlx
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - mlx
7
+ base_model: tencent/Hy3-preview
8
+ ---
9
+
10
+ # mlx-community/Hy3-preview-6bit
11
+
12
+ This model [mlx-community/Hy3-preview-6bit](https://huggingface.co/mlx-community/Hy3-preview-6bit) was
13
+ converted to MLX format from [tencent/Hy3-preview](https://huggingface.co/tencent/Hy3-preview)
14
+ using mlx-lm version **0.31.3**.
15
+
16
+ ## Use with mlx
17
+
18
+ ```bash
19
+ pip install mlx-lm
20
+ ```
21
+
22
+ ```python
23
+ from mlx_lm import load, generate
24
+
25
+ model, tokenizer = load("mlx-community/Hy3-preview-6bit")
26
+
27
+ prompt = "hello"
28
+
29
+ if tokenizer.chat_template is not None:
30
+ messages = [{"role": "user", "content": prompt}]
31
+ prompt = tokenizer.apply_chat_template(
32
+ messages, add_generation_prompt=True, return_dict=False,
33
+ )
34
+
35
+ response = generate(model, tokenizer, prompt=prompt, verbose=True)
36
+ ```
chat_template.jinja ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {#- ----------‑‑‑ special token variables ‑‑‑---------- -#}
2
+ {%- set bos_token = '<|hy_begin▁of▁sentence|>' %}
3
+ {%- set pad_token = '<|hy_▁pad▁|>' %}
4
+ {%- set user_token = '<|hy_User|>' %}
5
+ {%- set assistant_token = '<|hy_Assistant|>' %}
6
+ {%- set eos_token = '<|hy_eos|>' %}
7
+ {%- set think_begin_token = '<think>' %}
8
+ {%- set think_end_token = '</think>' %}
9
+ {%- set toolcalls_begin_token = '<tool_calls>' %}
10
+ {%- set toolcalls_end_token = '</tool_calls>' %}
11
+ {%- set toolcall_begin_token = '<tool_call>' %}
12
+ {%- set toolcall_end_token = '</tool_call>' %}
13
+ {%- set toolsep_token = '<tool_sep>' %}
14
+ {%- set argkey_begin_token = '<arg_key>' %}
15
+ {%- set argkey_end_token = '</arg_key>' %}
16
+ {%- set argvalue_begin_token = '<arg_value>' %}
17
+ {%- set argvalue_end_token = '</arg_value>' %}
18
+ {%- set toolresponses_begin_token = '<tool_responses>' %}
19
+ {%- set toolresponses_end_token = '</tool_responses>' %}
20
+ {%- set toolresponse_begin_token = '<tool_response>' %}
21
+ {%- set toolresponse_end_token = '</tool_response>' %}
22
+ {%- set reasoning_mode_token = '<|reasoning_mode|>' %}
23
+ {#- ----------‑‑‑ hyperparameters variables ‑‑‑---------- -#}
24
+ {%- if not add_generation_prompt is defined %}
25
+ {%- set add_generation_prompt = false %}
26
+ {%- endif %}
27
+ {%- if not interleaved_thinking is defined %}
28
+ {%- set interleaved_thinking = false %}
29
+ {%- endif %}
30
+ {%- if not tools %}
31
+ {%- set interleaved_thinking = false %}
32
+ {%- endif %}
33
+ {%- if not is_training is defined %}
34
+ {%- set is_training = false %}
35
+ {%- endif %}
36
+ {%- if not reasoning_effort is defined or reasoning_effort not in ['high', 'low', 'no_think'] %}
37
+ {%- set reasoning_effort = 'no_think' %}
38
+ {%- endif %}
39
+
40
+ {%- macro visible_text(content) -%}
41
+ {%- if content is string -%}
42
+ {{- content }}
43
+ {%- elif content is iterable and content is not mapping -%}
44
+ {%- for item in content -%}
45
+ {%- if item is mapping and item.type == 'text' -%}
46
+ {{- item.text }}
47
+ {%- elif item is string -%}
48
+ {{- item }}
49
+ {%- endif -%}
50
+ {%- endfor -%}
51
+ {%- elif content is none -%}
52
+ {{- '' }}
53
+ {%- else -%}
54
+ {{- content }}
55
+ {%- endif -%}
56
+ {%- endmacro -%}
57
+
58
+ {%- set ns = namespace(last_user_index=-1) %}
59
+ {%- set sp_ns = namespace(system_prompt='', is_first_sp=true) %}
60
+ {%- for message in messages %}
61
+ {%- if message['role'] == 'system' %}
62
+ {%- set sp_ns.system_prompt = sp_ns.system_prompt + visible_text(message['content']) %}
63
+ {%- endif %}
64
+ {%- if message['role'] == 'user' %}
65
+ {%- set ns.last_user_index = loop.index0 %}
66
+ {%- endif %}
67
+ {%- endfor %}
68
+ {%- if reasoning_effort is defined and reasoning_effort is string and reasoning_effort != '' and not tools %}
69
+ {%- set sp_ns.system_prompt = sp_ns.system_prompt + reasoning_mode_token + 'reasoning_effort:' + reasoning_effort %}
70
+ {%- endif %}
71
+ {{- bos_token }}
72
+ {{- sp_ns.system_prompt }}
73
+ {%- if tools %}
74
+ {%- if sp_ns.system_prompt != '' %}
75
+ {{- '\n\n# Tools\n\nYou may call one or more functions to assist with the user query.' }}
76
+ {%- else %}
77
+ {{- '# Tools\n\nYou may call one or more functions to assist with the user query.' }}
78
+ {%- endif %}
79
+ {{- '\n\nYou are provided with function signatures within <tools></tools> XML tags:' }}
80
+ {{- '\n<tools>\n' }}
81
+ {%- for tool in tools %}
82
+ {%- if loop.index0 > 0 %}
83
+ {{- '\n' }}
84
+ {%- endif %}
85
+ {{- tool | tojson }}
86
+ {%- endfor %}
87
+ {{- '\n</tools>\n\n' }}
88
+ {{- 'For function call returns, you should first print ' + toolcalls_begin_token + '\n' }}
89
+ {{- 'For each function call, you should return object like:\n' }}
90
+ {{- toolcall_begin_token + '{function-name}' + toolsep_token + '\n' }}
91
+ {{- argkey_begin_token + '{arg-key-1}' + argkey_end_token + '\n' }}
92
+ {{- argvalue_begin_token + '{arg-value-1}' + argvalue_end_token + '\n' }}
93
+ {{- argkey_begin_token + '{arg-key-2}' + argkey_end_token + '\n' }}
94
+ {{- argvalue_begin_token + '{arg-value-2}' + argvalue_end_token + '\n' }}
95
+ {{- '...\n' }}
96
+ {{- toolcall_end_token + '\n' }}
97
+ {%- if reasoning_effort is defined and reasoning_effort is string and reasoning_effort != '' %}
98
+ {{- 'At the end of function call returns, you should print ' + toolcalls_end_token + reasoning_mode_token + 'reasoning_effort:' + reasoning_effort }}
99
+ {%- else %}
100
+ {{- 'At the end of function call returns, you should print ' + toolcalls_end_token }}
101
+ {%- endif %}
102
+ {%- endif %}
103
+
104
+ {%- set prev_ns = namespace(is_tool=false, is_tool_first=true) %}
105
+ {%- set last_ns = namespace(last_is_assistant=false) %}
106
+ {%- for message in messages %}
107
+ {%- if message['role'] == 'user' %}
108
+ {%- if prev_ns.is_tool %}
109
+ {{- toolresponses_end_token }}
110
+ {%- endif %}
111
+ {{- user_token + visible_text(message['content']) }}
112
+ {%- set prev_ns.is_tool = false %}
113
+ {%- endif %}
114
+ {%- if message['role'] == 'assistant' %}
115
+ {%- if 'reasoning_content' in message and message['reasoning_content'] is string %}
116
+ {%- set rc = message['reasoning_content'] %}
117
+ {%- elif 'reasoning' in message and message['reasoning'] is string %}
118
+ {%- set rc = message['reasoning'] %}
119
+ {%- else %}
120
+ {%- set rc = none %}
121
+ {%- endif %}
122
+ {%- if is_training %}
123
+ {%- if rc is not none %}
124
+ {%- set content = think_begin_token + rc + think_end_token + visible_text(message['content']) %}
125
+ {%- else %}
126
+ {%- set content = think_begin_token + think_end_token + visible_text(message['content']) %}
127
+ {%- endif %}
128
+ {%- else %}
129
+ {%- if interleaved_thinking %}
130
+ {%- if loop.index0 > ns.last_user_index and rc is not none %}
131
+ {%- set content = think_begin_token + rc + think_end_token + visible_text(message['content']) %}
132
+ {%- else %}
133
+ {%- set content = think_begin_token + think_end_token + visible_text(message['content']) %}
134
+ {%- endif %}
135
+ {%- else %}
136
+ {%- set content = think_begin_token + think_end_token + visible_text(message['content']) %}
137
+ {%- endif %}
138
+ {%- endif %}
139
+ {%- if prev_ns.is_tool %}
140
+ {{- toolresponses_end_token }}
141
+ {%- endif %}
142
+ {{- assistant_token }}
143
+ {%- if message['tool_calls'] is defined and message['tool_calls'] %}
144
+ {%- set prev_ns.is_tool_first = true %}
145
+ {{- content }}
146
+ {{- toolcalls_begin_token + '\n' }}
147
+ {%- for tool in message['tool_calls'] %}
148
+ {%- set arguments = tool['function']['arguments'] %}
149
+ {{- toolcall_begin_token + tool['function']['name'] + toolsep_token + '\n' }}
150
+ {%- for key, value in arguments.items() %}
151
+ {{- argkey_begin_token + key + argkey_end_token + '\n' }}
152
+ {%- if value is not string %}
153
+ {%- set value = value | tojson(ensure_ascii=False) %}
154
+ {%- endif %}
155
+ {{- argvalue_begin_token + value + argvalue_end_token + '\n' }}
156
+ {%- endfor %}
157
+ {{- toolcall_end_token + '\n' }}
158
+ {%- endfor %}
159
+ {{- toolcalls_end_token + eos_token }}
160
+ {%- else %}
161
+ {%- if not loop.last or is_training %}
162
+ {{- content + eos_token }}
163
+ {%- else %}
164
+ {{- content }}
165
+ {%- endif %}
166
+ {%- endif %}
167
+ {%- set prev_ns.is_tool = false %}
168
+ {%- endif %}
169
+ {%- if message['role'] == 'tool' %}
170
+ {%- set prev_ns.is_tool = true %}
171
+ {%- if prev_ns.is_tool_first %}
172
+ {{- toolresponses_begin_token + '\n' }}
173
+ {%- set prev_ns.is_tool_first = false %}
174
+ {%- endif %}
175
+ {{- toolresponse_begin_token + '\n' + visible_text(message['content']) + '\n' + toolresponse_end_token + '\n' }}
176
+ {%- endif %}
177
+ {%- if loop.last and message['role'] == 'assistant' %}
178
+ {%- set last_ns.last_is_assistant = true %}
179
+ {%- endif %}
180
+
181
+ {%- endfor %}
182
+ {%- if prev_ns.is_tool %}
183
+ {{- toolresponses_end_token }}
184
+ {%- endif %}
185
+ {%- if add_generation_prompt %}
186
+ {%- if not last_ns.last_is_assistant %}
187
+ {%- if reasoning_effort is defined and reasoning_effort in ['low', 'high'] %}
188
+ {{- assistant_token + think_begin_token }}
189
+ {%- elif reasoning_effort is defined and reasoning_effort == 'no_think' %}
190
+ {{- assistant_token + think_begin_token + think_end_token }}
191
+ {%- else %}
192
+ {{- assistant_token }}
193
+ {%- endif %}
194
+ {%- endif %}
195
+ {%- endif %}
config.json ADDED
@@ -0,0 +1,688 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "HYV3ForCausalLM"
4
+ ],
5
+ "bos_token_id": 120000,
6
+ "enable_attention_fp32_softmax": false,
7
+ "enable_lm_head_fp32": true,
8
+ "enable_moe_fp32_combine": false,
9
+ "eod_token_id": 120026,
10
+ "eos_token_id": 120025,
11
+ "expert_hidden_dim": 1536,
12
+ "first_k_dense_replace": 1,
13
+ "head_dim": 128,
14
+ "hidden_act": "silu",
15
+ "hidden_size": 4096,
16
+ "initializer_range": 0.006,
17
+ "intermediate_size": 13312,
18
+ "max_position_embeddings": 262144,
19
+ "model_type": "hy_v3",
20
+ "moe_intermediate_size": 1536,
21
+ "moe_router_enable_expert_bias": true,
22
+ "moe_router_use_sigmoid": true,
23
+ "num_attention_heads": 64,
24
+ "num_experts": 192,
25
+ "num_experts_per_tok": 8,
26
+ "num_hidden_layers": 80,
27
+ "num_key_value_heads": 8,
28
+ "num_nextn_predict_layers": 1,
29
+ "num_shared_experts": 1,
30
+ "output_router_logits": true,
31
+ "pad_token_id": 120002,
32
+ "qk_norm": true,
33
+ "quantization": {
34
+ "group_size": 64,
35
+ "bits": 6,
36
+ "mode": "affine",
37
+ "model.layers.1.mlp.router.gate": {
38
+ "group_size": 64,
39
+ "bits": 8
40
+ },
41
+ "model.layers.2.mlp.router.gate": {
42
+ "group_size": 64,
43
+ "bits": 8
44
+ },
45
+ "model.layers.3.mlp.router.gate": {
46
+ "group_size": 64,
47
+ "bits": 8
48
+ },
49
+ "model.layers.4.mlp.router.gate": {
50
+ "group_size": 64,
51
+ "bits": 8
52
+ },
53
+ "model.layers.5.mlp.router.gate": {
54
+ "group_size": 64,
55
+ "bits": 8
56
+ },
57
+ "model.layers.6.mlp.router.gate": {
58
+ "group_size": 64,
59
+ "bits": 8
60
+ },
61
+ "model.layers.7.mlp.router.gate": {
62
+ "group_size": 64,
63
+ "bits": 8
64
+ },
65
+ "model.layers.8.mlp.router.gate": {
66
+ "group_size": 64,
67
+ "bits": 8
68
+ },
69
+ "model.layers.9.mlp.router.gate": {
70
+ "group_size": 64,
71
+ "bits": 8
72
+ },
73
+ "model.layers.10.mlp.router.gate": {
74
+ "group_size": 64,
75
+ "bits": 8
76
+ },
77
+ "model.layers.11.mlp.router.gate": {
78
+ "group_size": 64,
79
+ "bits": 8
80
+ },
81
+ "model.layers.12.mlp.router.gate": {
82
+ "group_size": 64,
83
+ "bits": 8
84
+ },
85
+ "model.layers.13.mlp.router.gate": {
86
+ "group_size": 64,
87
+ "bits": 8
88
+ },
89
+ "model.layers.14.mlp.router.gate": {
90
+ "group_size": 64,
91
+ "bits": 8
92
+ },
93
+ "model.layers.15.mlp.router.gate": {
94
+ "group_size": 64,
95
+ "bits": 8
96
+ },
97
+ "model.layers.16.mlp.router.gate": {
98
+ "group_size": 64,
99
+ "bits": 8
100
+ },
101
+ "model.layers.17.mlp.router.gate": {
102
+ "group_size": 64,
103
+ "bits": 8
104
+ },
105
+ "model.layers.18.mlp.router.gate": {
106
+ "group_size": 64,
107
+ "bits": 8
108
+ },
109
+ "model.layers.19.mlp.router.gate": {
110
+ "group_size": 64,
111
+ "bits": 8
112
+ },
113
+ "model.layers.20.mlp.router.gate": {
114
+ "group_size": 64,
115
+ "bits": 8
116
+ },
117
+ "model.layers.21.mlp.router.gate": {
118
+ "group_size": 64,
119
+ "bits": 8
120
+ },
121
+ "model.layers.22.mlp.router.gate": {
122
+ "group_size": 64,
123
+ "bits": 8
124
+ },
125
+ "model.layers.23.mlp.router.gate": {
126
+ "group_size": 64,
127
+ "bits": 8
128
+ },
129
+ "model.layers.24.mlp.router.gate": {
130
+ "group_size": 64,
131
+ "bits": 8
132
+ },
133
+ "model.layers.25.mlp.router.gate": {
134
+ "group_size": 64,
135
+ "bits": 8
136
+ },
137
+ "model.layers.26.mlp.router.gate": {
138
+ "group_size": 64,
139
+ "bits": 8
140
+ },
141
+ "model.layers.27.mlp.router.gate": {
142
+ "group_size": 64,
143
+ "bits": 8
144
+ },
145
+ "model.layers.28.mlp.router.gate": {
146
+ "group_size": 64,
147
+ "bits": 8
148
+ },
149
+ "model.layers.29.mlp.router.gate": {
150
+ "group_size": 64,
151
+ "bits": 8
152
+ },
153
+ "model.layers.30.mlp.router.gate": {
154
+ "group_size": 64,
155
+ "bits": 8
156
+ },
157
+ "model.layers.31.mlp.router.gate": {
158
+ "group_size": 64,
159
+ "bits": 8
160
+ },
161
+ "model.layers.32.mlp.router.gate": {
162
+ "group_size": 64,
163
+ "bits": 8
164
+ },
165
+ "model.layers.33.mlp.router.gate": {
166
+ "group_size": 64,
167
+ "bits": 8
168
+ },
169
+ "model.layers.34.mlp.router.gate": {
170
+ "group_size": 64,
171
+ "bits": 8
172
+ },
173
+ "model.layers.35.mlp.router.gate": {
174
+ "group_size": 64,
175
+ "bits": 8
176
+ },
177
+ "model.layers.36.mlp.router.gate": {
178
+ "group_size": 64,
179
+ "bits": 8
180
+ },
181
+ "model.layers.37.mlp.router.gate": {
182
+ "group_size": 64,
183
+ "bits": 8
184
+ },
185
+ "model.layers.38.mlp.router.gate": {
186
+ "group_size": 64,
187
+ "bits": 8
188
+ },
189
+ "model.layers.39.mlp.router.gate": {
190
+ "group_size": 64,
191
+ "bits": 8
192
+ },
193
+ "model.layers.40.mlp.router.gate": {
194
+ "group_size": 64,
195
+ "bits": 8
196
+ },
197
+ "model.layers.41.mlp.router.gate": {
198
+ "group_size": 64,
199
+ "bits": 8
200
+ },
201
+ "model.layers.42.mlp.router.gate": {
202
+ "group_size": 64,
203
+ "bits": 8
204
+ },
205
+ "model.layers.43.mlp.router.gate": {
206
+ "group_size": 64,
207
+ "bits": 8
208
+ },
209
+ "model.layers.44.mlp.router.gate": {
210
+ "group_size": 64,
211
+ "bits": 8
212
+ },
213
+ "model.layers.45.mlp.router.gate": {
214
+ "group_size": 64,
215
+ "bits": 8
216
+ },
217
+ "model.layers.46.mlp.router.gate": {
218
+ "group_size": 64,
219
+ "bits": 8
220
+ },
221
+ "model.layers.47.mlp.router.gate": {
222
+ "group_size": 64,
223
+ "bits": 8
224
+ },
225
+ "model.layers.48.mlp.router.gate": {
226
+ "group_size": 64,
227
+ "bits": 8
228
+ },
229
+ "model.layers.49.mlp.router.gate": {
230
+ "group_size": 64,
231
+ "bits": 8
232
+ },
233
+ "model.layers.50.mlp.router.gate": {
234
+ "group_size": 64,
235
+ "bits": 8
236
+ },
237
+ "model.layers.51.mlp.router.gate": {
238
+ "group_size": 64,
239
+ "bits": 8
240
+ },
241
+ "model.layers.52.mlp.router.gate": {
242
+ "group_size": 64,
243
+ "bits": 8
244
+ },
245
+ "model.layers.53.mlp.router.gate": {
246
+ "group_size": 64,
247
+ "bits": 8
248
+ },
249
+ "model.layers.54.mlp.router.gate": {
250
+ "group_size": 64,
251
+ "bits": 8
252
+ },
253
+ "model.layers.55.mlp.router.gate": {
254
+ "group_size": 64,
255
+ "bits": 8
256
+ },
257
+ "model.layers.56.mlp.router.gate": {
258
+ "group_size": 64,
259
+ "bits": 8
260
+ },
261
+ "model.layers.57.mlp.router.gate": {
262
+ "group_size": 64,
263
+ "bits": 8
264
+ },
265
+ "model.layers.58.mlp.router.gate": {
266
+ "group_size": 64,
267
+ "bits": 8
268
+ },
269
+ "model.layers.59.mlp.router.gate": {
270
+ "group_size": 64,
271
+ "bits": 8
272
+ },
273
+ "model.layers.60.mlp.router.gate": {
274
+ "group_size": 64,
275
+ "bits": 8
276
+ },
277
+ "model.layers.61.mlp.router.gate": {
278
+ "group_size": 64,
279
+ "bits": 8
280
+ },
281
+ "model.layers.62.mlp.router.gate": {
282
+ "group_size": 64,
283
+ "bits": 8
284
+ },
285
+ "model.layers.63.mlp.router.gate": {
286
+ "group_size": 64,
287
+ "bits": 8
288
+ },
289
+ "model.layers.64.mlp.router.gate": {
290
+ "group_size": 64,
291
+ "bits": 8
292
+ },
293
+ "model.layers.65.mlp.router.gate": {
294
+ "group_size": 64,
295
+ "bits": 8
296
+ },
297
+ "model.layers.66.mlp.router.gate": {
298
+ "group_size": 64,
299
+ "bits": 8
300
+ },
301
+ "model.layers.67.mlp.router.gate": {
302
+ "group_size": 64,
303
+ "bits": 8
304
+ },
305
+ "model.layers.68.mlp.router.gate": {
306
+ "group_size": 64,
307
+ "bits": 8
308
+ },
309
+ "model.layers.69.mlp.router.gate": {
310
+ "group_size": 64,
311
+ "bits": 8
312
+ },
313
+ "model.layers.70.mlp.router.gate": {
314
+ "group_size": 64,
315
+ "bits": 8
316
+ },
317
+ "model.layers.71.mlp.router.gate": {
318
+ "group_size": 64,
319
+ "bits": 8
320
+ },
321
+ "model.layers.72.mlp.router.gate": {
322
+ "group_size": 64,
323
+ "bits": 8
324
+ },
325
+ "model.layers.73.mlp.router.gate": {
326
+ "group_size": 64,
327
+ "bits": 8
328
+ },
329
+ "model.layers.74.mlp.router.gate": {
330
+ "group_size": 64,
331
+ "bits": 8
332
+ },
333
+ "model.layers.75.mlp.router.gate": {
334
+ "group_size": 64,
335
+ "bits": 8
336
+ },
337
+ "model.layers.76.mlp.router.gate": {
338
+ "group_size": 64,
339
+ "bits": 8
340
+ },
341
+ "model.layers.77.mlp.router.gate": {
342
+ "group_size": 64,
343
+ "bits": 8
344
+ },
345
+ "model.layers.78.mlp.router.gate": {
346
+ "group_size": 64,
347
+ "bits": 8
348
+ },
349
+ "model.layers.79.mlp.router.gate": {
350
+ "group_size": 64,
351
+ "bits": 8
352
+ }
353
+ },
354
+ "quantization_config": {
355
+ "group_size": 64,
356
+ "bits": 6,
357
+ "mode": "affine",
358
+ "model.layers.1.mlp.router.gate": {
359
+ "group_size": 64,
360
+ "bits": 8
361
+ },
362
+ "model.layers.2.mlp.router.gate": {
363
+ "group_size": 64,
364
+ "bits": 8
365
+ },
366
+ "model.layers.3.mlp.router.gate": {
367
+ "group_size": 64,
368
+ "bits": 8
369
+ },
370
+ "model.layers.4.mlp.router.gate": {
371
+ "group_size": 64,
372
+ "bits": 8
373
+ },
374
+ "model.layers.5.mlp.router.gate": {
375
+ "group_size": 64,
376
+ "bits": 8
377
+ },
378
+ "model.layers.6.mlp.router.gate": {
379
+ "group_size": 64,
380
+ "bits": 8
381
+ },
382
+ "model.layers.7.mlp.router.gate": {
383
+ "group_size": 64,
384
+ "bits": 8
385
+ },
386
+ "model.layers.8.mlp.router.gate": {
387
+ "group_size": 64,
388
+ "bits": 8
389
+ },
390
+ "model.layers.9.mlp.router.gate": {
391
+ "group_size": 64,
392
+ "bits": 8
393
+ },
394
+ "model.layers.10.mlp.router.gate": {
395
+ "group_size": 64,
396
+ "bits": 8
397
+ },
398
+ "model.layers.11.mlp.router.gate": {
399
+ "group_size": 64,
400
+ "bits": 8
401
+ },
402
+ "model.layers.12.mlp.router.gate": {
403
+ "group_size": 64,
404
+ "bits": 8
405
+ },
406
+ "model.layers.13.mlp.router.gate": {
407
+ "group_size": 64,
408
+ "bits": 8
409
+ },
410
+ "model.layers.14.mlp.router.gate": {
411
+ "group_size": 64,
412
+ "bits": 8
413
+ },
414
+ "model.layers.15.mlp.router.gate": {
415
+ "group_size": 64,
416
+ "bits": 8
417
+ },
418
+ "model.layers.16.mlp.router.gate": {
419
+ "group_size": 64,
420
+ "bits": 8
421
+ },
422
+ "model.layers.17.mlp.router.gate": {
423
+ "group_size": 64,
424
+ "bits": 8
425
+ },
426
+ "model.layers.18.mlp.router.gate": {
427
+ "group_size": 64,
428
+ "bits": 8
429
+ },
430
+ "model.layers.19.mlp.router.gate": {
431
+ "group_size": 64,
432
+ "bits": 8
433
+ },
434
+ "model.layers.20.mlp.router.gate": {
435
+ "group_size": 64,
436
+ "bits": 8
437
+ },
438
+ "model.layers.21.mlp.router.gate": {
439
+ "group_size": 64,
440
+ "bits": 8
441
+ },
442
+ "model.layers.22.mlp.router.gate": {
443
+ "group_size": 64,
444
+ "bits": 8
445
+ },
446
+ "model.layers.23.mlp.router.gate": {
447
+ "group_size": 64,
448
+ "bits": 8
449
+ },
450
+ "model.layers.24.mlp.router.gate": {
451
+ "group_size": 64,
452
+ "bits": 8
453
+ },
454
+ "model.layers.25.mlp.router.gate": {
455
+ "group_size": 64,
456
+ "bits": 8
457
+ },
458
+ "model.layers.26.mlp.router.gate": {
459
+ "group_size": 64,
460
+ "bits": 8
461
+ },
462
+ "model.layers.27.mlp.router.gate": {
463
+ "group_size": 64,
464
+ "bits": 8
465
+ },
466
+ "model.layers.28.mlp.router.gate": {
467
+ "group_size": 64,
468
+ "bits": 8
469
+ },
470
+ "model.layers.29.mlp.router.gate": {
471
+ "group_size": 64,
472
+ "bits": 8
473
+ },
474
+ "model.layers.30.mlp.router.gate": {
475
+ "group_size": 64,
476
+ "bits": 8
477
+ },
478
+ "model.layers.31.mlp.router.gate": {
479
+ "group_size": 64,
480
+ "bits": 8
481
+ },
482
+ "model.layers.32.mlp.router.gate": {
483
+ "group_size": 64,
484
+ "bits": 8
485
+ },
486
+ "model.layers.33.mlp.router.gate": {
487
+ "group_size": 64,
488
+ "bits": 8
489
+ },
490
+ "model.layers.34.mlp.router.gate": {
491
+ "group_size": 64,
492
+ "bits": 8
493
+ },
494
+ "model.layers.35.mlp.router.gate": {
495
+ "group_size": 64,
496
+ "bits": 8
497
+ },
498
+ "model.layers.36.mlp.router.gate": {
499
+ "group_size": 64,
500
+ "bits": 8
501
+ },
502
+ "model.layers.37.mlp.router.gate": {
503
+ "group_size": 64,
504
+ "bits": 8
505
+ },
506
+ "model.layers.38.mlp.router.gate": {
507
+ "group_size": 64,
508
+ "bits": 8
509
+ },
510
+ "model.layers.39.mlp.router.gate": {
511
+ "group_size": 64,
512
+ "bits": 8
513
+ },
514
+ "model.layers.40.mlp.router.gate": {
515
+ "group_size": 64,
516
+ "bits": 8
517
+ },
518
+ "model.layers.41.mlp.router.gate": {
519
+ "group_size": 64,
520
+ "bits": 8
521
+ },
522
+ "model.layers.42.mlp.router.gate": {
523
+ "group_size": 64,
524
+ "bits": 8
525
+ },
526
+ "model.layers.43.mlp.router.gate": {
527
+ "group_size": 64,
528
+ "bits": 8
529
+ },
530
+ "model.layers.44.mlp.router.gate": {
531
+ "group_size": 64,
532
+ "bits": 8
533
+ },
534
+ "model.layers.45.mlp.router.gate": {
535
+ "group_size": 64,
536
+ "bits": 8
537
+ },
538
+ "model.layers.46.mlp.router.gate": {
539
+ "group_size": 64,
540
+ "bits": 8
541
+ },
542
+ "model.layers.47.mlp.router.gate": {
543
+ "group_size": 64,
544
+ "bits": 8
545
+ },
546
+ "model.layers.48.mlp.router.gate": {
547
+ "group_size": 64,
548
+ "bits": 8
549
+ },
550
+ "model.layers.49.mlp.router.gate": {
551
+ "group_size": 64,
552
+ "bits": 8
553
+ },
554
+ "model.layers.50.mlp.router.gate": {
555
+ "group_size": 64,
556
+ "bits": 8
557
+ },
558
+ "model.layers.51.mlp.router.gate": {
559
+ "group_size": 64,
560
+ "bits": 8
561
+ },
562
+ "model.layers.52.mlp.router.gate": {
563
+ "group_size": 64,
564
+ "bits": 8
565
+ },
566
+ "model.layers.53.mlp.router.gate": {
567
+ "group_size": 64,
568
+ "bits": 8
569
+ },
570
+ "model.layers.54.mlp.router.gate": {
571
+ "group_size": 64,
572
+ "bits": 8
573
+ },
574
+ "model.layers.55.mlp.router.gate": {
575
+ "group_size": 64,
576
+ "bits": 8
577
+ },
578
+ "model.layers.56.mlp.router.gate": {
579
+ "group_size": 64,
580
+ "bits": 8
581
+ },
582
+ "model.layers.57.mlp.router.gate": {
583
+ "group_size": 64,
584
+ "bits": 8
585
+ },
586
+ "model.layers.58.mlp.router.gate": {
587
+ "group_size": 64,
588
+ "bits": 8
589
+ },
590
+ "model.layers.59.mlp.router.gate": {
591
+ "group_size": 64,
592
+ "bits": 8
593
+ },
594
+ "model.layers.60.mlp.router.gate": {
595
+ "group_size": 64,
596
+ "bits": 8
597
+ },
598
+ "model.layers.61.mlp.router.gate": {
599
+ "group_size": 64,
600
+ "bits": 8
601
+ },
602
+ "model.layers.62.mlp.router.gate": {
603
+ "group_size": 64,
604
+ "bits": 8
605
+ },
606
+ "model.layers.63.mlp.router.gate": {
607
+ "group_size": 64,
608
+ "bits": 8
609
+ },
610
+ "model.layers.64.mlp.router.gate": {
611
+ "group_size": 64,
612
+ "bits": 8
613
+ },
614
+ "model.layers.65.mlp.router.gate": {
615
+ "group_size": 64,
616
+ "bits": 8
617
+ },
618
+ "model.layers.66.mlp.router.gate": {
619
+ "group_size": 64,
620
+ "bits": 8
621
+ },
622
+ "model.layers.67.mlp.router.gate": {
623
+ "group_size": 64,
624
+ "bits": 8
625
+ },
626
+ "model.layers.68.mlp.router.gate": {
627
+ "group_size": 64,
628
+ "bits": 8
629
+ },
630
+ "model.layers.69.mlp.router.gate": {
631
+ "group_size": 64,
632
+ "bits": 8
633
+ },
634
+ "model.layers.70.mlp.router.gate": {
635
+ "group_size": 64,
636
+ "bits": 8
637
+ },
638
+ "model.layers.71.mlp.router.gate": {
639
+ "group_size": 64,
640
+ "bits": 8
641
+ },
642
+ "model.layers.72.mlp.router.gate": {
643
+ "group_size": 64,
644
+ "bits": 8
645
+ },
646
+ "model.layers.73.mlp.router.gate": {
647
+ "group_size": 64,
648
+ "bits": 8
649
+ },
650
+ "model.layers.74.mlp.router.gate": {
651
+ "group_size": 64,
652
+ "bits": 8
653
+ },
654
+ "model.layers.75.mlp.router.gate": {
655
+ "group_size": 64,
656
+ "bits": 8
657
+ },
658
+ "model.layers.76.mlp.router.gate": {
659
+ "group_size": 64,
660
+ "bits": 8
661
+ },
662
+ "model.layers.77.mlp.router.gate": {
663
+ "group_size": 64,
664
+ "bits": 8
665
+ },
666
+ "model.layers.78.mlp.router.gate": {
667
+ "group_size": 64,
668
+ "bits": 8
669
+ },
670
+ "model.layers.79.mlp.router.gate": {
671
+ "group_size": 64,
672
+ "bits": 8
673
+ }
674
+ },
675
+ "rms_norm_eps": 1e-05,
676
+ "rope_parameters": {
677
+ "rope_theta": 11158840.0,
678
+ "rope_type": "default"
679
+ },
680
+ "route_norm": true,
681
+ "router_scaling_factor": 2.826,
682
+ "sep_token_id": 120007,
683
+ "tie_word_embeddings": false,
684
+ "transformers_version": "5.6.0",
685
+ "use_cache": true,
686
+ "use_grouped_mm": false,
687
+ "vocab_size": 120832
688
+ }
generation_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 120000,
3
+ "do_sample": true,
4
+ "eos_token_id": 120025,
5
+ "pad_token_id": 120002,
6
+ "temperature": 0.9,
7
+ "top_k": -1,
8
+ "top_p": 1,
9
+ "transformers_version": "5.6.0"
10
+ }
model-00001-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26a3f771f8a6e1a4475c31b4dc09ccf86b11dbf7dbd41ba9203d2d64c0fb34df
3
+ size 4661145304
model-00002-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76e9a80d0b353f2ab4aa4240779be2e00e3f008c92209a00e0a16b80a572579b
3
+ size 5062405029
model-00003-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60dfeffea988c883b84766531013aba0fd01fa58fd26036f9943723a4dda58e4
3
+ size 4984871347
model-00004-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bce140fa5fb9dc918f915b60f123d4be6a570073c4f4454261157918b47bce68
3
+ size 5062405055
model-00005-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b487f25a3e1cf8b819c29b1d9272a2b2e2ad5a84c9557d7e25308d031e0239e
3
+ size 5062405021
model-00006-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4c3c4afab0293982e7373322ce04678330723c223c969988a0ca9e1a3679ff3
3
+ size 4984871367
model-00007-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:378c489fc46beb44bfaabdc661acbb728695b9e0d00d946ad686e7107318b319
3
+ size 5062405126
model-00008-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f11e63ce55c58607cd35376618327a4660091871877e725fac3e5c472f20f4d6
3
+ size 5062405062
model-00009-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78ac0fa4561df8e3e0cd1a66a74ab3f994b322896f1a65be1ba926ea9e185c85
3
+ size 4984871379
model-00010-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:920f46d601dd3df0aaaecdddf2543ce1c6b5f71f7cab4aab3c041de21840bff2
3
+ size 5062405146
model-00011-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd2eeaf6597ce05456cd1041d855aac792856c4521a705b3622a645dc6aae574
3
+ size 5062405050
model-00012-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6eb7a997d06b957fd5f6dcd08eef2af3719787d05e782cc5d3601562c40001d9
3
+ size 4984871397
model-00013-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c502931ee29377a555f41a4d906e5184c911244a24d724d20cbee5a3a84f2010
3
+ size 5062405132
model-00014-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:930d129a43a12f53394b0b66da8cb74818ef975aa8e2f61161fbe54a6e43a35e
3
+ size 5062405072
model-00015-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5260a7bfaea3dfc531d3334ff41b64265b4c317da5b4ef7bf04bfa5442b8de2e
3
+ size 4984871395
model-00016-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a3022d2a47bd68acf30200a167c13edaf779c59f045b9214e42ac7a0a05838e
3
+ size 5062405148
model-00017-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c193345ca0c7e8af6b3509c426de5ea4dac4bf511691043d7001a2f99edfe484
3
+ size 5062405116
model-00018-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff45ca068ac95c3fa158cd1c5671086a8d24025e36a0fda701107112213e9528
3
+ size 4984871349
model-00020-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77e4a9a6f726bd811b629961e94108d265f4920d6928ee1fae355adb676f0149
3
+ size 5062405108
model-00021-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32b354b7f324ff82d44897e8601c0c1c05ca745e4cce3be4d8a7e05c414ccf49
3
+ size 4984871391
model-00022-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:979fd5feda8147b79002a0ab02ba7973bb457ac907e61ff863104e8e689dc08f
3
+ size 5062405178
model-00023-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:130852b3bcf3c1e7eb08211ce3039e4ec57f07fe99ce1a58a85f54f06096aff4
3
+ size 5062405072
model-00026-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fe1ad000ed5365b6cd91bd37ed707df024c546a4ec1a40a48031dc657fa2a46
3
+ size 5062405100
model-00027-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e34810cbba394379ef0030e003409ca6c7424cc57e03eaf27cec5a2f4a59970f
3
+ size 4984871375
model-00028-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45b5ad2b43c3acd24325a072f7175237d6eb94f9a5f4cfd0698187bba28b05fb
3
+ size 5062405088
model-00029-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2890dc24343ea418e83fcbdc33e7c56442e61771a5270800d418a7fd79c4a9ff
3
+ size 5062405106
model-00030-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a030fe22adcb1d52e8930c54e3ac64a611aa20600d69d2ff1af3cf9cf268fd5
3
+ size 4984871393
model-00031-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d16800c689012aba6fce040955f531dc8b782f6ab9df8aca91c1beb3019fd7d8
3
+ size 5062405160
model-00032-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58c7f31af6b912d3b3431a06089db42cc3f871b54b1f432c318ec2100b8ddcfe
3
+ size 5062405070
model-00033-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:859329d92b68e5760546d7009b8267b95e687ef80eb7d65126c8c874f5cd3cbe
3
+ size 4984871393
model-00034-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f78863853796cffcad0de010f7db63121d85301c080e9fa24b971c70dcd4a470
3
+ size 5062405148
model-00035-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed0f1eeee2043e3ad579fce48baabea2a79d1a705db824718ae4801b998de028
3
+ size 5062405064
model-00036-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d71b99d5baea2c75ce516ce53d1fa43f6fe7bc0d3b45900e297b9efca58622d
3
+ size 4984871395
model-00039-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d945ebe940a4c1a9d9f11e94cced8e80956470b5e3d8583ce8772bf60e9ccdf9
3
+ size 4984871377
model-00040-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbd44751a3cb1d39e595b6783b31425db8772237a1dc5da346e04dfff30b92a7
3
+ size 5062405158
model-00041-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:472f212be22480b89531eb8733c594d38f6f5360048359e092670cbcd88494dc
3
+ size 5062405060
model-00042-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:837c83c332bfa171d8290f9bfd9044ca6962b69af33b65eb7d4c239e2df9c98e
3
+ size 4984871349
model-00043-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8323916cb122c29d9127401ca03fe55ec008868a5cfcb710d8c00dfba2a79ea9
3
+ size 5062405144
model-00044-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4030cdbd12a30fe462148ad9e9cd176f516859db7d25f5ab2554874b3a0cb188
3
+ size 5062405096
model-00045-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efa7bd314717e436640365aa4603f65d58bae0d1e4dc38d956aed4792611a486
3
+ size 4984871375
model-00046-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3443598d3ac1559a3c60b29ecc0ea05a12af7c9c7614854c886208cc2e96d8e
3
+ size 5062405160
model-00047-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f30f4b76cb6fe925543b3dba291683c931824b31efbb1e695392b7c594e550c2
3
+ size 5062405106
model-00048-of-00048.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63b569314f8b5cbcb9f4b47b389e8f6befd37ff42e82b94a161967213f9dbc16
3
+ size 3362729986
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "bos_token": "<|hy_begin▁of▁sentence|>",
4
+ "clean_up_tokenization_spaces": false,
5
+ "eos_token": "<|hy_eos|>",
6
+ "is_local": true,
7
+ "model_max_length": 1000000000000000019884624838656,
8
+ "pad_token": "<|hy_▁pad▁|>",
9
+ "tokenizer_class": "TokenizersBackend",
10
+ "tool_parser_type": "hy_v3"
11
+ }