dylanlanigan commited on
Commit
79f8c7e
·
verified ·
1 Parent(s): 138f82e

Add files using upload-large-folder tool

Browse files
Files changed (2) hide show
  1. mlc-chat-config.json +9 -2
  2. mlc-chat-config.json.base +91 -0
mlc-chat-config.json CHANGED
@@ -12,9 +12,16 @@
12
  "num_key_value_heads": 4,
13
  "rms_norm_eps": 1e-06,
14
  "rope_theta": 1000000.0,
 
 
 
 
 
 
 
15
  "vocab_size": 151936,
16
  "tie_word_embeddings": false,
17
- "context_window_size": 40960,
18
  "prefill_chunk_size": 2048,
19
  "tensor_parallel_shards": 1,
20
  "head_dim": 128,
@@ -28,7 +35,7 @@
28
  "norm_topk_prob": true
29
  },
30
  "vocab_size": 151936,
31
- "context_window_size": 40960,
32
  "sliding_window_size": -1,
33
  "prefill_chunk_size": 2048,
34
  "attention_sink_size": -1,
 
12
  "num_key_value_heads": 4,
13
  "rms_norm_eps": 1e-06,
14
  "rope_theta": 1000000.0,
15
+ "rope_scaling": {
16
+ "factor": 4,
17
+ "mscale": 0.707,
18
+ "mscale_all_dim": 0.707,
19
+ "original_max_position_embeddings": 32768,
20
+ "type": "yarn"
21
+ },
22
  "vocab_size": 151936,
23
  "tie_word_embeddings": false,
24
+ "context_window_size": 131072,
25
  "prefill_chunk_size": 2048,
26
  "tensor_parallel_shards": 1,
27
  "head_dim": 128,
 
35
  "norm_topk_prob": true
36
  },
37
  "vocab_size": 151936,
38
+ "context_window_size": 131072,
39
  "sliding_window_size": -1,
40
  "prefill_chunk_size": 2048,
41
  "attention_sink_size": -1,
mlc-chat-config.json.base ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": "0.1.0",
3
+ "model_type": "qwen3_moe",
4
+ "quantization": "q4f16_1",
5
+ "model_config": {
6
+ "hidden_act": "silu",
7
+ "hidden_size": 2048,
8
+ "intermediate_size": 6144,
9
+ "attention_bias": false,
10
+ "num_attention_heads": 32,
11
+ "num_hidden_layers": 48,
12
+ "num_key_value_heads": 4,
13
+ "rms_norm_eps": 1e-06,
14
+ "rope_theta": 1000000.0,
15
+ "vocab_size": 151936,
16
+ "tie_word_embeddings": false,
17
+ "context_window_size": 40960,
18
+ "prefill_chunk_size": 2048,
19
+ "tensor_parallel_shards": 1,
20
+ "head_dim": 128,
21
+ "dtype": "float32",
22
+ "max_batch_size": 128,
23
+ "weight_block_size": null,
24
+ "moe_intermediate_size": 768,
25
+ "num_experts_per_tok": 8,
26
+ "num_experts": 128,
27
+ "decoder_sparse_step": 1,
28
+ "norm_topk_prob": true
29
+ },
30
+ "vocab_size": 151936,
31
+ "context_window_size": 40960,
32
+ "sliding_window_size": -1,
33
+ "prefill_chunk_size": 2048,
34
+ "attention_sink_size": -1,
35
+ "tensor_parallel_shards": 1,
36
+ "pipeline_parallel_stages": 1,
37
+ "temperature": 0.6,
38
+ "presence_penalty": 0.0,
39
+ "frequency_penalty": 0.0,
40
+ "repetition_penalty": 1.0,
41
+ "top_p": 0.95,
42
+ "tokenizer_files": [
43
+ "tokenizer.json",
44
+ "vocab.json",
45
+ "merges.txt",
46
+ "tokenizer_config.json"
47
+ ],
48
+ "tokenizer_info": {
49
+ "token_postproc_method": "byte_level",
50
+ "prepend_space_in_encode": false,
51
+ "strip_space_in_decode": false
52
+ },
53
+ "conv_template": {
54
+ "name": "qwen2",
55
+ "system_template": "<|im_start|>system\n{system_message}<|im_end|>\n",
56
+ "system_message": "You are a helpful assistant.",
57
+ "system_prefix_token_ids": null,
58
+ "add_role_after_system_message": true,
59
+ "roles": {
60
+ "user": "<|im_start|>user",
61
+ "assistant": "<|im_start|>assistant"
62
+ },
63
+ "role_templates": {
64
+ "user": "{user_message}",
65
+ "assistant": "{assistant_message}",
66
+ "tool": "{tool_message}"
67
+ },
68
+ "messages": [],
69
+ "seps": [
70
+ "<|im_end|>\n"
71
+ ],
72
+ "role_content_sep": "\n",
73
+ "role_empty_sep": "\n",
74
+ "stop_str": [
75
+ "<|endoftext|>",
76
+ "<|im_end|>"
77
+ ],
78
+ "stop_token_ids": [
79
+ 151643,
80
+ 151645
81
+ ],
82
+ "function_string": "",
83
+ "use_function_calling": false
84
+ },
85
+ "pad_token_id": 151643,
86
+ "bos_token_id": 151643,
87
+ "eos_token_id": [
88
+ 151645,
89
+ 151643
90
+ ]
91
+ }