Osaurus-AI commited on
Commit
cec83e9
·
verified ·
1 Parent(s): 9a4bae5

Upload jang_config.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. jang_config.json +73 -0
jang_config.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "weight_format": "mxtq",
3
+ "profile": "JANGTQ_V3",
4
+ "variant": "V3",
5
+ "mxtq_seed": 42,
6
+ "drop_mtp": true,
7
+ "source_model": "/Users/eric/models/Sources/DeepSeek-V4-Flash",
8
+ "source_config": {
9
+ "n_routed_experts": 256,
10
+ "num_hidden_layers": 43,
11
+ "n_hash_layers": 3
12
+ },
13
+ "mxtq_bits": {
14
+ "routed_expert": 2,
15
+ "attention": 8,
16
+ "shared_expert": 8,
17
+ "compressor": 8,
18
+ "indexer": 8,
19
+ "embed_tokens": 8,
20
+ "lm_head": 8,
21
+ "norms_router_hc": 16
22
+ },
23
+ "model_family": "deepseek_v4",
24
+ "chat": {
25
+ "encoder": "encoding_dsv4",
26
+ "encoder_fn": "encode_messages",
27
+ "chat_template_source": "builtin_encoding_module",
28
+ "has_tokenizer_chat_template": false,
29
+ "bos_token": "<\uff5cbegin\u2581of\u2581sentence\uff5c>",
30
+ "eos_token": "<\uff5cend\u2581of\u2581sentence\uff5c>",
31
+ "bos_token_id": 0,
32
+ "eos_token_id": 1,
33
+ "role_tokens": {
34
+ "user": "<\uff5cUser\uff5c>",
35
+ "assistant": "<\uff5cAssistant\uff5c>",
36
+ "latest_reminder": "<\uff5clatest_reminder\uff5c>"
37
+ },
38
+ "reasoning": {
39
+ "supported": true,
40
+ "modes": [
41
+ "chat",
42
+ "thinking"
43
+ ],
44
+ "default_mode": "chat",
45
+ "thinking_start": "<think>",
46
+ "thinking_end": "</think>",
47
+ "reasoning_effort_levels": [
48
+ "max",
49
+ "high",
50
+ null
51
+ ],
52
+ "drop_earlier_reasoning": true
53
+ },
54
+ "tool_calling": {
55
+ "supported": true,
56
+ "parser": "dsml",
57
+ "dsml_token": "\uff5cDSML\uff5c",
58
+ "tool_calls_block": "tool_calls",
59
+ "invoke_block": "invoke",
60
+ "parameter_block": "parameter",
61
+ "tool_output_tag": "tool_result"
62
+ },
63
+ "sampling_defaults": {
64
+ "temperature": 0.6,
65
+ "top_p": 0.95,
66
+ "repetition_penalty": 1.0,
67
+ "repetition_penalty_thinking": 1.0,
68
+ "repetition_penalty_chat": 1.05,
69
+ "max_new_tokens": 4096
70
+ }
71
+ },
72
+ "routed_expert_layout": "prestacked"
73
+ }