semran1 commited on
Commit
fc6232e
·
verified ·
1 Parent(s): 26fd4d7

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
added_tokens.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</think>": 151668,
3
+ "</tool_call>": 151658,
4
+ "</tool_response>": 151666,
5
+ "<think>": 151667,
6
+ "<tool_call>": 151657,
7
+ "<tool_response>": 151665,
8
+ "<|box_end|>": 151649,
9
+ "<|box_start|>": 151648,
10
+ "<|endoftext|>": 151643,
11
+ "<|file_sep|>": 151664,
12
+ "<|fim_middle|>": 151660,
13
+ "<|fim_pad|>": 151662,
14
+ "<|fim_prefix|>": 151659,
15
+ "<|fim_suffix|>": 151661,
16
+ "<|im_end|>": 151645,
17
+ "<|im_start|>": 151644,
18
+ "<|image_pad|>": 151655,
19
+ "<|object_ref_end|>": 151647,
20
+ "<|object_ref_start|>": 151646,
21
+ "<|quad_end|>": 151651,
22
+ "<|quad_start|>": 151650,
23
+ "<|repo_name|>": 151663,
24
+ "<|video_pad|>": 151656,
25
+ "<|vision_end|>": 151653,
26
+ "<|vision_pad|>": 151654,
27
+ "<|vision_start|>": 151652
28
+ }
chat_template.jinja ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0].role == 'system' %}
4
+ {{- messages[0].content + '\n\n' }}
5
+ {%- endif %}
6
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
7
+ {%- for tool in tools %}
8
+ {{- "\n" }}
9
+ {{- tool | tojson }}
10
+ {%- endfor %}
11
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
12
+ {%- else %}
13
+ {%- if messages[0].role == 'system' %}
14
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
+ {%- endif %}
16
+ {%- endif %}
17
+ {%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
18
+ {%- for message in messages[::-1] %}
19
+ {%- set index = (messages|length - 1) - loop.index0 %}
20
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}
21
+ {%- set ns.multi_step_tool = false %}
22
+ {%- set ns.last_query_index = index %}
23
+ {%- endif %}
24
+ {%- endfor %}
25
+ {%- for message in messages %}
26
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
27
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
28
+ {%- elif message.role == "assistant" %}
29
+ {%- set content = message.content %}
30
+ {%- set reasoning_content = '' %}
31
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
32
+ {%- set reasoning_content = message.reasoning_content %}
33
+ {%- else %}
34
+ {%- if '</think>' in message.content %}
35
+ {%- set content = message.content.split('</think>')[-1].lstrip('\n') %}
36
+ {%- set reasoning_content = message.content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
37
+ {%- endif %}
38
+ {%- endif %}
39
+ {%- if loop.index0 > ns.last_query_index %}
40
+ {%- if loop.last or (not loop.last and reasoning_content) %}
41
+ {{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content.strip('\n') + '\n</think>\n\n' + content.lstrip('\n') }}
42
+ {%- else %}
43
+ {{- '<|im_start|>' + message.role + '\n' + content }}
44
+ {%- endif %}
45
+ {%- else %}
46
+ {{- '<|im_start|>' + message.role + '\n' + content }}
47
+ {%- endif %}
48
+ {%- if message.tool_calls %}
49
+ {%- for tool_call in message.tool_calls %}
50
+ {%- if (loop.first and content) or (not loop.first) %}
51
+ {{- '\n' }}
52
+ {%- endif %}
53
+ {%- if tool_call.function %}
54
+ {%- set tool_call = tool_call.function %}
55
+ {%- endif %}
56
+ {{- '<tool_call>\n{"name": "' }}
57
+ {{- tool_call.name }}
58
+ {{- '", "arguments": ' }}
59
+ {%- if tool_call.arguments is string %}
60
+ {{- tool_call.arguments }}
61
+ {%- else %}
62
+ {{- tool_call.arguments | tojson }}
63
+ {%- endif %}
64
+ {{- '}\n</tool_call>' }}
65
+ {%- endfor %}
66
+ {%- endif %}
67
+ {{- '<|im_end|>\n' }}
68
+ {%- elif message.role == "tool" %}
69
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
70
+ {{- '<|im_start|>user' }}
71
+ {%- endif %}
72
+ {{- '\n<tool_response>\n' }}
73
+ {{- message.content }}
74
+ {{- '\n</tool_response>' }}
75
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
76
+ {{- '<|im_end|>\n' }}
77
+ {%- endif %}
78
+ {%- endif %}
79
+ {%- endfor %}
80
+ {%- if add_generation_prompt %}
81
+ {{- '<|im_start|>assistant\n' }}
82
+ {%- if enable_thinking is defined and enable_thinking is false %}
83
+ {{- '<think>\n\n</think>\n\n' }}
84
+ {%- endif %}
85
+ {%- endif %}
config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen3ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "eos_token_id": 151643,
9
+ "head_dim": 128,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 2560,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 9728,
14
+ "max_position_embeddings": 32768,
15
+ "max_window_layers": 36,
16
+ "model_type": "qwen3",
17
+ "num_attention_heads": 32,
18
+ "num_hidden_layers": 36,
19
+ "num_key_value_heads": 8,
20
+ "rms_norm_eps": 1e-06,
21
+ "rope_scaling": null,
22
+ "rope_theta": 1000000,
23
+ "sliding_window": null,
24
+ "tie_word_embeddings": true,
25
+ "torch_dtype": "bfloat16",
26
+ "transformers_version": "4.52.4",
27
+ "use_cache": true,
28
+ "use_sliding_window": false,
29
+ "vocab_size": 151936
30
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "eos_token_id": 151643,
4
+ "max_new_tokens": 2048,
5
+ "transformers_version": "4.52.4"
6
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ee4e9b1b78262c5ea3b460b7ed6231ab00a0b13a149fac3077af78160742a26
3
+ size 4967215360
model-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59620eabaf438533c845d54d79160abeff7b8dd05bc41d8f751964dd14d6c003
3
+ size 3077766632
model.safetensors.index.json ADDED
@@ -0,0 +1,405 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 8044936192
4
+ },
5
+ "weight_map": {
6
+ "model.embed_tokens.weight": "model-00001-of-00002.safetensors",
7
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00002.safetensors",
8
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
9
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
10
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
11
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
12
+ "model.layers.0.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
13
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
14
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
15
+ "model.layers.0.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
16
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
17
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
18
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00002.safetensors",
19
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
20
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
21
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
22
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
23
+ "model.layers.1.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
24
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
25
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
26
+ "model.layers.1.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
27
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
28
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
29
+ "model.layers.10.input_layernorm.weight": "model-00001-of-00002.safetensors",
30
+ "model.layers.10.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
31
+ "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
32
+ "model.layers.10.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
33
+ "model.layers.10.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
34
+ "model.layers.10.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
35
+ "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
36
+ "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
37
+ "model.layers.10.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
38
+ "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
39
+ "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
40
+ "model.layers.11.input_layernorm.weight": "model-00001-of-00002.safetensors",
41
+ "model.layers.11.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
42
+ "model.layers.11.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
43
+ "model.layers.11.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
44
+ "model.layers.11.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
45
+ "model.layers.11.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
46
+ "model.layers.11.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
47
+ "model.layers.11.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
48
+ "model.layers.11.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
49
+ "model.layers.11.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
50
+ "model.layers.11.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
51
+ "model.layers.12.input_layernorm.weight": "model-00001-of-00002.safetensors",
52
+ "model.layers.12.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
53
+ "model.layers.12.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
54
+ "model.layers.12.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
55
+ "model.layers.12.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
56
+ "model.layers.12.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
57
+ "model.layers.12.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
58
+ "model.layers.12.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
59
+ "model.layers.12.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
60
+ "model.layers.12.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
61
+ "model.layers.12.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
62
+ "model.layers.13.input_layernorm.weight": "model-00001-of-00002.safetensors",
63
+ "model.layers.13.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
64
+ "model.layers.13.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
65
+ "model.layers.13.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
66
+ "model.layers.13.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
67
+ "model.layers.13.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
68
+ "model.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
69
+ "model.layers.13.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
70
+ "model.layers.13.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
71
+ "model.layers.13.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
72
+ "model.layers.13.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
73
+ "model.layers.14.input_layernorm.weight": "model-00001-of-00002.safetensors",
74
+ "model.layers.14.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
75
+ "model.layers.14.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
76
+ "model.layers.14.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
77
+ "model.layers.14.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
78
+ "model.layers.14.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
79
+ "model.layers.14.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
80
+ "model.layers.14.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
81
+ "model.layers.14.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
82
+ "model.layers.14.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
83
+ "model.layers.14.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
84
+ "model.layers.15.input_layernorm.weight": "model-00001-of-00002.safetensors",
85
+ "model.layers.15.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
86
+ "model.layers.15.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
87
+ "model.layers.15.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
88
+ "model.layers.15.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
89
+ "model.layers.15.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
90
+ "model.layers.15.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
91
+ "model.layers.15.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
92
+ "model.layers.15.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
93
+ "model.layers.15.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
94
+ "model.layers.15.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
95
+ "model.layers.16.input_layernorm.weight": "model-00001-of-00002.safetensors",
96
+ "model.layers.16.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
97
+ "model.layers.16.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
98
+ "model.layers.16.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
99
+ "model.layers.16.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
100
+ "model.layers.16.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
101
+ "model.layers.16.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
102
+ "model.layers.16.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
103
+ "model.layers.16.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
104
+ "model.layers.16.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
105
+ "model.layers.16.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
106
+ "model.layers.17.input_layernorm.weight": "model-00001-of-00002.safetensors",
107
+ "model.layers.17.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
108
+ "model.layers.17.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
109
+ "model.layers.17.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
110
+ "model.layers.17.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
111
+ "model.layers.17.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
112
+ "model.layers.17.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
113
+ "model.layers.17.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
114
+ "model.layers.17.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
115
+ "model.layers.17.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
116
+ "model.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
117
+ "model.layers.18.input_layernorm.weight": "model-00001-of-00002.safetensors",
118
+ "model.layers.18.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
119
+ "model.layers.18.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
120
+ "model.layers.18.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
121
+ "model.layers.18.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
122
+ "model.layers.18.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
123
+ "model.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
124
+ "model.layers.18.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
125
+ "model.layers.18.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
126
+ "model.layers.18.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
127
+ "model.layers.18.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
128
+ "model.layers.19.input_layernorm.weight": "model-00001-of-00002.safetensors",
129
+ "model.layers.19.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
130
+ "model.layers.19.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
131
+ "model.layers.19.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
132
+ "model.layers.19.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
133
+ "model.layers.19.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
134
+ "model.layers.19.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
135
+ "model.layers.19.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
136
+ "model.layers.19.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
137
+ "model.layers.19.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
138
+ "model.layers.19.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
139
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00002.safetensors",
140
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
141
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
142
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
143
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
144
+ "model.layers.2.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
145
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
146
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
147
+ "model.layers.2.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
148
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
149
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
150
+ "model.layers.20.input_layernorm.weight": "model-00002-of-00002.safetensors",
151
+ "model.layers.20.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
152
+ "model.layers.20.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
153
+ "model.layers.20.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
154
+ "model.layers.20.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
155
+ "model.layers.20.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
156
+ "model.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
157
+ "model.layers.20.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
158
+ "model.layers.20.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
159
+ "model.layers.20.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
160
+ "model.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
161
+ "model.layers.21.input_layernorm.weight": "model-00002-of-00002.safetensors",
162
+ "model.layers.21.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
163
+ "model.layers.21.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
164
+ "model.layers.21.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
165
+ "model.layers.21.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
166
+ "model.layers.21.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
167
+ "model.layers.21.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
168
+ "model.layers.21.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
169
+ "model.layers.21.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
170
+ "model.layers.21.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
171
+ "model.layers.21.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
172
+ "model.layers.22.input_layernorm.weight": "model-00002-of-00002.safetensors",
173
+ "model.layers.22.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
174
+ "model.layers.22.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
175
+ "model.layers.22.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
176
+ "model.layers.22.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
177
+ "model.layers.22.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
178
+ "model.layers.22.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
179
+ "model.layers.22.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
180
+ "model.layers.22.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
181
+ "model.layers.22.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
182
+ "model.layers.22.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
183
+ "model.layers.23.input_layernorm.weight": "model-00002-of-00002.safetensors",
184
+ "model.layers.23.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
185
+ "model.layers.23.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
186
+ "model.layers.23.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
187
+ "model.layers.23.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
188
+ "model.layers.23.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
189
+ "model.layers.23.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
190
+ "model.layers.23.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
191
+ "model.layers.23.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
192
+ "model.layers.23.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
193
+ "model.layers.23.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
194
+ "model.layers.24.input_layernorm.weight": "model-00002-of-00002.safetensors",
195
+ "model.layers.24.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
196
+ "model.layers.24.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
197
+ "model.layers.24.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
198
+ "model.layers.24.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
199
+ "model.layers.24.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
200
+ "model.layers.24.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
201
+ "model.layers.24.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
202
+ "model.layers.24.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
203
+ "model.layers.24.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
204
+ "model.layers.24.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
205
+ "model.layers.25.input_layernorm.weight": "model-00002-of-00002.safetensors",
206
+ "model.layers.25.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
207
+ "model.layers.25.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
208
+ "model.layers.25.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
209
+ "model.layers.25.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
210
+ "model.layers.25.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
211
+ "model.layers.25.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
212
+ "model.layers.25.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
213
+ "model.layers.25.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
214
+ "model.layers.25.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
215
+ "model.layers.25.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
216
+ "model.layers.26.input_layernorm.weight": "model-00002-of-00002.safetensors",
217
+ "model.layers.26.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
218
+ "model.layers.26.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
219
+ "model.layers.26.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
220
+ "model.layers.26.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
221
+ "model.layers.26.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
222
+ "model.layers.26.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
223
+ "model.layers.26.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
224
+ "model.layers.26.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
225
+ "model.layers.26.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
226
+ "model.layers.26.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
227
+ "model.layers.27.input_layernorm.weight": "model-00002-of-00002.safetensors",
228
+ "model.layers.27.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
229
+ "model.layers.27.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
230
+ "model.layers.27.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
231
+ "model.layers.27.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
232
+ "model.layers.27.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
233
+ "model.layers.27.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
234
+ "model.layers.27.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
235
+ "model.layers.27.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
236
+ "model.layers.27.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
237
+ "model.layers.27.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
238
+ "model.layers.28.input_layernorm.weight": "model-00002-of-00002.safetensors",
239
+ "model.layers.28.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
240
+ "model.layers.28.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
241
+ "model.layers.28.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
242
+ "model.layers.28.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
243
+ "model.layers.28.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
244
+ "model.layers.28.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
245
+ "model.layers.28.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
246
+ "model.layers.28.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
247
+ "model.layers.28.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
248
+ "model.layers.28.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
249
+ "model.layers.29.input_layernorm.weight": "model-00002-of-00002.safetensors",
250
+ "model.layers.29.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
251
+ "model.layers.29.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
252
+ "model.layers.29.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
253
+ "model.layers.29.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
254
+ "model.layers.29.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
255
+ "model.layers.29.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
256
+ "model.layers.29.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
257
+ "model.layers.29.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
258
+ "model.layers.29.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
259
+ "model.layers.29.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
260
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00002.safetensors",
261
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
262
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
263
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
264
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
265
+ "model.layers.3.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
266
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
267
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
268
+ "model.layers.3.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
269
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
270
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
271
+ "model.layers.30.input_layernorm.weight": "model-00002-of-00002.safetensors",
272
+ "model.layers.30.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
273
+ "model.layers.30.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
274
+ "model.layers.30.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
275
+ "model.layers.30.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
276
+ "model.layers.30.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
277
+ "model.layers.30.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
278
+ "model.layers.30.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
279
+ "model.layers.30.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
280
+ "model.layers.30.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
281
+ "model.layers.30.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
282
+ "model.layers.31.input_layernorm.weight": "model-00002-of-00002.safetensors",
283
+ "model.layers.31.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
284
+ "model.layers.31.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
285
+ "model.layers.31.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
286
+ "model.layers.31.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
287
+ "model.layers.31.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
288
+ "model.layers.31.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
289
+ "model.layers.31.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
290
+ "model.layers.31.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
291
+ "model.layers.31.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
292
+ "model.layers.31.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
293
+ "model.layers.32.input_layernorm.weight": "model-00002-of-00002.safetensors",
294
+ "model.layers.32.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
295
+ "model.layers.32.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
296
+ "model.layers.32.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
297
+ "model.layers.32.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
298
+ "model.layers.32.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
299
+ "model.layers.32.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
300
+ "model.layers.32.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
301
+ "model.layers.32.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
302
+ "model.layers.32.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
303
+ "model.layers.32.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
304
+ "model.layers.33.input_layernorm.weight": "model-00002-of-00002.safetensors",
305
+ "model.layers.33.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
306
+ "model.layers.33.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
307
+ "model.layers.33.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
308
+ "model.layers.33.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
309
+ "model.layers.33.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
310
+ "model.layers.33.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
311
+ "model.layers.33.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
312
+ "model.layers.33.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
313
+ "model.layers.33.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
314
+ "model.layers.33.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
315
+ "model.layers.34.input_layernorm.weight": "model-00002-of-00002.safetensors",
316
+ "model.layers.34.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
317
+ "model.layers.34.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
318
+ "model.layers.34.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
319
+ "model.layers.34.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
320
+ "model.layers.34.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
321
+ "model.layers.34.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
322
+ "model.layers.34.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
323
+ "model.layers.34.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
324
+ "model.layers.34.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
325
+ "model.layers.34.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
326
+ "model.layers.35.input_layernorm.weight": "model-00002-of-00002.safetensors",
327
+ "model.layers.35.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
328
+ "model.layers.35.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
329
+ "model.layers.35.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
330
+ "model.layers.35.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
331
+ "model.layers.35.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
332
+ "model.layers.35.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
333
+ "model.layers.35.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
334
+ "model.layers.35.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
335
+ "model.layers.35.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
336
+ "model.layers.35.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
337
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00002.safetensors",
338
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
339
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
340
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
341
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
342
+ "model.layers.4.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
343
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
344
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
345
+ "model.layers.4.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
346
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
347
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
348
+ "model.layers.5.input_layernorm.weight": "model-00001-of-00002.safetensors",
349
+ "model.layers.5.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
350
+ "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
351
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
352
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
353
+ "model.layers.5.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
354
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
355
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
356
+ "model.layers.5.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
357
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
358
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
359
+ "model.layers.6.input_layernorm.weight": "model-00001-of-00002.safetensors",
360
+ "model.layers.6.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
361
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
362
+ "model.layers.6.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
363
+ "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
364
+ "model.layers.6.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
365
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
366
+ "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
367
+ "model.layers.6.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
368
+ "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
369
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
370
+ "model.layers.7.input_layernorm.weight": "model-00001-of-00002.safetensors",
371
+ "model.layers.7.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
372
+ "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
373
+ "model.layers.7.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
374
+ "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
375
+ "model.layers.7.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
376
+ "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
377
+ "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
378
+ "model.layers.7.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
379
+ "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
380
+ "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
381
+ "model.layers.8.input_layernorm.weight": "model-00001-of-00002.safetensors",
382
+ "model.layers.8.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
383
+ "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
384
+ "model.layers.8.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
385
+ "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
386
+ "model.layers.8.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
387
+ "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
388
+ "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
389
+ "model.layers.8.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
390
+ "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
391
+ "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
392
+ "model.layers.9.input_layernorm.weight": "model-00001-of-00002.safetensors",
393
+ "model.layers.9.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
394
+ "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
395
+ "model.layers.9.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
396
+ "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
397
+ "model.layers.9.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
398
+ "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
399
+ "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
400
+ "model.layers.9.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
401
+ "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
402
+ "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
403
+ "model.norm.weight": "model-00002-of-00002.safetensors"
404
+ }
405
+ }
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4d7643b30a00af1add3757b968e9ae598f906e0fadf68a354cd3629fbcd9660
3
+ size 14534279706
rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd50dcddd76a88c1d10f985b5a08df6e34744c003b6feff687f4774bf36a1513
3
+ size 15984
rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:096e351ace65ff91008be171a45173ed6369cc639fce73a288f671041e24b0ec
3
+ size 15984
rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f21c61b1a7e793bbdec183de3b52da90042305234bc7e5887986655cd3fc2192
3
+ size 15984
rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:348742498d81780506d9760c655a7a7555185b5fbd70a7ae296d88fd9aeecd84
3
+ size 15984
rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:399c4700dab9ae7b754110ce307fb7e26e22cb49b5bb233c435f0f12b77c202f
3
+ size 15984
rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:864ea2379cc907eb4189c52706cb978150d9c26e18abf74679590729a8f0c8e8
3
+ size 15984
rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25755ba07299ee6ff45936dd04df329596319c9f8095af71e6f3a219e7543e26
3
+ size 15984
rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:908f018cd701ed629c41299726da4a25f202f20a1d4bc2075a2266ed4013db3a
3
+ size 15984
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9149eb1a87bd99c883a7b9adf5ffab0fe75c86a60665dba5feb76d7679e2eab1
3
+ size 1064
special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c3dfe474a8bbe89b0e83627fd9ff784ad71027f12fd8c618708c818e808789d
3
+ size 11422648
tokenizer_config.json ADDED
@@ -0,0 +1,240 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ },
181
+ "151665": {
182
+ "content": "<tool_response>",
183
+ "lstrip": false,
184
+ "normalized": false,
185
+ "rstrip": false,
186
+ "single_word": false,
187
+ "special": false
188
+ },
189
+ "151666": {
190
+ "content": "</tool_response>",
191
+ "lstrip": false,
192
+ "normalized": false,
193
+ "rstrip": false,
194
+ "single_word": false,
195
+ "special": false
196
+ },
197
+ "151667": {
198
+ "content": "<think>",
199
+ "lstrip": false,
200
+ "normalized": false,
201
+ "rstrip": false,
202
+ "single_word": false,
203
+ "special": false
204
+ },
205
+ "151668": {
206
+ "content": "</think>",
207
+ "lstrip": false,
208
+ "normalized": false,
209
+ "rstrip": false,
210
+ "single_word": false,
211
+ "special": false
212
+ }
213
+ },
214
+ "additional_special_tokens": [
215
+ "<|im_start|>",
216
+ "<|im_end|>",
217
+ "<|object_ref_start|>",
218
+ "<|object_ref_end|>",
219
+ "<|box_start|>",
220
+ "<|box_end|>",
221
+ "<|quad_start|>",
222
+ "<|quad_end|>",
223
+ "<|vision_start|>",
224
+ "<|vision_end|>",
225
+ "<|vision_pad|>",
226
+ "<|image_pad|>",
227
+ "<|video_pad|>"
228
+ ],
229
+ "bos_token": null,
230
+ "clean_up_tokenization_spaces": false,
231
+ "eos_token": "<|endoftext|>",
232
+ "errors": "replace",
233
+ "extra_special_tokens": {},
234
+ "legacy": false,
235
+ "model_max_length": 131072,
236
+ "pad_token": "<|endoftext|>",
237
+ "split_special_tokens": false,
238
+ "tokenizer_class": "Qwen2Tokenizer",
239
+ "unk_token": null
240
+ }
trainer_state.json ADDED
@@ -0,0 +1,2433 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 0.2,
6
+ "eval_steps": 2000,
7
+ "global_step": 2000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.001,
14
+ "grad_norm": 4736.0,
15
+ "learning_rate": 1.9e-05,
16
+ "loss": 132.1055,
17
+ "loss/crossentropy": 12.246079635620116,
18
+ "loss/hidden": 18.7125,
19
+ "loss/jsd": 0.0,
20
+ "loss/logits": 10.372939014434815,
21
+ "step": 10
22
+ },
23
+ {
24
+ "epoch": 0.002,
25
+ "grad_norm": 330.0,
26
+ "grad_norm_var": 91640269.18333334,
27
+ "learning_rate": 2.8000000000000003e-05,
28
+ "loss": 95.9731,
29
+ "loss/crossentropy": 8.862393474578857,
30
+ "loss/hidden": 18.675,
31
+ "loss/jsd": 0.0,
32
+ "loss/logits": 6.677179157733917,
33
+ "step": 20
34
+ },
35
+ {
36
+ "epoch": 0.003,
37
+ "grad_norm": 394.0,
38
+ "grad_norm_var": 237715.45,
39
+ "learning_rate": 3.7e-05,
40
+ "loss": 86.3778,
41
+ "loss/crossentropy": 8.083840227127075,
42
+ "loss/hidden": 18.259375,
43
+ "loss/jsd": 0.0,
44
+ "loss/logits": 6.130921971797943,
45
+ "step": 30
46
+ },
47
+ {
48
+ "epoch": 0.004,
49
+ "grad_norm": 924.0,
50
+ "grad_norm_var": 2.6757682503402172e+16,
51
+ "learning_rate": 4.600000000000001e-05,
52
+ "loss": 82.5914,
53
+ "loss/crossentropy": 7.802511918544769,
54
+ "loss/hidden": 17.440625,
55
+ "loss/jsd": 0.0,
56
+ "loss/logits": 5.772503018379211,
57
+ "step": 40
58
+ },
59
+ {
60
+ "epoch": 0.005,
61
+ "grad_norm": 516.0,
62
+ "grad_norm_var": 38597.583333333336,
63
+ "learning_rate": 5.500000000000001e-05,
64
+ "loss": 75.3397,
65
+ "loss/crossentropy": 7.156700026988983,
66
+ "loss/hidden": 17.253125,
67
+ "loss/jsd": 0.0,
68
+ "loss/logits": 5.156575608253479,
69
+ "step": 50
70
+ },
71
+ {
72
+ "epoch": 0.006,
73
+ "grad_norm": 1232.0,
74
+ "grad_norm_var": 68241.45,
75
+ "learning_rate": 6.400000000000001e-05,
76
+ "loss": 61.2745,
77
+ "loss/crossentropy": 6.0138510942459105,
78
+ "loss/hidden": 15.80625,
79
+ "loss/jsd": 0.0,
80
+ "loss/logits": 3.8037488579750063,
81
+ "step": 60
82
+ },
83
+ {
84
+ "epoch": 0.007,
85
+ "grad_norm": 376.0,
86
+ "grad_norm_var": 626103.4,
87
+ "learning_rate": 7.3e-05,
88
+ "loss": 41.3695,
89
+ "loss/crossentropy": 4.422797441482544,
90
+ "loss/hidden": 13.1125,
91
+ "loss/jsd": 0.0,
92
+ "loss/logits": 2.4006322652101515,
93
+ "step": 70
94
+ },
95
+ {
96
+ "epoch": 0.008,
97
+ "grad_norm": 272.0,
98
+ "grad_norm_var": 674923.45,
99
+ "learning_rate": 8.200000000000001e-05,
100
+ "loss": 27.4755,
101
+ "loss/crossentropy": 3.3576226443052293,
102
+ "loss/hidden": 10.7359375,
103
+ "loss/jsd": 0.0,
104
+ "loss/logits": 1.3968962401151657,
105
+ "step": 80
106
+ },
107
+ {
108
+ "epoch": 0.009,
109
+ "grad_norm": 296.0,
110
+ "grad_norm_var": 15426.383333333333,
111
+ "learning_rate": 9.1e-05,
112
+ "loss": 22.6607,
113
+ "loss/crossentropy": 3.217679074406624,
114
+ "loss/hidden": 9.2140625,
115
+ "loss/jsd": 0.0,
116
+ "loss/logits": 1.055714099109173,
117
+ "step": 90
118
+ },
119
+ {
120
+ "epoch": 0.01,
121
+ "grad_norm": 328.0,
122
+ "grad_norm_var": 9349.666666666666,
123
+ "learning_rate": 0.0001,
124
+ "loss": 20.3108,
125
+ "loss/crossentropy": 2.934060016274452,
126
+ "loss/hidden": 8.40703125,
127
+ "loss/jsd": 0.0,
128
+ "loss/logits": 0.8702833190560341,
129
+ "step": 100
130
+ },
131
+ {
132
+ "epoch": 0.011,
133
+ "grad_norm": 194.0,
134
+ "grad_norm_var": 5992.866666666667,
135
+ "learning_rate": 0.0001,
136
+ "loss": 18.8852,
137
+ "loss/crossentropy": 2.8450062334537507,
138
+ "loss/hidden": 8.221875,
139
+ "loss/jsd": 0.0,
140
+ "loss/logits": 0.8380498677492142,
141
+ "step": 110
142
+ },
143
+ {
144
+ "epoch": 0.012,
145
+ "grad_norm": 244.0,
146
+ "grad_norm_var": 1176.5333333333333,
147
+ "learning_rate": 0.0001,
148
+ "loss": 17.97,
149
+ "loss/crossentropy": 2.612249107658863,
150
+ "loss/hidden": 7.578125,
151
+ "loss/jsd": 0.0,
152
+ "loss/logits": 0.686215291172266,
153
+ "step": 120
154
+ },
155
+ {
156
+ "epoch": 0.013,
157
+ "grad_norm": 242.0,
158
+ "grad_norm_var": 1168.8958333333333,
159
+ "learning_rate": 0.0001,
160
+ "loss": 17.2904,
161
+ "loss/crossentropy": 2.8242316216230394,
162
+ "loss/hidden": 7.7390625,
163
+ "loss/jsd": 0.0,
164
+ "loss/logits": 0.7805894792079926,
165
+ "step": 130
166
+ },
167
+ {
168
+ "epoch": 0.014,
169
+ "grad_norm": 179.0,
170
+ "grad_norm_var": 1465.1333333333334,
171
+ "learning_rate": 0.0001,
172
+ "loss": 16.5581,
173
+ "loss/crossentropy": 2.737143725156784,
174
+ "loss/hidden": 7.3421875,
175
+ "loss/jsd": 0.0,
176
+ "loss/logits": 0.6888546235859394,
177
+ "step": 140
178
+ },
179
+ {
180
+ "epoch": 0.015,
181
+ "grad_norm": 175.0,
182
+ "grad_norm_var": 1119.8625,
183
+ "learning_rate": 0.0001,
184
+ "loss": 16.0501,
185
+ "loss/crossentropy": 2.7599751561880113,
186
+ "loss/hidden": 7.05703125,
187
+ "loss/jsd": 0.0,
188
+ "loss/logits": 0.6640767879784107,
189
+ "step": 150
190
+ },
191
+ {
192
+ "epoch": 0.016,
193
+ "grad_norm": 186.0,
194
+ "grad_norm_var": 1044.5166666666667,
195
+ "learning_rate": 0.0001,
196
+ "loss": 15.4631,
197
+ "loss/crossentropy": 2.6100075274705885,
198
+ "loss/hidden": 6.8203125,
199
+ "loss/jsd": 0.0,
200
+ "loss/logits": 0.5824844464659691,
201
+ "step": 160
202
+ },
203
+ {
204
+ "epoch": 0.017,
205
+ "grad_norm": 179.0,
206
+ "grad_norm_var": 1082.8,
207
+ "learning_rate": 0.0001,
208
+ "loss": 15.2201,
209
+ "loss/crossentropy": 2.4276285111904143,
210
+ "loss/hidden": 6.8203125,
211
+ "loss/jsd": 0.0,
212
+ "loss/logits": 0.5915141828358174,
213
+ "step": 170
214
+ },
215
+ {
216
+ "epoch": 0.018,
217
+ "grad_norm": 153.0,
218
+ "grad_norm_var": 622.6625,
219
+ "learning_rate": 0.0001,
220
+ "loss": 14.9606,
221
+ "loss/crossentropy": 2.630460512638092,
222
+ "loss/hidden": 6.52578125,
223
+ "loss/jsd": 0.0,
224
+ "loss/logits": 0.5396774187684059,
225
+ "step": 180
226
+ },
227
+ {
228
+ "epoch": 0.019,
229
+ "grad_norm": 176.0,
230
+ "grad_norm_var": 1093.2,
231
+ "learning_rate": 0.0001,
232
+ "loss": 14.6255,
233
+ "loss/crossentropy": 2.3158223152160646,
234
+ "loss/hidden": 6.50390625,
235
+ "loss/jsd": 0.0,
236
+ "loss/logits": 0.4905257746577263,
237
+ "step": 190
238
+ },
239
+ {
240
+ "epoch": 0.02,
241
+ "grad_norm": 112.0,
242
+ "grad_norm_var": 695.7291666666666,
243
+ "learning_rate": 0.0001,
244
+ "loss": 14.3647,
245
+ "loss/crossentropy": 2.586851382255554,
246
+ "loss/hidden": 6.42265625,
247
+ "loss/jsd": 0.0,
248
+ "loss/logits": 0.5586091712117195,
249
+ "step": 200
250
+ },
251
+ {
252
+ "epoch": 0.021,
253
+ "grad_norm": 118.5,
254
+ "grad_norm_var": 574.3072916666666,
255
+ "learning_rate": 0.0001,
256
+ "loss": 14.0867,
257
+ "loss/crossentropy": 2.5010055124759676,
258
+ "loss/hidden": 6.34453125,
259
+ "loss/jsd": 0.0,
260
+ "loss/logits": 0.4965482771396637,
261
+ "step": 210
262
+ },
263
+ {
264
+ "epoch": 0.022,
265
+ "grad_norm": 88.5,
266
+ "grad_norm_var": 662.65,
267
+ "learning_rate": 0.0001,
268
+ "loss": 13.6551,
269
+ "loss/crossentropy": 2.573444625735283,
270
+ "loss/hidden": 6.33125,
271
+ "loss/jsd": 0.0,
272
+ "loss/logits": 0.5534068010747433,
273
+ "step": 220
274
+ },
275
+ {
276
+ "epoch": 0.023,
277
+ "grad_norm": 118.0,
278
+ "grad_norm_var": 412.1958333333333,
279
+ "learning_rate": 0.0001,
280
+ "loss": 13.4715,
281
+ "loss/crossentropy": 2.4142292886972427,
282
+ "loss/hidden": 5.96640625,
283
+ "loss/jsd": 0.0,
284
+ "loss/logits": 0.44360905699431896,
285
+ "step": 230
286
+ },
287
+ {
288
+ "epoch": 0.024,
289
+ "grad_norm": 134.0,
290
+ "grad_norm_var": 242.9,
291
+ "learning_rate": 0.0001,
292
+ "loss": 13.3289,
293
+ "loss/crossentropy": 2.4670142769813537,
294
+ "loss/hidden": 5.98671875,
295
+ "loss/jsd": 0.0,
296
+ "loss/logits": 0.47392544001340864,
297
+ "step": 240
298
+ },
299
+ {
300
+ "epoch": 0.025,
301
+ "grad_norm": 137.0,
302
+ "grad_norm_var": 158.4625,
303
+ "learning_rate": 0.0001,
304
+ "loss": 13.0031,
305
+ "loss/crossentropy": 2.416000656783581,
306
+ "loss/hidden": 5.7859375,
307
+ "loss/jsd": 0.0,
308
+ "loss/logits": 0.44607544504106045,
309
+ "step": 250
310
+ },
311
+ {
312
+ "epoch": 0.026,
313
+ "grad_norm": 109.0,
314
+ "grad_norm_var": 279.990625,
315
+ "learning_rate": 0.0001,
316
+ "loss": 13.0076,
317
+ "loss/crossentropy": 2.370332670211792,
318
+ "loss/hidden": 5.9984375,
319
+ "loss/jsd": 0.0,
320
+ "loss/logits": 0.5006627842783928,
321
+ "step": 260
322
+ },
323
+ {
324
+ "epoch": 0.027,
325
+ "grad_norm": 129.0,
326
+ "grad_norm_var": 427.37395833333335,
327
+ "learning_rate": 0.0001,
328
+ "loss": 12.8809,
329
+ "loss/crossentropy": 2.281908763945103,
330
+ "loss/hidden": 5.98671875,
331
+ "loss/jsd": 0.0,
332
+ "loss/logits": 0.45061586182564495,
333
+ "step": 270
334
+ },
335
+ {
336
+ "epoch": 0.028,
337
+ "grad_norm": 98.0,
338
+ "grad_norm_var": 278.1489583333333,
339
+ "learning_rate": 0.0001,
340
+ "loss": 12.8942,
341
+ "loss/crossentropy": 2.3922384053468706,
342
+ "loss/hidden": 5.6984375,
343
+ "loss/jsd": 0.0,
344
+ "loss/logits": 0.44376694336533545,
345
+ "step": 280
346
+ },
347
+ {
348
+ "epoch": 0.029,
349
+ "grad_norm": 99.5,
350
+ "grad_norm_var": 303.55,
351
+ "learning_rate": 0.0001,
352
+ "loss": 12.7122,
353
+ "loss/crossentropy": 2.730095013976097,
354
+ "loss/hidden": 5.49140625,
355
+ "loss/jsd": 0.0,
356
+ "loss/logits": 0.4411045670509338,
357
+ "step": 290
358
+ },
359
+ {
360
+ "epoch": 0.03,
361
+ "grad_norm": 112.5,
362
+ "grad_norm_var": 359.56666666666666,
363
+ "learning_rate": 0.0001,
364
+ "loss": 12.5618,
365
+ "loss/crossentropy": 2.3741705983877184,
366
+ "loss/hidden": 5.43203125,
367
+ "loss/jsd": 0.0,
368
+ "loss/logits": 0.40091707594692705,
369
+ "step": 300
370
+ },
371
+ {
372
+ "epoch": 0.031,
373
+ "grad_norm": 84.5,
374
+ "grad_norm_var": 245.25729166666667,
375
+ "learning_rate": 0.0001,
376
+ "loss": 12.2525,
377
+ "loss/crossentropy": 2.2781229317188263,
378
+ "loss/hidden": 5.53515625,
379
+ "loss/jsd": 0.0,
380
+ "loss/logits": 0.4274128321558237,
381
+ "step": 310
382
+ },
383
+ {
384
+ "epoch": 0.032,
385
+ "grad_norm": 108.5,
386
+ "grad_norm_var": 140.59583333333333,
387
+ "learning_rate": 0.0001,
388
+ "loss": 12.2935,
389
+ "loss/crossentropy": 2.5757294684648513,
390
+ "loss/hidden": 5.4609375,
391
+ "loss/jsd": 0.0,
392
+ "loss/logits": 0.42916890494525434,
393
+ "step": 320
394
+ },
395
+ {
396
+ "epoch": 0.033,
397
+ "grad_norm": 108.0,
398
+ "grad_norm_var": 70.89895833333334,
399
+ "learning_rate": 0.0001,
400
+ "loss": 12.1545,
401
+ "loss/crossentropy": 2.527638339996338,
402
+ "loss/hidden": 5.378125,
403
+ "loss/jsd": 0.0,
404
+ "loss/logits": 0.4032053742557764,
405
+ "step": 330
406
+ },
407
+ {
408
+ "epoch": 0.034,
409
+ "grad_norm": 210.0,
410
+ "grad_norm_var": 1272.465625,
411
+ "learning_rate": 0.0001,
412
+ "loss": 12.2482,
413
+ "loss/crossentropy": 2.5401821002364158,
414
+ "loss/hidden": 5.390625,
415
+ "loss/jsd": 0.0,
416
+ "loss/logits": 0.4444709587842226,
417
+ "step": 340
418
+ },
419
+ {
420
+ "epoch": 0.035,
421
+ "grad_norm": 79.5,
422
+ "grad_norm_var": 1376.5958333333333,
423
+ "learning_rate": 0.0001,
424
+ "loss": 12.08,
425
+ "loss/crossentropy": 2.514840933680534,
426
+ "loss/hidden": 5.2640625,
427
+ "loss/jsd": 0.0,
428
+ "loss/logits": 0.4077944982796907,
429
+ "step": 350
430
+ },
431
+ {
432
+ "epoch": 0.036,
433
+ "grad_norm": 87.0,
434
+ "grad_norm_var": 418.83229166666666,
435
+ "learning_rate": 0.0001,
436
+ "loss": 12.0245,
437
+ "loss/crossentropy": 2.420889538526535,
438
+ "loss/hidden": 5.34921875,
439
+ "loss/jsd": 0.0,
440
+ "loss/logits": 0.44222328886389733,
441
+ "step": 360
442
+ },
443
+ {
444
+ "epoch": 0.037,
445
+ "grad_norm": 76.5,
446
+ "grad_norm_var": 138.5625,
447
+ "learning_rate": 0.0001,
448
+ "loss": 11.7097,
449
+ "loss/crossentropy": 2.2826619133353234,
450
+ "loss/hidden": 5.3296875,
451
+ "loss/jsd": 0.0,
452
+ "loss/logits": 0.3849468305706978,
453
+ "step": 370
454
+ },
455
+ {
456
+ "epoch": 0.038,
457
+ "grad_norm": 96.5,
458
+ "grad_norm_var": 184.93229166666666,
459
+ "learning_rate": 0.0001,
460
+ "loss": 11.465,
461
+ "loss/crossentropy": 2.4052042722702027,
462
+ "loss/hidden": 5.16796875,
463
+ "loss/jsd": 0.0,
464
+ "loss/logits": 0.40173302926123144,
465
+ "step": 380
466
+ },
467
+ {
468
+ "epoch": 0.039,
469
+ "grad_norm": 125.5,
470
+ "grad_norm_var": 183.09583333333333,
471
+ "learning_rate": 0.0001,
472
+ "loss": 11.6273,
473
+ "loss/crossentropy": 2.540145033597946,
474
+ "loss/hidden": 5.215625,
475
+ "loss/jsd": 0.0,
476
+ "loss/logits": 0.41224894523620603,
477
+ "step": 390
478
+ },
479
+ {
480
+ "epoch": 0.04,
481
+ "grad_norm": 83.5,
482
+ "grad_norm_var": 258.315625,
483
+ "learning_rate": 0.0001,
484
+ "loss": 11.397,
485
+ "loss/crossentropy": 2.207468980550766,
486
+ "loss/hidden": 5.09296875,
487
+ "loss/jsd": 0.0,
488
+ "loss/logits": 0.3590874429792166,
489
+ "step": 400
490
+ },
491
+ {
492
+ "epoch": 0.041,
493
+ "grad_norm": 94.5,
494
+ "grad_norm_var": 184.5625,
495
+ "learning_rate": 0.0001,
496
+ "loss": 11.443,
497
+ "loss/crossentropy": 2.4378984421491623,
498
+ "loss/hidden": 5.21171875,
499
+ "loss/jsd": 0.0,
500
+ "loss/logits": 0.40493359677493573,
501
+ "step": 410
502
+ },
503
+ {
504
+ "epoch": 0.042,
505
+ "grad_norm": 106.5,
506
+ "grad_norm_var": 125.590625,
507
+ "learning_rate": 0.0001,
508
+ "loss": 11.5678,
509
+ "loss/crossentropy": 2.518555220961571,
510
+ "loss/hidden": 5.07265625,
511
+ "loss/jsd": 0.0,
512
+ "loss/logits": 0.4297170080244541,
513
+ "step": 420
514
+ },
515
+ {
516
+ "epoch": 0.043,
517
+ "grad_norm": 87.5,
518
+ "grad_norm_var": 115.765625,
519
+ "learning_rate": 0.0001,
520
+ "loss": 11.3132,
521
+ "loss/crossentropy": 2.490597203373909,
522
+ "loss/hidden": 5.11171875,
523
+ "loss/jsd": 0.0,
524
+ "loss/logits": 0.403754598274827,
525
+ "step": 430
526
+ },
527
+ {
528
+ "epoch": 0.044,
529
+ "grad_norm": 92.5,
530
+ "grad_norm_var": 156.35729166666667,
531
+ "learning_rate": 0.0001,
532
+ "loss": 11.1476,
533
+ "loss/crossentropy": 2.037529316544533,
534
+ "loss/hidden": 5.07421875,
535
+ "loss/jsd": 0.0,
536
+ "loss/logits": 0.35246654506772757,
537
+ "step": 440
538
+ },
539
+ {
540
+ "epoch": 0.045,
541
+ "grad_norm": 80.5,
542
+ "grad_norm_var": 210.66666666666666,
543
+ "learning_rate": 0.0001,
544
+ "loss": 11.3038,
545
+ "loss/crossentropy": 2.3201738983392715,
546
+ "loss/hidden": 5.0828125,
547
+ "loss/jsd": 0.0,
548
+ "loss/logits": 0.38196625709533694,
549
+ "step": 450
550
+ },
551
+ {
552
+ "epoch": 0.046,
553
+ "grad_norm": 107.5,
554
+ "grad_norm_var": 284.1666666666667,
555
+ "learning_rate": 0.0001,
556
+ "loss": 11.3625,
557
+ "loss/crossentropy": 2.4791718110442162,
558
+ "loss/hidden": 4.95546875,
559
+ "loss/jsd": 0.0,
560
+ "loss/logits": 0.36495909169316293,
561
+ "step": 460
562
+ },
563
+ {
564
+ "epoch": 0.047,
565
+ "grad_norm": 91.5,
566
+ "grad_norm_var": 247.39895833333333,
567
+ "learning_rate": 0.0001,
568
+ "loss": 11.0542,
569
+ "loss/crossentropy": 2.3155667960643767,
570
+ "loss/hidden": 4.93828125,
571
+ "loss/jsd": 0.0,
572
+ "loss/logits": 0.362844867631793,
573
+ "step": 470
574
+ },
575
+ {
576
+ "epoch": 0.048,
577
+ "grad_norm": 95.0,
578
+ "grad_norm_var": 194.79895833333333,
579
+ "learning_rate": 0.0001,
580
+ "loss": 11.2413,
581
+ "loss/crossentropy": 2.496318203210831,
582
+ "loss/hidden": 4.840625,
583
+ "loss/jsd": 0.0,
584
+ "loss/logits": 0.3887303464114666,
585
+ "step": 480
586
+ },
587
+ {
588
+ "epoch": 0.049,
589
+ "grad_norm": 74.5,
590
+ "grad_norm_var": 243.840625,
591
+ "learning_rate": 0.0001,
592
+ "loss": 10.9416,
593
+ "loss/crossentropy": 2.385223904252052,
594
+ "loss/hidden": 4.85234375,
595
+ "loss/jsd": 0.0,
596
+ "loss/logits": 0.3598880790174007,
597
+ "step": 490
598
+ },
599
+ {
600
+ "epoch": 0.05,
601
+ "grad_norm": 79.0,
602
+ "grad_norm_var": 105.990625,
603
+ "learning_rate": 0.0001,
604
+ "loss": 10.9114,
605
+ "loss/crossentropy": 2.2462552055716514,
606
+ "loss/hidden": 4.80859375,
607
+ "loss/jsd": 0.0,
608
+ "loss/logits": 0.3265662036836147,
609
+ "step": 500
610
+ },
611
+ {
612
+ "epoch": 0.051,
613
+ "grad_norm": 96.5,
614
+ "grad_norm_var": 138.43229166666666,
615
+ "learning_rate": 0.0001,
616
+ "loss": 10.8821,
617
+ "loss/crossentropy": 2.297148121893406,
618
+ "loss/hidden": 4.8609375,
619
+ "loss/jsd": 0.0,
620
+ "loss/logits": 0.3467547960579395,
621
+ "step": 510
622
+ },
623
+ {
624
+ "epoch": 0.052,
625
+ "grad_norm": 97.5,
626
+ "grad_norm_var": 129.365625,
627
+ "learning_rate": 0.0001,
628
+ "loss": 10.9299,
629
+ "loss/crossentropy": 2.4197026968002318,
630
+ "loss/hidden": 4.7921875,
631
+ "loss/jsd": 0.0,
632
+ "loss/logits": 0.3632193084806204,
633
+ "step": 520
634
+ },
635
+ {
636
+ "epoch": 0.053,
637
+ "grad_norm": 81.5,
638
+ "grad_norm_var": 99.47395833333333,
639
+ "learning_rate": 0.0001,
640
+ "loss": 10.787,
641
+ "loss/crossentropy": 2.36982424557209,
642
+ "loss/hidden": 4.825,
643
+ "loss/jsd": 0.0,
644
+ "loss/logits": 0.3405680742114782,
645
+ "step": 530
646
+ },
647
+ {
648
+ "epoch": 0.054,
649
+ "grad_norm": 85.5,
650
+ "grad_norm_var": 48.340625,
651
+ "learning_rate": 0.0001,
652
+ "loss": 10.8675,
653
+ "loss/crossentropy": 2.4611779801547526,
654
+ "loss/hidden": 4.8625,
655
+ "loss/jsd": 0.0,
656
+ "loss/logits": 0.36872007288038733,
657
+ "step": 540
658
+ },
659
+ {
660
+ "epoch": 0.055,
661
+ "grad_norm": 93.5,
662
+ "grad_norm_var": 84.24895833333333,
663
+ "learning_rate": 0.0001,
664
+ "loss": 10.64,
665
+ "loss/crossentropy": 2.1758567959070207,
666
+ "loss/hidden": 4.7484375,
667
+ "loss/jsd": 0.0,
668
+ "loss/logits": 0.3336840860545635,
669
+ "step": 550
670
+ },
671
+ {
672
+ "epoch": 0.056,
673
+ "grad_norm": 114.0,
674
+ "grad_norm_var": 129.53098958333334,
675
+ "learning_rate": 0.0001,
676
+ "loss": 10.5615,
677
+ "loss/crossentropy": 2.3970536097884176,
678
+ "loss/hidden": 4.7625,
679
+ "loss/jsd": 0.0,
680
+ "loss/logits": 0.34276723079383375,
681
+ "step": 560
682
+ },
683
+ {
684
+ "epoch": 0.057,
685
+ "grad_norm": 80.0,
686
+ "grad_norm_var": 579.57890625,
687
+ "learning_rate": 0.0001,
688
+ "loss": 10.8999,
689
+ "loss/crossentropy": 2.4695185527205465,
690
+ "loss/hidden": 4.9453125,
691
+ "loss/jsd": 0.0,
692
+ "loss/logits": 0.42829814068973066,
693
+ "step": 570
694
+ },
695
+ {
696
+ "epoch": 0.058,
697
+ "grad_norm": 85.0,
698
+ "grad_norm_var": 596.9572916666667,
699
+ "learning_rate": 0.0001,
700
+ "loss": 10.8802,
701
+ "loss/crossentropy": 2.3520184576511385,
702
+ "loss/hidden": 4.790625,
703
+ "loss/jsd": 0.0,
704
+ "loss/logits": 0.3662864986807108,
705
+ "step": 580
706
+ },
707
+ {
708
+ "epoch": 0.059,
709
+ "grad_norm": 73.0,
710
+ "grad_norm_var": 181.69583333333333,
711
+ "learning_rate": 0.0001,
712
+ "loss": 10.6744,
713
+ "loss/crossentropy": 2.2842736929655074,
714
+ "loss/hidden": 4.71484375,
715
+ "loss/jsd": 0.0,
716
+ "loss/logits": 0.3500846643000841,
717
+ "step": 590
718
+ },
719
+ {
720
+ "epoch": 0.06,
721
+ "grad_norm": 97.0,
722
+ "grad_norm_var": 160.58307291666668,
723
+ "learning_rate": 0.0001,
724
+ "loss": 10.6987,
725
+ "loss/crossentropy": 2.29906165599823,
726
+ "loss/hidden": 4.602734375,
727
+ "loss/jsd": 0.0,
728
+ "loss/logits": 0.334361494705081,
729
+ "step": 600
730
+ },
731
+ {
732
+ "epoch": 0.061,
733
+ "grad_norm": 89.0,
734
+ "grad_norm_var": 162.67682291666668,
735
+ "learning_rate": 0.0001,
736
+ "loss": 10.6143,
737
+ "loss/crossentropy": 2.3032930195331573,
738
+ "loss/hidden": 4.6703125,
739
+ "loss/jsd": 0.0,
740
+ "loss/logits": 0.3258141163736582,
741
+ "step": 610
742
+ },
743
+ {
744
+ "epoch": 0.062,
745
+ "grad_norm": 77.5,
746
+ "grad_norm_var": 97.12916666666666,
747
+ "learning_rate": 0.0001,
748
+ "loss": 10.5946,
749
+ "loss/crossentropy": 2.452244046330452,
750
+ "loss/hidden": 4.7109375,
751
+ "loss/jsd": 0.0,
752
+ "loss/logits": 0.3432691916823387,
753
+ "step": 620
754
+ },
755
+ {
756
+ "epoch": 0.063,
757
+ "grad_norm": 75.5,
758
+ "grad_norm_var": 227.69973958333333,
759
+ "learning_rate": 0.0001,
760
+ "loss": 10.6287,
761
+ "loss/crossentropy": 2.2894835874438284,
762
+ "loss/hidden": 4.74609375,
763
+ "loss/jsd": 0.0,
764
+ "loss/logits": 0.35672005768865345,
765
+ "step": 630
766
+ },
767
+ {
768
+ "epoch": 0.064,
769
+ "grad_norm": 70.0,
770
+ "grad_norm_var": 541.2322916666667,
771
+ "learning_rate": 0.0001,
772
+ "loss": 10.6195,
773
+ "loss/crossentropy": 2.4114772886037827,
774
+ "loss/hidden": 4.70546875,
775
+ "loss/jsd": 0.0,
776
+ "loss/logits": 0.35591375902295114,
777
+ "step": 640
778
+ },
779
+ {
780
+ "epoch": 0.065,
781
+ "grad_norm": 77.0,
782
+ "grad_norm_var": 435.15390625,
783
+ "learning_rate": 0.0001,
784
+ "loss": 10.4142,
785
+ "loss/crossentropy": 2.332440134882927,
786
+ "loss/hidden": 4.634375,
787
+ "loss/jsd": 0.0,
788
+ "loss/logits": 0.339809150993824,
789
+ "step": 650
790
+ },
791
+ {
792
+ "epoch": 0.066,
793
+ "grad_norm": 71.5,
794
+ "grad_norm_var": 118.03307291666667,
795
+ "learning_rate": 0.0001,
796
+ "loss": 10.4602,
797
+ "loss/crossentropy": 2.154422373324633,
798
+ "loss/hidden": 4.54140625,
799
+ "loss/jsd": 0.0,
800
+ "loss/logits": 0.3334257358685136,
801
+ "step": 660
802
+ },
803
+ {
804
+ "epoch": 0.067,
805
+ "grad_norm": 73.5,
806
+ "grad_norm_var": 144.94166666666666,
807
+ "learning_rate": 0.0001,
808
+ "loss": 10.5185,
809
+ "loss/crossentropy": 2.3223402693867685,
810
+ "loss/hidden": 4.795703125,
811
+ "loss/jsd": 0.0,
812
+ "loss/logits": 0.37188967503607273,
813
+ "step": 670
814
+ },
815
+ {
816
+ "epoch": 0.068,
817
+ "grad_norm": 61.5,
818
+ "grad_norm_var": 169.65598958333334,
819
+ "learning_rate": 0.0001,
820
+ "loss": 10.5323,
821
+ "loss/crossentropy": 2.332353001832962,
822
+ "loss/hidden": 4.50625,
823
+ "loss/jsd": 0.0,
824
+ "loss/logits": 0.31948004066944125,
825
+ "step": 680
826
+ },
827
+ {
828
+ "epoch": 0.069,
829
+ "grad_norm": 74.0,
830
+ "grad_norm_var": 155.94140625,
831
+ "learning_rate": 0.0001,
832
+ "loss": 10.4359,
833
+ "loss/crossentropy": 2.4077556908130644,
834
+ "loss/hidden": 4.623828125,
835
+ "loss/jsd": 0.0,
836
+ "loss/logits": 0.339173823595047,
837
+ "step": 690
838
+ },
839
+ {
840
+ "epoch": 0.07,
841
+ "grad_norm": 82.5,
842
+ "grad_norm_var": 125.55416666666666,
843
+ "learning_rate": 0.0001,
844
+ "loss": 10.4493,
845
+ "loss/crossentropy": 2.292634981870651,
846
+ "loss/hidden": 4.571875,
847
+ "loss/jsd": 0.0,
848
+ "loss/logits": 0.3477486100047827,
849
+ "step": 700
850
+ },
851
+ {
852
+ "epoch": 0.071,
853
+ "grad_norm": 88.0,
854
+ "grad_norm_var": 155.84166666666667,
855
+ "learning_rate": 0.0001,
856
+ "loss": 10.2041,
857
+ "loss/crossentropy": 2.4034020826220512,
858
+ "loss/hidden": 4.53046875,
859
+ "loss/jsd": 0.0,
860
+ "loss/logits": 0.3406600248068571,
861
+ "step": 710
862
+ },
863
+ {
864
+ "epoch": 0.072,
865
+ "grad_norm": 124.0,
866
+ "grad_norm_var": 230.83307291666668,
867
+ "learning_rate": 0.0001,
868
+ "loss": 10.3489,
869
+ "loss/crossentropy": 2.333241228759289,
870
+ "loss/hidden": 4.6015625,
871
+ "loss/jsd": 0.0,
872
+ "loss/logits": 0.3285223826766014,
873
+ "step": 720
874
+ },
875
+ {
876
+ "epoch": 0.073,
877
+ "grad_norm": 71.0,
878
+ "grad_norm_var": 278.95390625,
879
+ "learning_rate": 0.0001,
880
+ "loss": 10.1548,
881
+ "loss/crossentropy": 2.4066421508789064,
882
+ "loss/hidden": 4.682421875,
883
+ "loss/jsd": 0.0,
884
+ "loss/logits": 0.338771004602313,
885
+ "step": 730
886
+ },
887
+ {
888
+ "epoch": 0.074,
889
+ "grad_norm": 84.5,
890
+ "grad_norm_var": 166.85729166666667,
891
+ "learning_rate": 0.0001,
892
+ "loss": 10.2647,
893
+ "loss/crossentropy": 2.2724754482507707,
894
+ "loss/hidden": 4.567578125,
895
+ "loss/jsd": 0.0,
896
+ "loss/logits": 0.3267147310078144,
897
+ "step": 740
898
+ },
899
+ {
900
+ "epoch": 0.075,
901
+ "grad_norm": 67.5,
902
+ "grad_norm_var": 343.5247395833333,
903
+ "learning_rate": 0.0001,
904
+ "loss": 10.2815,
905
+ "loss/crossentropy": 2.3046080738306047,
906
+ "loss/hidden": 4.473828125,
907
+ "loss/jsd": 0.0,
908
+ "loss/logits": 0.33236319161951544,
909
+ "step": 750
910
+ },
911
+ {
912
+ "epoch": 0.076,
913
+ "grad_norm": 68.5,
914
+ "grad_norm_var": 306.540625,
915
+ "learning_rate": 0.0001,
916
+ "loss": 10.2479,
917
+ "loss/crossentropy": 2.2831736013293265,
918
+ "loss/hidden": 4.62734375,
919
+ "loss/jsd": 0.0,
920
+ "loss/logits": 0.3329113606363535,
921
+ "step": 760
922
+ },
923
+ {
924
+ "epoch": 0.077,
925
+ "grad_norm": 88.5,
926
+ "grad_norm_var": 111.57473958333334,
927
+ "learning_rate": 0.0001,
928
+ "loss": 10.2161,
929
+ "loss/crossentropy": 2.3853780582547186,
930
+ "loss/hidden": 4.541015625,
931
+ "loss/jsd": 0.0,
932
+ "loss/logits": 0.31959532871842383,
933
+ "step": 770
934
+ },
935
+ {
936
+ "epoch": 0.078,
937
+ "grad_norm": 80.5,
938
+ "grad_norm_var": 110.65729166666667,
939
+ "learning_rate": 0.0001,
940
+ "loss": 10.2076,
941
+ "loss/crossentropy": 2.3982744574546815,
942
+ "loss/hidden": 4.55859375,
943
+ "loss/jsd": 0.0,
944
+ "loss/logits": 0.3542841043323278,
945
+ "step": 780
946
+ },
947
+ {
948
+ "epoch": 0.079,
949
+ "grad_norm": 66.0,
950
+ "grad_norm_var": 275.6322916666667,
951
+ "learning_rate": 0.0001,
952
+ "loss": 10.1697,
953
+ "loss/crossentropy": 2.4292824655771255,
954
+ "loss/hidden": 4.632421875,
955
+ "loss/jsd": 0.0,
956
+ "loss/logits": 0.36711033545434474,
957
+ "step": 790
958
+ },
959
+ {
960
+ "epoch": 0.08,
961
+ "grad_norm": 57.25,
962
+ "grad_norm_var": 290.9291666666667,
963
+ "learning_rate": 0.0001,
964
+ "loss": 10.2176,
965
+ "loss/crossentropy": 2.380542576313019,
966
+ "loss/hidden": 4.509375,
967
+ "loss/jsd": 0.0,
968
+ "loss/logits": 0.3368827097117901,
969
+ "step": 800
970
+ },
971
+ {
972
+ "epoch": 0.081,
973
+ "grad_norm": 60.75,
974
+ "grad_norm_var": 52.67916666666667,
975
+ "learning_rate": 0.0001,
976
+ "loss": 10.2311,
977
+ "loss/crossentropy": 2.4212940514087675,
978
+ "loss/hidden": 4.55546875,
979
+ "loss/jsd": 0.0,
980
+ "loss/logits": 0.355662290379405,
981
+ "step": 810
982
+ },
983
+ {
984
+ "epoch": 0.082,
985
+ "grad_norm": 60.75,
986
+ "grad_norm_var": 65.81666666666666,
987
+ "learning_rate": 0.0001,
988
+ "loss": 10.1866,
989
+ "loss/crossentropy": 2.4809795886278154,
990
+ "loss/hidden": 4.49140625,
991
+ "loss/jsd": 0.0,
992
+ "loss/logits": 0.3553234666585922,
993
+ "step": 820
994
+ },
995
+ {
996
+ "epoch": 0.083,
997
+ "grad_norm": 56.5,
998
+ "grad_norm_var": 98.2875,
999
+ "learning_rate": 0.0001,
1000
+ "loss": 9.9805,
1001
+ "loss/crossentropy": 2.306653854250908,
1002
+ "loss/hidden": 4.40078125,
1003
+ "loss/jsd": 0.0,
1004
+ "loss/logits": 0.3146494958549738,
1005
+ "step": 830
1006
+ },
1007
+ {
1008
+ "epoch": 0.084,
1009
+ "grad_norm": 68.0,
1010
+ "grad_norm_var": 38.51015625,
1011
+ "learning_rate": 0.0001,
1012
+ "loss": 10.1087,
1013
+ "loss/crossentropy": 2.250006601214409,
1014
+ "loss/hidden": 4.422265625,
1015
+ "loss/jsd": 0.0,
1016
+ "loss/logits": 0.30200174674391744,
1017
+ "step": 840
1018
+ },
1019
+ {
1020
+ "epoch": 0.085,
1021
+ "grad_norm": 70.5,
1022
+ "grad_norm_var": 43.483072916666664,
1023
+ "learning_rate": 0.0001,
1024
+ "loss": 10.0526,
1025
+ "loss/crossentropy": 2.211633677780628,
1026
+ "loss/hidden": 4.47421875,
1027
+ "loss/jsd": 0.0,
1028
+ "loss/logits": 0.31178686060011385,
1029
+ "step": 850
1030
+ },
1031
+ {
1032
+ "epoch": 0.086,
1033
+ "grad_norm": 61.0,
1034
+ "grad_norm_var": 41.545572916666664,
1035
+ "learning_rate": 0.0001,
1036
+ "loss": 10.1915,
1037
+ "loss/crossentropy": 2.5281356513500213,
1038
+ "loss/hidden": 4.389453125,
1039
+ "loss/jsd": 0.0,
1040
+ "loss/logits": 0.34625968635082244,
1041
+ "step": 860
1042
+ },
1043
+ {
1044
+ "epoch": 0.087,
1045
+ "grad_norm": 72.5,
1046
+ "grad_norm_var": 54.475,
1047
+ "learning_rate": 0.0001,
1048
+ "loss": 10.0007,
1049
+ "loss/crossentropy": 2.4020907685160635,
1050
+ "loss/hidden": 4.326171875,
1051
+ "loss/jsd": 0.0,
1052
+ "loss/logits": 0.32252500094473363,
1053
+ "step": 870
1054
+ },
1055
+ {
1056
+ "epoch": 0.088,
1057
+ "grad_norm": 142.0,
1058
+ "grad_norm_var": 499.1375,
1059
+ "learning_rate": 0.0001,
1060
+ "loss": 9.99,
1061
+ "loss/crossentropy": 2.384984764456749,
1062
+ "loss/hidden": 4.38515625,
1063
+ "loss/jsd": 0.0,
1064
+ "loss/logits": 0.3189360786229372,
1065
+ "step": 880
1066
+ },
1067
+ {
1068
+ "epoch": 0.089,
1069
+ "grad_norm": 57.75,
1070
+ "grad_norm_var": 527.23515625,
1071
+ "learning_rate": 0.0001,
1072
+ "loss": 9.9879,
1073
+ "loss/crossentropy": 2.3401281625032424,
1074
+ "loss/hidden": 4.46328125,
1075
+ "loss/jsd": 0.0,
1076
+ "loss/logits": 0.3382201848551631,
1077
+ "step": 890
1078
+ },
1079
+ {
1080
+ "epoch": 0.09,
1081
+ "grad_norm": 71.5,
1082
+ "grad_norm_var": 95.97265625,
1083
+ "learning_rate": 0.0001,
1084
+ "loss": 9.9352,
1085
+ "loss/crossentropy": 2.3969784706830977,
1086
+ "loss/hidden": 4.384375,
1087
+ "loss/jsd": 0.0,
1088
+ "loss/logits": 0.336395762488246,
1089
+ "step": 900
1090
+ },
1091
+ {
1092
+ "epoch": 0.091,
1093
+ "grad_norm": 79.0,
1094
+ "grad_norm_var": 144.96666666666667,
1095
+ "learning_rate": 0.0001,
1096
+ "loss": 10.149,
1097
+ "loss/crossentropy": 2.4599110893905163,
1098
+ "loss/hidden": 4.30703125,
1099
+ "loss/jsd": 0.0,
1100
+ "loss/logits": 0.3240171395242214,
1101
+ "step": 910
1102
+ },
1103
+ {
1104
+ "epoch": 0.092,
1105
+ "grad_norm": 65.5,
1106
+ "grad_norm_var": 119.2375,
1107
+ "learning_rate": 0.0001,
1108
+ "loss": 9.9634,
1109
+ "loss/crossentropy": 2.4210876494646074,
1110
+ "loss/hidden": 4.30390625,
1111
+ "loss/jsd": 0.0,
1112
+ "loss/logits": 0.32166178375482557,
1113
+ "step": 920
1114
+ },
1115
+ {
1116
+ "epoch": 0.093,
1117
+ "grad_norm": 63.0,
1118
+ "grad_norm_var": 41.47083333333333,
1119
+ "learning_rate": 0.0001,
1120
+ "loss": 9.744,
1121
+ "loss/crossentropy": 2.2256636448204516,
1122
+ "loss/hidden": 4.284765625,
1123
+ "loss/jsd": 0.0,
1124
+ "loss/logits": 0.29795306362211704,
1125
+ "step": 930
1126
+ },
1127
+ {
1128
+ "epoch": 0.094,
1129
+ "grad_norm": 53.5,
1130
+ "grad_norm_var": 192.55807291666667,
1131
+ "learning_rate": 0.0001,
1132
+ "loss": 9.8636,
1133
+ "loss/crossentropy": 2.297808923572302,
1134
+ "loss/hidden": 4.31640625,
1135
+ "loss/jsd": 0.0,
1136
+ "loss/logits": 0.30742434673011304,
1137
+ "step": 940
1138
+ },
1139
+ {
1140
+ "epoch": 0.095,
1141
+ "grad_norm": 61.0,
1142
+ "grad_norm_var": 81.95729166666666,
1143
+ "learning_rate": 0.0001,
1144
+ "loss": 9.798,
1145
+ "loss/crossentropy": 2.3219059616327287,
1146
+ "loss/hidden": 4.211328125,
1147
+ "loss/jsd": 0.0,
1148
+ "loss/logits": 0.30037002861499784,
1149
+ "step": 950
1150
+ },
1151
+ {
1152
+ "epoch": 0.096,
1153
+ "grad_norm": 56.75,
1154
+ "grad_norm_var": 61.55807291666667,
1155
+ "learning_rate": 0.0001,
1156
+ "loss": 9.7449,
1157
+ "loss/crossentropy": 2.3104363679885864,
1158
+ "loss/hidden": 4.388671875,
1159
+ "loss/jsd": 0.0,
1160
+ "loss/logits": 0.327311984449625,
1161
+ "step": 960
1162
+ },
1163
+ {
1164
+ "epoch": 0.097,
1165
+ "grad_norm": 60.0,
1166
+ "grad_norm_var": 56.18932291666667,
1167
+ "learning_rate": 0.0001,
1168
+ "loss": 9.9668,
1169
+ "loss/crossentropy": 2.308886554837227,
1170
+ "loss/hidden": 4.407421875,
1171
+ "loss/jsd": 0.0,
1172
+ "loss/logits": 0.3183224782347679,
1173
+ "step": 970
1174
+ },
1175
+ {
1176
+ "epoch": 0.098,
1177
+ "grad_norm": 66.5,
1178
+ "grad_norm_var": 42.05416666666667,
1179
+ "learning_rate": 0.0001,
1180
+ "loss": 9.7807,
1181
+ "loss/crossentropy": 2.3363482102751734,
1182
+ "loss/hidden": 4.2921875,
1183
+ "loss/jsd": 0.0,
1184
+ "loss/logits": 0.3384779039770365,
1185
+ "step": 980
1186
+ },
1187
+ {
1188
+ "epoch": 0.099,
1189
+ "grad_norm": 57.25,
1190
+ "grad_norm_var": 56.891666666666666,
1191
+ "learning_rate": 0.0001,
1192
+ "loss": 9.7501,
1193
+ "loss/crossentropy": 2.1767295479774473,
1194
+ "loss/hidden": 4.466015625,
1195
+ "loss/jsd": 0.0,
1196
+ "loss/logits": 0.31410733237862587,
1197
+ "step": 990
1198
+ },
1199
+ {
1200
+ "epoch": 0.1,
1201
+ "grad_norm": 50.25,
1202
+ "grad_norm_var": 75.85598958333334,
1203
+ "learning_rate": 0.0001,
1204
+ "loss": 9.9273,
1205
+ "loss/crossentropy": 2.505411845445633,
1206
+ "loss/hidden": 4.36015625,
1207
+ "loss/jsd": 0.0,
1208
+ "loss/logits": 0.33212706074118614,
1209
+ "step": 1000
1210
+ },
1211
+ {
1212
+ "epoch": 0.101,
1213
+ "grad_norm": 77.5,
1214
+ "grad_norm_var": 196.21848958333334,
1215
+ "learning_rate": 0.0001,
1216
+ "loss": 9.9237,
1217
+ "loss/crossentropy": 2.3281257838010787,
1218
+ "loss/hidden": 4.35546875,
1219
+ "loss/jsd": 0.0,
1220
+ "loss/logits": 0.32293859515339135,
1221
+ "step": 1010
1222
+ },
1223
+ {
1224
+ "epoch": 0.102,
1225
+ "grad_norm": 63.25,
1226
+ "grad_norm_var": 167.42395833333333,
1227
+ "learning_rate": 0.0001,
1228
+ "loss": 9.7592,
1229
+ "loss/crossentropy": 2.3165650010108947,
1230
+ "loss/hidden": 4.32890625,
1231
+ "loss/jsd": 0.0,
1232
+ "loss/logits": 0.31759811006486416,
1233
+ "step": 1020
1234
+ },
1235
+ {
1236
+ "epoch": 0.103,
1237
+ "grad_norm": 60.0,
1238
+ "grad_norm_var": 153.80833333333334,
1239
+ "learning_rate": 0.0001,
1240
+ "loss": 9.7366,
1241
+ "loss/crossentropy": 2.3203016728162766,
1242
+ "loss/hidden": 4.28515625,
1243
+ "loss/jsd": 0.0,
1244
+ "loss/logits": 0.31944827549159527,
1245
+ "step": 1030
1246
+ },
1247
+ {
1248
+ "epoch": 0.104,
1249
+ "grad_norm": 66.5,
1250
+ "grad_norm_var": 3319.3958333333335,
1251
+ "learning_rate": 0.0001,
1252
+ "loss": 10.0035,
1253
+ "loss/crossentropy": 2.4188640087842943,
1254
+ "loss/hidden": 4.38828125,
1255
+ "loss/jsd": 0.0,
1256
+ "loss/logits": 0.3581279247999191,
1257
+ "step": 1040
1258
+ },
1259
+ {
1260
+ "epoch": 0.105,
1261
+ "grad_norm": 60.25,
1262
+ "grad_norm_var": 3338.31640625,
1263
+ "learning_rate": 0.0001,
1264
+ "loss": 9.6837,
1265
+ "loss/crossentropy": 2.2860016629099844,
1266
+ "loss/hidden": 4.325390625,
1267
+ "loss/jsd": 0.0,
1268
+ "loss/logits": 0.318701284006238,
1269
+ "step": 1050
1270
+ },
1271
+ {
1272
+ "epoch": 0.106,
1273
+ "grad_norm": 74.0,
1274
+ "grad_norm_var": 112.4,
1275
+ "learning_rate": 0.0001,
1276
+ "loss": 9.517,
1277
+ "loss/crossentropy": 2.4143033266067504,
1278
+ "loss/hidden": 4.319140625,
1279
+ "loss/jsd": 0.0,
1280
+ "loss/logits": 0.3072842717170715,
1281
+ "step": 1060
1282
+ },
1283
+ {
1284
+ "epoch": 0.107,
1285
+ "grad_norm": 71.5,
1286
+ "grad_norm_var": 68.60598958333334,
1287
+ "learning_rate": 0.0001,
1288
+ "loss": 9.8549,
1289
+ "loss/crossentropy": 2.351083371043205,
1290
+ "loss/hidden": 4.398046875,
1291
+ "loss/jsd": 0.0,
1292
+ "loss/logits": 0.33429058492183683,
1293
+ "step": 1070
1294
+ },
1295
+ {
1296
+ "epoch": 0.108,
1297
+ "grad_norm": 53.25,
1298
+ "grad_norm_var": 43.83229166666667,
1299
+ "learning_rate": 0.0001,
1300
+ "loss": 9.7738,
1301
+ "loss/crossentropy": 2.4011227190494537,
1302
+ "loss/hidden": 4.29453125,
1303
+ "loss/jsd": 0.0,
1304
+ "loss/logits": 0.3128178097307682,
1305
+ "step": 1080
1306
+ },
1307
+ {
1308
+ "epoch": 0.109,
1309
+ "grad_norm": 72.0,
1310
+ "grad_norm_var": 34.82890625,
1311
+ "learning_rate": 0.0001,
1312
+ "loss": 9.7432,
1313
+ "loss/crossentropy": 2.310031126439571,
1314
+ "loss/hidden": 4.38984375,
1315
+ "loss/jsd": 0.0,
1316
+ "loss/logits": 0.3273486144840717,
1317
+ "step": 1090
1318
+ },
1319
+ {
1320
+ "epoch": 0.11,
1321
+ "grad_norm": 66.5,
1322
+ "grad_norm_var": 111.64895833333334,
1323
+ "learning_rate": 0.0001,
1324
+ "loss": 9.6743,
1325
+ "loss/crossentropy": 2.3055127263069153,
1326
+ "loss/hidden": 4.21796875,
1327
+ "loss/jsd": 0.0,
1328
+ "loss/logits": 0.32233874313533306,
1329
+ "step": 1100
1330
+ },
1331
+ {
1332
+ "epoch": 0.111,
1333
+ "grad_norm": 51.5,
1334
+ "grad_norm_var": 46.70729166666667,
1335
+ "learning_rate": 0.0001,
1336
+ "loss": 9.8026,
1337
+ "loss/crossentropy": 2.314373381435871,
1338
+ "loss/hidden": 4.256640625,
1339
+ "loss/jsd": 0.0,
1340
+ "loss/logits": 0.3083756107836962,
1341
+ "step": 1110
1342
+ },
1343
+ {
1344
+ "epoch": 0.112,
1345
+ "grad_norm": 57.75,
1346
+ "grad_norm_var": 7292.4375,
1347
+ "learning_rate": 0.0001,
1348
+ "loss": 9.7291,
1349
+ "loss/crossentropy": 2.5138203650712967,
1350
+ "loss/hidden": 4.19921875,
1351
+ "loss/jsd": 0.0,
1352
+ "loss/logits": 0.30809955932199956,
1353
+ "step": 1120
1354
+ },
1355
+ {
1356
+ "epoch": 0.113,
1357
+ "grad_norm": 56.5,
1358
+ "grad_norm_var": 29.190625,
1359
+ "learning_rate": 0.0001,
1360
+ "loss": 9.6823,
1361
+ "loss/crossentropy": 2.2719234466552733,
1362
+ "loss/hidden": 4.294140625,
1363
+ "loss/jsd": 0.0,
1364
+ "loss/logits": 0.3143883816897869,
1365
+ "step": 1130
1366
+ },
1367
+ {
1368
+ "epoch": 0.114,
1369
+ "grad_norm": 60.5,
1370
+ "grad_norm_var": 45.925,
1371
+ "learning_rate": 0.0001,
1372
+ "loss": 9.7564,
1373
+ "loss/crossentropy": 2.4254489660263063,
1374
+ "loss/hidden": 4.261328125,
1375
+ "loss/jsd": 0.0,
1376
+ "loss/logits": 0.3154076419770718,
1377
+ "step": 1140
1378
+ },
1379
+ {
1380
+ "epoch": 0.115,
1381
+ "grad_norm": 56.0,
1382
+ "grad_norm_var": 71.74583333333334,
1383
+ "learning_rate": 0.0001,
1384
+ "loss": 9.7001,
1385
+ "loss/crossentropy": 2.28252642005682,
1386
+ "loss/hidden": 4.323046875,
1387
+ "loss/jsd": 0.0,
1388
+ "loss/logits": 0.3203336976468563,
1389
+ "step": 1150
1390
+ },
1391
+ {
1392
+ "epoch": 0.116,
1393
+ "grad_norm": 67.0,
1394
+ "grad_norm_var": 46.040625,
1395
+ "learning_rate": 0.0001,
1396
+ "loss": 9.7436,
1397
+ "loss/crossentropy": 2.391976150870323,
1398
+ "loss/hidden": 4.225390625,
1399
+ "loss/jsd": 0.0,
1400
+ "loss/logits": 0.31455044373869895,
1401
+ "step": 1160
1402
+ },
1403
+ {
1404
+ "epoch": 0.117,
1405
+ "grad_norm": 46.0,
1406
+ "grad_norm_var": 47.06640625,
1407
+ "learning_rate": 0.0001,
1408
+ "loss": 9.5622,
1409
+ "loss/crossentropy": 2.3361207604408265,
1410
+ "loss/hidden": 4.19296875,
1411
+ "loss/jsd": 0.0,
1412
+ "loss/logits": 0.30060703232884406,
1413
+ "step": 1170
1414
+ },
1415
+ {
1416
+ "epoch": 0.118,
1417
+ "grad_norm": 56.25,
1418
+ "grad_norm_var": 49.264322916666664,
1419
+ "learning_rate": 0.0001,
1420
+ "loss": 9.6834,
1421
+ "loss/crossentropy": 2.297483670711517,
1422
+ "loss/hidden": 4.2890625,
1423
+ "loss/jsd": 0.0,
1424
+ "loss/logits": 0.2907493541017175,
1425
+ "step": 1180
1426
+ },
1427
+ {
1428
+ "epoch": 0.119,
1429
+ "grad_norm": 52.5,
1430
+ "grad_norm_var": 12.27890625,
1431
+ "learning_rate": 0.0001,
1432
+ "loss": 9.6207,
1433
+ "loss/crossentropy": 2.2364058643579483,
1434
+ "loss/hidden": 4.277734375,
1435
+ "loss/jsd": 0.0,
1436
+ "loss/logits": 0.3097097765654325,
1437
+ "step": 1190
1438
+ },
1439
+ {
1440
+ "epoch": 0.12,
1441
+ "grad_norm": 68.5,
1442
+ "grad_norm_var": 35.055989583333336,
1443
+ "learning_rate": 0.0001,
1444
+ "loss": 9.6018,
1445
+ "loss/crossentropy": 2.2412969201803206,
1446
+ "loss/hidden": 4.287109375,
1447
+ "loss/jsd": 0.0,
1448
+ "loss/logits": 0.31851550191640854,
1449
+ "step": 1200
1450
+ },
1451
+ {
1452
+ "epoch": 0.121,
1453
+ "grad_norm": 60.5,
1454
+ "grad_norm_var": 25.774739583333332,
1455
+ "learning_rate": 0.0001,
1456
+ "loss": 9.6979,
1457
+ "loss/crossentropy": 2.3062032952904703,
1458
+ "loss/hidden": 4.258984375,
1459
+ "loss/jsd": 0.0,
1460
+ "loss/logits": 0.3124631106853485,
1461
+ "step": 1210
1462
+ },
1463
+ {
1464
+ "epoch": 0.122,
1465
+ "grad_norm": 59.25,
1466
+ "grad_norm_var": 20.026822916666667,
1467
+ "learning_rate": 0.0001,
1468
+ "loss": 9.7129,
1469
+ "loss/crossentropy": 2.4036868065595627,
1470
+ "loss/hidden": 4.20234375,
1471
+ "loss/jsd": 0.0,
1472
+ "loss/logits": 0.31101155243813994,
1473
+ "step": 1220
1474
+ },
1475
+ {
1476
+ "epoch": 0.123,
1477
+ "grad_norm": 53.25,
1478
+ "grad_norm_var": 75.30833333333334,
1479
+ "learning_rate": 0.0001,
1480
+ "loss": 9.7047,
1481
+ "loss/crossentropy": 2.3730016142129897,
1482
+ "loss/hidden": 4.193359375,
1483
+ "loss/jsd": 0.0,
1484
+ "loss/logits": 0.3105484452098608,
1485
+ "step": 1230
1486
+ },
1487
+ {
1488
+ "epoch": 0.124,
1489
+ "grad_norm": 62.25,
1490
+ "grad_norm_var": 33.27682291666667,
1491
+ "learning_rate": 0.0001,
1492
+ "loss": 9.6313,
1493
+ "loss/crossentropy": 2.2872567594051363,
1494
+ "loss/hidden": 4.319140625,
1495
+ "loss/jsd": 0.0,
1496
+ "loss/logits": 0.3244694545865059,
1497
+ "step": 1240
1498
+ },
1499
+ {
1500
+ "epoch": 0.125,
1501
+ "grad_norm": 61.25,
1502
+ "grad_norm_var": 25.673958333333335,
1503
+ "learning_rate": 0.0001,
1504
+ "loss": 9.6217,
1505
+ "loss/crossentropy": 2.3013710603117943,
1506
+ "loss/hidden": 4.29140625,
1507
+ "loss/jsd": 0.0,
1508
+ "loss/logits": 0.32178852558135984,
1509
+ "step": 1250
1510
+ },
1511
+ {
1512
+ "epoch": 0.126,
1513
+ "grad_norm": 49.5,
1514
+ "grad_norm_var": 46.97265625,
1515
+ "learning_rate": 0.0001,
1516
+ "loss": 9.6151,
1517
+ "loss/crossentropy": 2.2743802405893803,
1518
+ "loss/hidden": 4.19921875,
1519
+ "loss/jsd": 0.0,
1520
+ "loss/logits": 0.3104738780297339,
1521
+ "step": 1260
1522
+ },
1523
+ {
1524
+ "epoch": 0.127,
1525
+ "grad_norm": 52.25,
1526
+ "grad_norm_var": 318.94557291666666,
1527
+ "learning_rate": 0.0001,
1528
+ "loss": 9.635,
1529
+ "loss/crossentropy": 2.2751111879944803,
1530
+ "loss/hidden": 4.17578125,
1531
+ "loss/jsd": 0.0,
1532
+ "loss/logits": 0.2947248375043273,
1533
+ "step": 1270
1534
+ },
1535
+ {
1536
+ "epoch": 0.128,
1537
+ "grad_norm": 59.5,
1538
+ "grad_norm_var": 201.26848958333332,
1539
+ "learning_rate": 0.0001,
1540
+ "loss": 9.605,
1541
+ "loss/crossentropy": 2.3590754181146623,
1542
+ "loss/hidden": 4.116015625,
1543
+ "loss/jsd": 0.0,
1544
+ "loss/logits": 0.29773430675268175,
1545
+ "step": 1280
1546
+ },
1547
+ {
1548
+ "epoch": 0.129,
1549
+ "grad_norm": 50.0,
1550
+ "grad_norm_var": 25.795833333333334,
1551
+ "learning_rate": 0.0001,
1552
+ "loss": 9.4314,
1553
+ "loss/crossentropy": 2.165515697002411,
1554
+ "loss/hidden": 4.148046875,
1555
+ "loss/jsd": 0.0,
1556
+ "loss/logits": 0.2729496695101261,
1557
+ "step": 1290
1558
+ },
1559
+ {
1560
+ "epoch": 0.13,
1561
+ "grad_norm": 51.5,
1562
+ "grad_norm_var": 65.69557291666666,
1563
+ "learning_rate": 0.0001,
1564
+ "loss": 9.4579,
1565
+ "loss/crossentropy": 2.425456903874874,
1566
+ "loss/hidden": 4.1140625,
1567
+ "loss/jsd": 0.0,
1568
+ "loss/logits": 0.3002984166145325,
1569
+ "step": 1300
1570
+ },
1571
+ {
1572
+ "epoch": 0.131,
1573
+ "grad_norm": 55.75,
1574
+ "grad_norm_var": 74.63515625,
1575
+ "learning_rate": 0.0001,
1576
+ "loss": 9.562,
1577
+ "loss/crossentropy": 2.3212677478790282,
1578
+ "loss/hidden": 4.209765625,
1579
+ "loss/jsd": 0.0,
1580
+ "loss/logits": 0.28645528480410576,
1581
+ "step": 1310
1582
+ },
1583
+ {
1584
+ "epoch": 0.132,
1585
+ "grad_norm": 44.75,
1586
+ "grad_norm_var": 39.139322916666664,
1587
+ "learning_rate": 0.0001,
1588
+ "loss": 9.305,
1589
+ "loss/crossentropy": 2.2911602184176445,
1590
+ "loss/hidden": 4.133984375,
1591
+ "loss/jsd": 0.0,
1592
+ "loss/logits": 0.28404638059437276,
1593
+ "step": 1320
1594
+ },
1595
+ {
1596
+ "epoch": 0.133,
1597
+ "grad_norm": 52.25,
1598
+ "grad_norm_var": 76.19583333333334,
1599
+ "learning_rate": 0.0001,
1600
+ "loss": 9.3122,
1601
+ "loss/crossentropy": 2.3109163105487824,
1602
+ "loss/hidden": 4.137109375,
1603
+ "loss/jsd": 0.0,
1604
+ "loss/logits": 0.2864396806806326,
1605
+ "step": 1330
1606
+ },
1607
+ {
1608
+ "epoch": 0.134,
1609
+ "grad_norm": 47.0,
1610
+ "grad_norm_var": 41.66015625,
1611
+ "learning_rate": 0.0001,
1612
+ "loss": 9.4629,
1613
+ "loss/crossentropy": 2.353537403047085,
1614
+ "loss/hidden": 4.08203125,
1615
+ "loss/jsd": 0.0,
1616
+ "loss/logits": 0.2971150416880846,
1617
+ "step": 1340
1618
+ },
1619
+ {
1620
+ "epoch": 0.135,
1621
+ "grad_norm": 46.25,
1622
+ "grad_norm_var": 45.31848958333333,
1623
+ "learning_rate": 0.0001,
1624
+ "loss": 9.365,
1625
+ "loss/crossentropy": 2.3774181246757506,
1626
+ "loss/hidden": 4.09296875,
1627
+ "loss/jsd": 0.0,
1628
+ "loss/logits": 0.2799839396029711,
1629
+ "step": 1350
1630
+ },
1631
+ {
1632
+ "epoch": 0.136,
1633
+ "grad_norm": 51.0,
1634
+ "grad_norm_var": 17.93515625,
1635
+ "learning_rate": 0.0001,
1636
+ "loss": 9.3498,
1637
+ "loss/crossentropy": 2.246833881735802,
1638
+ "loss/hidden": 4.18828125,
1639
+ "loss/jsd": 0.0,
1640
+ "loss/logits": 0.2903384942561388,
1641
+ "step": 1360
1642
+ },
1643
+ {
1644
+ "epoch": 0.137,
1645
+ "grad_norm": 51.25,
1646
+ "grad_norm_var": 12.420833333333333,
1647
+ "learning_rate": 0.0001,
1648
+ "loss": 9.4976,
1649
+ "loss/crossentropy": 2.453240838646889,
1650
+ "loss/hidden": 4.173828125,
1651
+ "loss/jsd": 0.0,
1652
+ "loss/logits": 0.3144164770841599,
1653
+ "step": 1370
1654
+ },
1655
+ {
1656
+ "epoch": 0.138,
1657
+ "grad_norm": 70.0,
1658
+ "grad_norm_var": 2011.0322916666667,
1659
+ "learning_rate": 0.0001,
1660
+ "loss": 9.5884,
1661
+ "loss/crossentropy": 2.174116183817387,
1662
+ "loss/hidden": 4.24921875,
1663
+ "loss/jsd": 0.0,
1664
+ "loss/logits": 0.2923248626291752,
1665
+ "step": 1380
1666
+ },
1667
+ {
1668
+ "epoch": 0.139,
1669
+ "grad_norm": 53.75,
1670
+ "grad_norm_var": 1988.8833333333334,
1671
+ "learning_rate": 0.0001,
1672
+ "loss": 9.5249,
1673
+ "loss/crossentropy": 2.3638354018330574,
1674
+ "loss/hidden": 4.184765625,
1675
+ "loss/jsd": 0.0,
1676
+ "loss/logits": 0.3066251628100872,
1677
+ "step": 1390
1678
+ },
1679
+ {
1680
+ "epoch": 0.14,
1681
+ "grad_norm": 55.5,
1682
+ "grad_norm_var": 22.779166666666665,
1683
+ "learning_rate": 0.0001,
1684
+ "loss": 9.3528,
1685
+ "loss/crossentropy": 2.4166768550872804,
1686
+ "loss/hidden": 4.123828125,
1687
+ "loss/jsd": 0.0,
1688
+ "loss/logits": 0.29636494982987643,
1689
+ "step": 1400
1690
+ },
1691
+ {
1692
+ "epoch": 0.141,
1693
+ "grad_norm": 60.5,
1694
+ "grad_norm_var": 66.59348958333334,
1695
+ "learning_rate": 0.0001,
1696
+ "loss": 9.5339,
1697
+ "loss/crossentropy": 2.3475931867957116,
1698
+ "loss/hidden": 4.1953125,
1699
+ "loss/jsd": 0.0,
1700
+ "loss/logits": 0.30608872696757317,
1701
+ "step": 1410
1702
+ },
1703
+ {
1704
+ "epoch": 0.142,
1705
+ "grad_norm": 51.25,
1706
+ "grad_norm_var": 62.49140625,
1707
+ "learning_rate": 0.0001,
1708
+ "loss": 9.3342,
1709
+ "loss/crossentropy": 2.1785849004983904,
1710
+ "loss/hidden": 4.161328125,
1711
+ "loss/jsd": 0.0,
1712
+ "loss/logits": 0.27641028352081776,
1713
+ "step": 1420
1714
+ },
1715
+ {
1716
+ "epoch": 0.143,
1717
+ "grad_norm": 54.25,
1718
+ "grad_norm_var": 30.154166666666665,
1719
+ "learning_rate": 0.0001,
1720
+ "loss": 9.3898,
1721
+ "loss/crossentropy": 2.3990818440914152,
1722
+ "loss/hidden": 4.18359375,
1723
+ "loss/jsd": 0.0,
1724
+ "loss/logits": 0.2944341886788607,
1725
+ "step": 1430
1726
+ },
1727
+ {
1728
+ "epoch": 0.144,
1729
+ "grad_norm": 51.75,
1730
+ "grad_norm_var": 38.93932291666667,
1731
+ "learning_rate": 0.0001,
1732
+ "loss": 9.4628,
1733
+ "loss/crossentropy": 2.4946817860007284,
1734
+ "loss/hidden": 4.198828125,
1735
+ "loss/jsd": 0.0,
1736
+ "loss/logits": 0.31867978498339655,
1737
+ "step": 1440
1738
+ },
1739
+ {
1740
+ "epoch": 0.145,
1741
+ "grad_norm": 53.75,
1742
+ "grad_norm_var": 33.9,
1743
+ "learning_rate": 0.0001,
1744
+ "loss": 9.3416,
1745
+ "loss/crossentropy": 2.2067521095275877,
1746
+ "loss/hidden": 4.235546875,
1747
+ "loss/jsd": 0.0,
1748
+ "loss/logits": 0.2976540043950081,
1749
+ "step": 1450
1750
+ },
1751
+ {
1752
+ "epoch": 0.146,
1753
+ "grad_norm": 60.75,
1754
+ "grad_norm_var": 142.08229166666666,
1755
+ "learning_rate": 0.0001,
1756
+ "loss": 9.4716,
1757
+ "loss/crossentropy": 2.4361192852258684,
1758
+ "loss/hidden": 4.1140625,
1759
+ "loss/jsd": 0.0,
1760
+ "loss/logits": 0.2877715673297644,
1761
+ "step": 1460
1762
+ },
1763
+ {
1764
+ "epoch": 0.147,
1765
+ "grad_norm": 58.75,
1766
+ "grad_norm_var": 44.35,
1767
+ "learning_rate": 0.0001,
1768
+ "loss": 9.4006,
1769
+ "loss/crossentropy": 2.239429622516036,
1770
+ "loss/hidden": 4.026171875,
1771
+ "loss/jsd": 0.0,
1772
+ "loss/logits": 0.27844256814569235,
1773
+ "step": 1470
1774
+ },
1775
+ {
1776
+ "epoch": 0.148,
1777
+ "grad_norm": 45.0,
1778
+ "grad_norm_var": 33.95390625,
1779
+ "learning_rate": 0.0001,
1780
+ "loss": 9.3993,
1781
+ "loss/crossentropy": 2.0759536787867545,
1782
+ "loss/hidden": 4.068359375,
1783
+ "loss/jsd": 0.0,
1784
+ "loss/logits": 0.2688772227615118,
1785
+ "step": 1480
1786
+ },
1787
+ {
1788
+ "epoch": 0.149,
1789
+ "grad_norm": 51.75,
1790
+ "grad_norm_var": 25.795833333333334,
1791
+ "learning_rate": 0.0001,
1792
+ "loss": 9.3786,
1793
+ "loss/crossentropy": 2.286362998187542,
1794
+ "loss/hidden": 4.15078125,
1795
+ "loss/jsd": 0.0,
1796
+ "loss/logits": 0.2942257083952427,
1797
+ "step": 1490
1798
+ },
1799
+ {
1800
+ "epoch": 0.15,
1801
+ "grad_norm": 46.75,
1802
+ "grad_norm_var": 20.520833333333332,
1803
+ "learning_rate": 0.0001,
1804
+ "loss": 9.2903,
1805
+ "loss/crossentropy": 2.312733788788319,
1806
+ "loss/hidden": 3.971484375,
1807
+ "loss/jsd": 0.0,
1808
+ "loss/logits": 0.2691910218447447,
1809
+ "step": 1500
1810
+ },
1811
+ {
1812
+ "epoch": 0.151,
1813
+ "grad_norm": 50.25,
1814
+ "grad_norm_var": 28.290625,
1815
+ "learning_rate": 0.0001,
1816
+ "loss": 9.3076,
1817
+ "loss/crossentropy": 2.2467628076672552,
1818
+ "loss/hidden": 4.105078125,
1819
+ "loss/jsd": 0.0,
1820
+ "loss/logits": 0.2887777745723724,
1821
+ "step": 1510
1822
+ },
1823
+ {
1824
+ "epoch": 0.152,
1825
+ "grad_norm": 63.5,
1826
+ "grad_norm_var": 33.73098958333333,
1827
+ "learning_rate": 0.0001,
1828
+ "loss": 9.4203,
1829
+ "loss/crossentropy": 2.372379180788994,
1830
+ "loss/hidden": 4.07578125,
1831
+ "loss/jsd": 0.0,
1832
+ "loss/logits": 0.3087839350104332,
1833
+ "step": 1520
1834
+ },
1835
+ {
1836
+ "epoch": 0.153,
1837
+ "grad_norm": 45.5,
1838
+ "grad_norm_var": 40.108333333333334,
1839
+ "learning_rate": 0.0001,
1840
+ "loss": 9.3215,
1841
+ "loss/crossentropy": 2.3452367037534714,
1842
+ "loss/hidden": 4.210546875,
1843
+ "loss/jsd": 0.0,
1844
+ "loss/logits": 0.3159611392766237,
1845
+ "step": 1530
1846
+ },
1847
+ {
1848
+ "epoch": 0.154,
1849
+ "grad_norm": 58.25,
1850
+ "grad_norm_var": 27.539322916666666,
1851
+ "learning_rate": 0.0001,
1852
+ "loss": 9.3755,
1853
+ "loss/crossentropy": 2.3029753446578978,
1854
+ "loss/hidden": 3.999609375,
1855
+ "loss/jsd": 0.0,
1856
+ "loss/logits": 0.2623455457389355,
1857
+ "step": 1540
1858
+ },
1859
+ {
1860
+ "epoch": 0.155,
1861
+ "grad_norm": 51.75,
1862
+ "grad_norm_var": 26.9,
1863
+ "learning_rate": 0.0001,
1864
+ "loss": 9.3578,
1865
+ "loss/crossentropy": 2.3988554388284684,
1866
+ "loss/hidden": 4.08828125,
1867
+ "loss/jsd": 0.0,
1868
+ "loss/logits": 0.2846154376864433,
1869
+ "step": 1550
1870
+ },
1871
+ {
1872
+ "epoch": 0.156,
1873
+ "grad_norm": 91.0,
1874
+ "grad_norm_var": 1307.8372395833333,
1875
+ "learning_rate": 0.0001,
1876
+ "loss": 9.432,
1877
+ "loss/crossentropy": 2.343544365465641,
1878
+ "loss/hidden": 4.021875,
1879
+ "loss/jsd": 0.0,
1880
+ "loss/logits": 0.2907770898193121,
1881
+ "step": 1560
1882
+ },
1883
+ {
1884
+ "epoch": 0.157,
1885
+ "grad_norm": 52.0,
1886
+ "grad_norm_var": 170.62890625,
1887
+ "learning_rate": 0.0001,
1888
+ "loss": 9.3432,
1889
+ "loss/crossentropy": 2.173108433187008,
1890
+ "loss/hidden": 4.10859375,
1891
+ "loss/jsd": 0.0,
1892
+ "loss/logits": 0.28518917988985776,
1893
+ "step": 1570
1894
+ },
1895
+ {
1896
+ "epoch": 0.158,
1897
+ "grad_norm": 42.0,
1898
+ "grad_norm_var": 47.56015625,
1899
+ "learning_rate": 0.0001,
1900
+ "loss": 9.367,
1901
+ "loss/crossentropy": 2.2230691239237785,
1902
+ "loss/hidden": 4.23515625,
1903
+ "loss/jsd": 0.0,
1904
+ "loss/logits": 0.30039387457072736,
1905
+ "step": 1580
1906
+ },
1907
+ {
1908
+ "epoch": 0.159,
1909
+ "grad_norm": 72.0,
1910
+ "grad_norm_var": 1.226104970407838e+18,
1911
+ "learning_rate": 0.0001,
1912
+ "loss": 9.3564,
1913
+ "loss/crossentropy": 2.263391149044037,
1914
+ "loss/hidden": 4.10625,
1915
+ "loss/jsd": 0.0,
1916
+ "loss/logits": 0.2923804897814989,
1917
+ "step": 1590
1918
+ },
1919
+ {
1920
+ "epoch": 0.16,
1921
+ "grad_norm": 52.5,
1922
+ "grad_norm_var": 1.2261049681378806e+18,
1923
+ "learning_rate": 0.0001,
1924
+ "loss": 9.4959,
1925
+ "loss/crossentropy": 2.113241518288851,
1926
+ "loss/hidden": 4.087109375,
1927
+ "loss/jsd": 0.0,
1928
+ "loss/logits": 0.2759646028280258,
1929
+ "step": 1600
1930
+ },
1931
+ {
1932
+ "epoch": 0.161,
1933
+ "grad_norm": 66.0,
1934
+ "grad_norm_var": 734.1489583333333,
1935
+ "learning_rate": 0.0001,
1936
+ "loss": 9.4743,
1937
+ "loss/crossentropy": 2.3895165085792542,
1938
+ "loss/hidden": 4.059765625,
1939
+ "loss/jsd": 0.0,
1940
+ "loss/logits": 0.2987998936325312,
1941
+ "step": 1610
1942
+ },
1943
+ {
1944
+ "epoch": 0.162,
1945
+ "grad_norm": 44.75,
1946
+ "grad_norm_var": 50.00182291666667,
1947
+ "learning_rate": 0.0001,
1948
+ "loss": 9.1919,
1949
+ "loss/crossentropy": 2.251766300201416,
1950
+ "loss/hidden": 4.04375,
1951
+ "loss/jsd": 0.0,
1952
+ "loss/logits": 0.2781111396849155,
1953
+ "step": 1620
1954
+ },
1955
+ {
1956
+ "epoch": 0.163,
1957
+ "grad_norm": 52.75,
1958
+ "grad_norm_var": 437.49583333333334,
1959
+ "learning_rate": 0.0001,
1960
+ "loss": 9.4572,
1961
+ "loss/crossentropy": 2.382322034239769,
1962
+ "loss/hidden": 4.07734375,
1963
+ "loss/jsd": 0.0,
1964
+ "loss/logits": 0.31318275928497313,
1965
+ "step": 1630
1966
+ },
1967
+ {
1968
+ "epoch": 0.164,
1969
+ "grad_norm": 61.0,
1970
+ "grad_norm_var": 40.301822916666666,
1971
+ "learning_rate": 0.0001,
1972
+ "loss": 9.2668,
1973
+ "loss/crossentropy": 2.1683703124523164,
1974
+ "loss/hidden": 4.07578125,
1975
+ "loss/jsd": 0.0,
1976
+ "loss/logits": 0.283413190767169,
1977
+ "step": 1640
1978
+ },
1979
+ {
1980
+ "epoch": 0.165,
1981
+ "grad_norm": 42.75,
1982
+ "grad_norm_var": 57.307291666666664,
1983
+ "learning_rate": 0.0001,
1984
+ "loss": 9.339,
1985
+ "loss/crossentropy": 2.3430400043725967,
1986
+ "loss/hidden": 4.036328125,
1987
+ "loss/jsd": 0.0,
1988
+ "loss/logits": 0.287694800645113,
1989
+ "step": 1650
1990
+ },
1991
+ {
1992
+ "epoch": 0.166,
1993
+ "grad_norm": 46.75,
1994
+ "grad_norm_var": 65.52395833333334,
1995
+ "learning_rate": 0.0001,
1996
+ "loss": 9.3768,
1997
+ "loss/crossentropy": 2.2867416352033616,
1998
+ "loss/hidden": 4.017578125,
1999
+ "loss/jsd": 0.0,
2000
+ "loss/logits": 0.29683431759476664,
2001
+ "step": 1660
2002
+ },
2003
+ {
2004
+ "epoch": 0.167,
2005
+ "grad_norm": 52.5,
2006
+ "grad_norm_var": 61.18932291666667,
2007
+ "learning_rate": 0.0001,
2008
+ "loss": 9.2451,
2009
+ "loss/crossentropy": 2.3707614041864873,
2010
+ "loss/hidden": 4.06015625,
2011
+ "loss/jsd": 0.0,
2012
+ "loss/logits": 0.29184688804671166,
2013
+ "step": 1670
2014
+ },
2015
+ {
2016
+ "epoch": 0.168,
2017
+ "grad_norm": 51.75,
2018
+ "grad_norm_var": 20.895572916666666,
2019
+ "learning_rate": 0.0001,
2020
+ "loss": 9.3601,
2021
+ "loss/crossentropy": 2.3268392831087112,
2022
+ "loss/hidden": 4.12734375,
2023
+ "loss/jsd": 0.0,
2024
+ "loss/logits": 0.29570323824882505,
2025
+ "step": 1680
2026
+ },
2027
+ {
2028
+ "epoch": 0.169,
2029
+ "grad_norm": 44.0,
2030
+ "grad_norm_var": 10.290625,
2031
+ "learning_rate": 0.0001,
2032
+ "loss": 9.4214,
2033
+ "loss/crossentropy": 2.324131193757057,
2034
+ "loss/hidden": 4.1984375,
2035
+ "loss/jsd": 0.0,
2036
+ "loss/logits": 0.3133995305746794,
2037
+ "step": 1690
2038
+ },
2039
+ {
2040
+ "epoch": 0.17,
2041
+ "grad_norm": 58.25,
2042
+ "grad_norm_var": 19.124739583333334,
2043
+ "learning_rate": 0.0001,
2044
+ "loss": 9.2465,
2045
+ "loss/crossentropy": 2.35849623978138,
2046
+ "loss/hidden": 4.00703125,
2047
+ "loss/jsd": 0.0,
2048
+ "loss/logits": 0.2762619823217392,
2049
+ "step": 1700
2050
+ },
2051
+ {
2052
+ "epoch": 0.171,
2053
+ "grad_norm": 45.75,
2054
+ "grad_norm_var": 53.89895833333333,
2055
+ "learning_rate": 0.0001,
2056
+ "loss": 9.1951,
2057
+ "loss/crossentropy": 2.3914038598537446,
2058
+ "loss/hidden": 3.9984375,
2059
+ "loss/jsd": 0.0,
2060
+ "loss/logits": 0.2871177852153778,
2061
+ "step": 1710
2062
+ },
2063
+ {
2064
+ "epoch": 0.172,
2065
+ "grad_norm": 43.25,
2066
+ "grad_norm_var": 16.479166666666668,
2067
+ "learning_rate": 0.0001,
2068
+ "loss": 9.1669,
2069
+ "loss/crossentropy": 2.152750685811043,
2070
+ "loss/hidden": 4.100390625,
2071
+ "loss/jsd": 0.0,
2072
+ "loss/logits": 0.28708020225167274,
2073
+ "step": 1720
2074
+ },
2075
+ {
2076
+ "epoch": 0.173,
2077
+ "grad_norm": 49.25,
2078
+ "grad_norm_var": 13.45390625,
2079
+ "learning_rate": 0.0001,
2080
+ "loss": 9.1015,
2081
+ "loss/crossentropy": 2.2946193665266037,
2082
+ "loss/hidden": 4.085546875,
2083
+ "loss/jsd": 0.0,
2084
+ "loss/logits": 0.3062314610928297,
2085
+ "step": 1730
2086
+ },
2087
+ {
2088
+ "epoch": 0.174,
2089
+ "grad_norm": 46.5,
2090
+ "grad_norm_var": 22.473958333333332,
2091
+ "learning_rate": 0.0001,
2092
+ "loss": 9.1287,
2093
+ "loss/crossentropy": 2.1538643553853034,
2094
+ "loss/hidden": 3.9421875,
2095
+ "loss/jsd": 0.0,
2096
+ "loss/logits": 0.2666194221004844,
2097
+ "step": 1740
2098
+ },
2099
+ {
2100
+ "epoch": 0.175,
2101
+ "grad_norm": 47.0,
2102
+ "grad_norm_var": 32.62057291666667,
2103
+ "learning_rate": 0.0001,
2104
+ "loss": 9.411,
2105
+ "loss/crossentropy": 2.387891933321953,
2106
+ "loss/hidden": 4.14921875,
2107
+ "loss/jsd": 0.0,
2108
+ "loss/logits": 0.29542505368590355,
2109
+ "step": 1750
2110
+ },
2111
+ {
2112
+ "epoch": 0.176,
2113
+ "grad_norm": 45.25,
2114
+ "grad_norm_var": 26.92265625,
2115
+ "learning_rate": 0.0001,
2116
+ "loss": 9.2833,
2117
+ "loss/crossentropy": 2.3024097591638566,
2118
+ "loss/hidden": 4.03984375,
2119
+ "loss/jsd": 0.0,
2120
+ "loss/logits": 0.29066667445003985,
2121
+ "step": 1760
2122
+ },
2123
+ {
2124
+ "epoch": 0.177,
2125
+ "grad_norm": 53.5,
2126
+ "grad_norm_var": 17.832291666666666,
2127
+ "learning_rate": 0.0001,
2128
+ "loss": 9.2665,
2129
+ "loss/crossentropy": 2.4454205125570296,
2130
+ "loss/hidden": 3.955859375,
2131
+ "loss/jsd": 0.0,
2132
+ "loss/logits": 0.2910691563040018,
2133
+ "step": 1770
2134
+ },
2135
+ {
2136
+ "epoch": 0.178,
2137
+ "grad_norm": 42.25,
2138
+ "grad_norm_var": 29.865625,
2139
+ "learning_rate": 0.0001,
2140
+ "loss": 9.1701,
2141
+ "loss/crossentropy": 2.2966391056776048,
2142
+ "loss/hidden": 4.027734375,
2143
+ "loss/jsd": 0.0,
2144
+ "loss/logits": 0.2789210833609104,
2145
+ "step": 1780
2146
+ },
2147
+ {
2148
+ "epoch": 0.179,
2149
+ "grad_norm": 48.25,
2150
+ "grad_norm_var": 17.548958333333335,
2151
+ "learning_rate": 0.0001,
2152
+ "loss": 9.1992,
2153
+ "loss/crossentropy": 2.395502945780754,
2154
+ "loss/hidden": 3.934765625,
2155
+ "loss/jsd": 0.0,
2156
+ "loss/logits": 0.2776679117232561,
2157
+ "step": 1790
2158
+ },
2159
+ {
2160
+ "epoch": 0.18,
2161
+ "grad_norm": 40.75,
2162
+ "grad_norm_var": 13.282291666666667,
2163
+ "learning_rate": 0.0001,
2164
+ "loss": 9.1046,
2165
+ "loss/crossentropy": 2.22285817861557,
2166
+ "loss/hidden": 3.9046875,
2167
+ "loss/jsd": 0.0,
2168
+ "loss/logits": 0.26667180880904195,
2169
+ "step": 1800
2170
+ },
2171
+ {
2172
+ "epoch": 0.181,
2173
+ "grad_norm": 36.25,
2174
+ "grad_norm_var": 34.90807291666667,
2175
+ "learning_rate": 0.0001,
2176
+ "loss": 9.3204,
2177
+ "loss/crossentropy": 2.3842350512743,
2178
+ "loss/hidden": 4.009375,
2179
+ "loss/jsd": 0.0,
2180
+ "loss/logits": 0.300260554254055,
2181
+ "step": 1810
2182
+ },
2183
+ {
2184
+ "epoch": 0.182,
2185
+ "grad_norm": 46.75,
2186
+ "grad_norm_var": 27.77890625,
2187
+ "learning_rate": 0.0001,
2188
+ "loss": 9.0943,
2189
+ "loss/crossentropy": 2.274762773513794,
2190
+ "loss/hidden": 4.01328125,
2191
+ "loss/jsd": 0.0,
2192
+ "loss/logits": 0.28360783979296683,
2193
+ "step": 1820
2194
+ },
2195
+ {
2196
+ "epoch": 0.183,
2197
+ "grad_norm": 55.5,
2198
+ "grad_norm_var": 27.298958333333335,
2199
+ "learning_rate": 0.0001,
2200
+ "loss": 9.1699,
2201
+ "loss/crossentropy": 2.1643219627439976,
2202
+ "loss/hidden": 3.9921875,
2203
+ "loss/jsd": 0.0,
2204
+ "loss/logits": 0.267458438500762,
2205
+ "step": 1830
2206
+ },
2207
+ {
2208
+ "epoch": 0.184,
2209
+ "grad_norm": 49.75,
2210
+ "grad_norm_var": 43.94583333333333,
2211
+ "learning_rate": 0.0001,
2212
+ "loss": 9.3022,
2213
+ "loss/crossentropy": 2.464679929614067,
2214
+ "loss/hidden": 3.9546875,
2215
+ "loss/jsd": 0.0,
2216
+ "loss/logits": 0.29758369028568266,
2217
+ "step": 1840
2218
+ },
2219
+ {
2220
+ "epoch": 0.185,
2221
+ "grad_norm": 51.0,
2222
+ "grad_norm_var": 37.90807291666667,
2223
+ "learning_rate": 0.0001,
2224
+ "loss": 9.1863,
2225
+ "loss/crossentropy": 2.3199010998010636,
2226
+ "loss/hidden": 3.99453125,
2227
+ "loss/jsd": 0.0,
2228
+ "loss/logits": 0.27702242247760295,
2229
+ "step": 1850
2230
+ },
2231
+ {
2232
+ "epoch": 0.186,
2233
+ "grad_norm": 46.5,
2234
+ "grad_norm_var": 40.920833333333334,
2235
+ "learning_rate": 0.0001,
2236
+ "loss": 9.2872,
2237
+ "loss/crossentropy": 2.4041683062911035,
2238
+ "loss/hidden": 4.09140625,
2239
+ "loss/jsd": 0.0,
2240
+ "loss/logits": 0.30005627647042277,
2241
+ "step": 1860
2242
+ },
2243
+ {
2244
+ "epoch": 0.187,
2245
+ "grad_norm": 39.75,
2246
+ "grad_norm_var": 40.723958333333336,
2247
+ "learning_rate": 0.0001,
2248
+ "loss": 9.1081,
2249
+ "loss/crossentropy": 2.273802790045738,
2250
+ "loss/hidden": 4.175,
2251
+ "loss/jsd": 0.0,
2252
+ "loss/logits": 0.3045934235677123,
2253
+ "step": 1870
2254
+ },
2255
+ {
2256
+ "epoch": 0.188,
2257
+ "grad_norm": 43.25,
2258
+ "grad_norm_var": 33.35729166666667,
2259
+ "learning_rate": 0.0001,
2260
+ "loss": 9.106,
2261
+ "loss/crossentropy": 2.3607766672968866,
2262
+ "loss/hidden": 3.9765625,
2263
+ "loss/jsd": 0.0,
2264
+ "loss/logits": 0.28193066976964476,
2265
+ "step": 1880
2266
+ },
2267
+ {
2268
+ "epoch": 0.189,
2269
+ "grad_norm": 48.25,
2270
+ "grad_norm_var": 14.915625,
2271
+ "learning_rate": 0.0001,
2272
+ "loss": 9.1437,
2273
+ "loss/crossentropy": 2.2798361241817475,
2274
+ "loss/hidden": 3.9890625,
2275
+ "loss/jsd": 0.0,
2276
+ "loss/logits": 0.2721746701747179,
2277
+ "step": 1890
2278
+ },
2279
+ {
2280
+ "epoch": 0.19,
2281
+ "grad_norm": 44.0,
2282
+ "grad_norm_var": 21.223958333333332,
2283
+ "learning_rate": 0.0001,
2284
+ "loss": 9.0972,
2285
+ "loss/crossentropy": 2.21695294380188,
2286
+ "loss/hidden": 4.0015625,
2287
+ "loss/jsd": 0.0,
2288
+ "loss/logits": 0.2832322970032692,
2289
+ "step": 1900
2290
+ },
2291
+ {
2292
+ "epoch": 0.191,
2293
+ "grad_norm": 39.5,
2294
+ "grad_norm_var": 27.808333333333334,
2295
+ "learning_rate": 0.0001,
2296
+ "loss": 9.1587,
2297
+ "loss/crossentropy": 2.1728454776108266,
2298
+ "loss/hidden": 3.98828125,
2299
+ "loss/jsd": 0.0,
2300
+ "loss/logits": 0.27170457877218723,
2301
+ "step": 1910
2302
+ },
2303
+ {
2304
+ "epoch": 0.192,
2305
+ "grad_norm": 41.75,
2306
+ "grad_norm_var": 13.315625,
2307
+ "learning_rate": 0.0001,
2308
+ "loss": 9.1326,
2309
+ "loss/crossentropy": 2.154237084835768,
2310
+ "loss/hidden": 4.063671875,
2311
+ "loss/jsd": 0.0,
2312
+ "loss/logits": 0.27950075305998323,
2313
+ "step": 1920
2314
+ },
2315
+ {
2316
+ "epoch": 0.193,
2317
+ "grad_norm": 43.0,
2318
+ "grad_norm_var": 25.240625,
2319
+ "learning_rate": 0.0001,
2320
+ "loss": 9.1013,
2321
+ "loss/crossentropy": 2.2507698431611063,
2322
+ "loss/hidden": 4.01171875,
2323
+ "loss/jsd": 0.0,
2324
+ "loss/logits": 0.2808088269084692,
2325
+ "step": 1930
2326
+ },
2327
+ {
2328
+ "epoch": 0.194,
2329
+ "grad_norm": 49.25,
2330
+ "grad_norm_var": 22.832291666666666,
2331
+ "learning_rate": 0.0001,
2332
+ "loss": 9.2429,
2333
+ "loss/crossentropy": 2.288056728243828,
2334
+ "loss/hidden": 4.1546875,
2335
+ "loss/jsd": 0.0,
2336
+ "loss/logits": 0.31668607220053674,
2337
+ "step": 1940
2338
+ },
2339
+ {
2340
+ "epoch": 0.195,
2341
+ "grad_norm": 48.5,
2342
+ "grad_norm_var": 58.09557291666667,
2343
+ "learning_rate": 0.0001,
2344
+ "loss": 9.1742,
2345
+ "loss/crossentropy": 2.2107961744070055,
2346
+ "loss/hidden": 4.05078125,
2347
+ "loss/jsd": 0.0,
2348
+ "loss/logits": 0.2858551822602749,
2349
+ "step": 1950
2350
+ },
2351
+ {
2352
+ "epoch": 0.196,
2353
+ "grad_norm": 39.25,
2354
+ "grad_norm_var": 49.50390625,
2355
+ "learning_rate": 0.0001,
2356
+ "loss": 9.1293,
2357
+ "loss/crossentropy": 2.224529256671667,
2358
+ "loss/hidden": 3.9703125,
2359
+ "loss/jsd": 0.0,
2360
+ "loss/logits": 0.27973891496658326,
2361
+ "step": 1960
2362
+ },
2363
+ {
2364
+ "epoch": 0.197,
2365
+ "grad_norm": 39.25,
2366
+ "grad_norm_var": 13.890625,
2367
+ "learning_rate": 0.0001,
2368
+ "loss": 9.0689,
2369
+ "loss/crossentropy": 2.363737019896507,
2370
+ "loss/hidden": 4.027734375,
2371
+ "loss/jsd": 0.0,
2372
+ "loss/logits": 0.2919711694121361,
2373
+ "step": 1970
2374
+ },
2375
+ {
2376
+ "epoch": 0.198,
2377
+ "grad_norm": 55.75,
2378
+ "grad_norm_var": 26.655989583333334,
2379
+ "learning_rate": 0.0001,
2380
+ "loss": 9.2228,
2381
+ "loss/crossentropy": 2.3380469545722007,
2382
+ "loss/hidden": 4.0171875,
2383
+ "loss/jsd": 0.0,
2384
+ "loss/logits": 0.28581551983952524,
2385
+ "step": 1980
2386
+ },
2387
+ {
2388
+ "epoch": 0.199,
2389
+ "grad_norm": 45.0,
2390
+ "grad_norm_var": 27.357291666666665,
2391
+ "learning_rate": 0.0001,
2392
+ "loss": 9.173,
2393
+ "loss/crossentropy": 2.43135461807251,
2394
+ "loss/hidden": 3.970703125,
2395
+ "loss/jsd": 0.0,
2396
+ "loss/logits": 0.28222124874591825,
2397
+ "step": 1990
2398
+ },
2399
+ {
2400
+ "epoch": 0.2,
2401
+ "grad_norm": 44.5,
2402
+ "grad_norm_var": 16.04140625,
2403
+ "learning_rate": 0.0001,
2404
+ "loss": 9.1554,
2405
+ "loss/crossentropy": 2.4415812104940415,
2406
+ "loss/hidden": 4.064453125,
2407
+ "loss/jsd": 0.0,
2408
+ "loss/logits": 0.2984179027378559,
2409
+ "step": 2000
2410
+ }
2411
+ ],
2412
+ "logging_steps": 10,
2413
+ "max_steps": 10000,
2414
+ "num_input_tokens_seen": 0,
2415
+ "num_train_epochs": 9223372036854775807,
2416
+ "save_steps": 2000,
2417
+ "stateful_callbacks": {
2418
+ "TrainerControl": {
2419
+ "args": {
2420
+ "should_epoch_stop": false,
2421
+ "should_evaluate": false,
2422
+ "should_log": false,
2423
+ "should_save": true,
2424
+ "should_training_stop": false
2425
+ },
2426
+ "attributes": {}
2427
+ }
2428
+ },
2429
+ "total_flos": 5.715020064017613e+18,
2430
+ "train_batch_size": 2,
2431
+ "trial_name": null,
2432
+ "trial_params": null
2433
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a25da17eff495e1e71f9cb40dc8e2af9b7ac9d1878cc59e1f0d23bc294798cd
3
+ size 5496
vocab.json ADDED
The diff for this file is too large to render. See raw diff