Farouk commited on
Commit
b1ac468
Β·
1 Parent(s): 7404361

Training in progress, step 9800

Browse files
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8014b65116d31eb74cc4642ec59d578770920c8fe633b022a95d330b92443d05
3
  size 319977229
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7a57b230b28a20902460c6c306a71e8c3836b77b53594454dcf33bc3fa498ac
3
  size 319977229
checkpoint-7600/adapter_model/adapter_model/README.md DELETED
@@ -1,44 +0,0 @@
1
- ---
2
- library_name: peft
3
- ---
4
- ## Training procedure
5
-
6
-
7
- The following `bitsandbytes` quantization config was used during training:
8
- - load_in_8bit: False
9
- - load_in_4bit: True
10
- - llm_int8_threshold: 6.0
11
- - llm_int8_skip_modules: None
12
- - llm_int8_enable_fp32_cpu_offload: False
13
- - llm_int8_has_fp16_weight: False
14
- - bnb_4bit_quant_type: nf4
15
- - bnb_4bit_use_double_quant: True
16
- - bnb_4bit_compute_dtype: bfloat16
17
-
18
- The following `bitsandbytes` quantization config was used during training:
19
- - load_in_8bit: False
20
- - load_in_4bit: True
21
- - llm_int8_threshold: 6.0
22
- - llm_int8_skip_modules: None
23
- - llm_int8_enable_fp32_cpu_offload: False
24
- - llm_int8_has_fp16_weight: False
25
- - bnb_4bit_quant_type: nf4
26
- - bnb_4bit_use_double_quant: True
27
- - bnb_4bit_compute_dtype: bfloat16
28
-
29
- The following `bitsandbytes` quantization config was used during training:
30
- - load_in_8bit: False
31
- - load_in_4bit: True
32
- - llm_int8_threshold: 6.0
33
- - llm_int8_skip_modules: None
34
- - llm_int8_enable_fp32_cpu_offload: False
35
- - llm_int8_has_fp16_weight: False
36
- - bnb_4bit_quant_type: nf4
37
- - bnb_4bit_use_double_quant: True
38
- - bnb_4bit_compute_dtype: bfloat16
39
- ### Framework versions
40
-
41
- - PEFT 0.4.0
42
- - PEFT 0.4.0
43
-
44
- - PEFT 0.4.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-7600/adapter_model/adapter_model/adapter_config.json DELETED
@@ -1,26 +0,0 @@
1
- {
2
- "auto_mapping": null,
3
- "base_model_name_or_path": "pankajmathur/orca_mini_v3_7b",
4
- "bias": "none",
5
- "fan_in_fan_out": false,
6
- "inference_mode": true,
7
- "init_lora_weights": true,
8
- "layers_pattern": null,
9
- "layers_to_transform": null,
10
- "lora_alpha": 16.0,
11
- "lora_dropout": 0.1,
12
- "modules_to_save": null,
13
- "peft_type": "LORA",
14
- "r": 64,
15
- "revision": null,
16
- "target_modules": [
17
- "k_proj",
18
- "q_proj",
19
- "down_proj",
20
- "v_proj",
21
- "o_proj",
22
- "up_proj",
23
- "gate_proj"
24
- ],
25
- "task_type": "CAUSAL_LM"
26
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-7600/adapter_model/adapter_model/adapter_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:836b7b1a71ccfc2ebaf85a9950f35c6cf3c5b743caad6bd98716af98f63d1967
3
- size 319977229
 
 
 
 
checkpoint-9400/adapter_model/adapter_model/README.md CHANGED
@@ -4,6 +4,17 @@ library_name: peft
4
  ## Training procedure
5
 
6
 
 
 
 
 
 
 
 
 
 
 
 
7
  The following `bitsandbytes` quantization config was used during training:
8
  - load_in_8bit: False
9
  - load_in_4bit: True
@@ -16,5 +27,6 @@ The following `bitsandbytes` quantization config was used during training:
16
  - bnb_4bit_compute_dtype: bfloat16
17
  ### Framework versions
18
 
 
19
 
20
  - PEFT 0.4.0
 
4
  ## Training procedure
5
 
6
 
7
+ The following `bitsandbytes` quantization config was used during training:
8
+ - load_in_8bit: False
9
+ - load_in_4bit: True
10
+ - llm_int8_threshold: 6.0
11
+ - llm_int8_skip_modules: None
12
+ - llm_int8_enable_fp32_cpu_offload: False
13
+ - llm_int8_has_fp16_weight: False
14
+ - bnb_4bit_quant_type: nf4
15
+ - bnb_4bit_use_double_quant: True
16
+ - bnb_4bit_compute_dtype: bfloat16
17
+
18
  The following `bitsandbytes` quantization config was used during training:
19
  - load_in_8bit: False
20
  - load_in_4bit: True
 
27
  - bnb_4bit_compute_dtype: bfloat16
28
  ### Framework versions
29
 
30
+ - PEFT 0.4.0
31
 
32
  - PEFT 0.4.0
checkpoint-9400/adapter_model/adapter_model/adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a2dbf941556387da2f794f520804558492f5e0196740bf83c3fac8b8bf3bd5ed
3
  size 319977229
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8014b65116d31eb74cc4642ec59d578770920c8fe633b022a95d330b92443d05
3
  size 319977229
{checkpoint-7600 β†’ checkpoint-9800}/README.md RENAMED
File without changes
{checkpoint-7600 β†’ checkpoint-9800}/adapter_config.json RENAMED
File without changes
{checkpoint-7600 β†’ checkpoint-9800}/adapter_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6de77fbee9a4f7bb0318dd19b2e190cae6e12a59dccd3b8afe4378f8e2e7a1e4
3
  size 319977229
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7a57b230b28a20902460c6c306a71e8c3836b77b53594454dcf33bc3fa498ac
3
  size 319977229
{checkpoint-7600 β†’ checkpoint-9800}/added_tokens.json RENAMED
File without changes
{checkpoint-7600 β†’ checkpoint-9800}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a744bd4bf3b6c1ab844cbb9455b41b41ab458b391a033afe3de864a5b4486655
3
  size 1279539973
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3ba4e8b41914b4de856902ea5c3f59ad3ee3d8eb62e32b24cf551eeb7299100
3
  size 1279539973
{checkpoint-7600 β†’ checkpoint-9800}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b034a995a91b08ada053d8addc2e57e479848d79364291759a3ff2fdf2367d7f
3
  size 14511
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c0bb26b2fde25c9eb45bb1cd6ea1f97efa2b6d631dd06263549a5c7c30486db
3
  size 14511
{checkpoint-7600 β†’ checkpoint-9800}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:28e84ce1056e951be7d81d2edd8521bf4fd1356b40fedd4b87bf74e02969be5b
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0dca521772e835c5c16030c5b4c0d3720f6592faeca96d824ba30cf9f318f744
3
  size 627
{checkpoint-7600 β†’ checkpoint-9800}/special_tokens_map.json RENAMED
File without changes
{checkpoint-7600 β†’ checkpoint-9800}/tokenizer.model RENAMED
File without changes
{checkpoint-7600 β†’ checkpoint-9800}/tokenizer_config.json RENAMED
File without changes
{checkpoint-7600 β†’ checkpoint-9800}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
- "best_metric": 0.719873309135437,
3
- "best_model_checkpoint": "experts/expert-13/checkpoint-7600",
4
- "epoch": 1.5973097940311054,
5
- "global_step": 7600,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -7264,11 +7264,2112 @@
7264
  "mmlu_eval_accuracy_world_religions": 0.7894736842105263,
7265
  "mmlu_loss": 1.142687739497999,
7266
  "step": 7600
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7267
  }
7268
  ],
7269
  "max_steps": 10000,
7270
  "num_train_epochs": 3,
7271
- "total_flos": 1.21469820132198e+18,
7272
  "trial_name": null,
7273
  "trial_params": null
7274
  }
 
1
  {
2
+ "best_metric": 0.7109686732292175,
3
+ "best_model_checkpoint": "experts/expert-13/checkpoint-9400",
4
+ "epoch": 2.0596889449348468,
5
+ "global_step": 9800,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
7264
  "mmlu_eval_accuracy_world_religions": 0.7894736842105263,
7265
  "mmlu_loss": 1.142687739497999,
7266
  "step": 7600
7267
+ },
7268
+ {
7269
+ "epoch": 1.6,
7270
+ "learning_rate": 0.0002,
7271
+ "loss": 0.675,
7272
+ "step": 7610
7273
+ },
7274
+ {
7275
+ "epoch": 1.6,
7276
+ "learning_rate": 0.0002,
7277
+ "loss": 0.7186,
7278
+ "step": 7620
7279
+ },
7280
+ {
7281
+ "epoch": 1.6,
7282
+ "learning_rate": 0.0002,
7283
+ "loss": 0.7125,
7284
+ "step": 7630
7285
+ },
7286
+ {
7287
+ "epoch": 1.61,
7288
+ "learning_rate": 0.0002,
7289
+ "loss": 0.5775,
7290
+ "step": 7640
7291
+ },
7292
+ {
7293
+ "epoch": 1.61,
7294
+ "learning_rate": 0.0002,
7295
+ "loss": 0.6777,
7296
+ "step": 7650
7297
+ },
7298
+ {
7299
+ "epoch": 1.61,
7300
+ "learning_rate": 0.0002,
7301
+ "loss": 0.6238,
7302
+ "step": 7660
7303
+ },
7304
+ {
7305
+ "epoch": 1.61,
7306
+ "learning_rate": 0.0002,
7307
+ "loss": 0.6249,
7308
+ "step": 7670
7309
+ },
7310
+ {
7311
+ "epoch": 1.61,
7312
+ "learning_rate": 0.0002,
7313
+ "loss": 0.6576,
7314
+ "step": 7680
7315
+ },
7316
+ {
7317
+ "epoch": 1.62,
7318
+ "learning_rate": 0.0002,
7319
+ "loss": 0.62,
7320
+ "step": 7690
7321
+ },
7322
+ {
7323
+ "epoch": 1.62,
7324
+ "learning_rate": 0.0002,
7325
+ "loss": 0.6442,
7326
+ "step": 7700
7327
+ },
7328
+ {
7329
+ "epoch": 1.62,
7330
+ "learning_rate": 0.0002,
7331
+ "loss": 0.6802,
7332
+ "step": 7710
7333
+ },
7334
+ {
7335
+ "epoch": 1.62,
7336
+ "learning_rate": 0.0002,
7337
+ "loss": 0.6526,
7338
+ "step": 7720
7339
+ },
7340
+ {
7341
+ "epoch": 1.62,
7342
+ "learning_rate": 0.0002,
7343
+ "loss": 0.606,
7344
+ "step": 7730
7345
+ },
7346
+ {
7347
+ "epoch": 1.63,
7348
+ "learning_rate": 0.0002,
7349
+ "loss": 0.6409,
7350
+ "step": 7740
7351
+ },
7352
+ {
7353
+ "epoch": 1.63,
7354
+ "learning_rate": 0.0002,
7355
+ "loss": 0.7021,
7356
+ "step": 7750
7357
+ },
7358
+ {
7359
+ "epoch": 1.63,
7360
+ "learning_rate": 0.0002,
7361
+ "loss": 0.6751,
7362
+ "step": 7760
7363
+ },
7364
+ {
7365
+ "epoch": 1.63,
7366
+ "learning_rate": 0.0002,
7367
+ "loss": 0.703,
7368
+ "step": 7770
7369
+ },
7370
+ {
7371
+ "epoch": 1.64,
7372
+ "learning_rate": 0.0002,
7373
+ "loss": 0.6279,
7374
+ "step": 7780
7375
+ },
7376
+ {
7377
+ "epoch": 1.64,
7378
+ "learning_rate": 0.0002,
7379
+ "loss": 0.6784,
7380
+ "step": 7790
7381
+ },
7382
+ {
7383
+ "epoch": 1.64,
7384
+ "learning_rate": 0.0002,
7385
+ "loss": 0.6603,
7386
+ "step": 7800
7387
+ },
7388
+ {
7389
+ "epoch": 1.64,
7390
+ "eval_loss": 0.7202572226524353,
7391
+ "eval_runtime": 68.9286,
7392
+ "eval_samples_per_second": 14.508,
7393
+ "eval_steps_per_second": 7.254,
7394
+ "step": 7800
7395
+ },
7396
+ {
7397
+ "epoch": 1.64,
7398
+ "mmlu_eval_accuracy": 0.49506372934866344,
7399
+ "mmlu_eval_accuracy_abstract_algebra": 0.36363636363636365,
7400
+ "mmlu_eval_accuracy_anatomy": 0.5714285714285714,
7401
+ "mmlu_eval_accuracy_astronomy": 0.5,
7402
+ "mmlu_eval_accuracy_business_ethics": 0.5454545454545454,
7403
+ "mmlu_eval_accuracy_clinical_knowledge": 0.5517241379310345,
7404
+ "mmlu_eval_accuracy_college_biology": 0.5,
7405
+ "mmlu_eval_accuracy_college_chemistry": 0.25,
7406
+ "mmlu_eval_accuracy_college_computer_science": 0.2727272727272727,
7407
+ "mmlu_eval_accuracy_college_mathematics": 0.18181818181818182,
7408
+ "mmlu_eval_accuracy_college_medicine": 0.4090909090909091,
7409
+ "mmlu_eval_accuracy_college_physics": 0.45454545454545453,
7410
+ "mmlu_eval_accuracy_computer_security": 0.45454545454545453,
7411
+ "mmlu_eval_accuracy_conceptual_physics": 0.38461538461538464,
7412
+ "mmlu_eval_accuracy_econometrics": 0.08333333333333333,
7413
+ "mmlu_eval_accuracy_electrical_engineering": 0.375,
7414
+ "mmlu_eval_accuracy_elementary_mathematics": 0.2926829268292683,
7415
+ "mmlu_eval_accuracy_formal_logic": 0.21428571428571427,
7416
+ "mmlu_eval_accuracy_global_facts": 0.4,
7417
+ "mmlu_eval_accuracy_high_school_biology": 0.46875,
7418
+ "mmlu_eval_accuracy_high_school_chemistry": 0.2727272727272727,
7419
+ "mmlu_eval_accuracy_high_school_computer_science": 0.5555555555555556,
7420
+ "mmlu_eval_accuracy_high_school_european_history": 0.5,
7421
+ "mmlu_eval_accuracy_high_school_geography": 0.8636363636363636,
7422
+ "mmlu_eval_accuracy_high_school_government_and_politics": 0.5714285714285714,
7423
+ "mmlu_eval_accuracy_high_school_macroeconomics": 0.4418604651162791,
7424
+ "mmlu_eval_accuracy_high_school_mathematics": 0.1724137931034483,
7425
+ "mmlu_eval_accuracy_high_school_microeconomics": 0.4230769230769231,
7426
+ "mmlu_eval_accuracy_high_school_physics": 0.29411764705882354,
7427
+ "mmlu_eval_accuracy_high_school_psychology": 0.8333333333333334,
7428
+ "mmlu_eval_accuracy_high_school_statistics": 0.391304347826087,
7429
+ "mmlu_eval_accuracy_high_school_us_history": 0.5909090909090909,
7430
+ "mmlu_eval_accuracy_high_school_world_history": 0.6923076923076923,
7431
+ "mmlu_eval_accuracy_human_aging": 0.6956521739130435,
7432
+ "mmlu_eval_accuracy_human_sexuality": 0.4166666666666667,
7433
+ "mmlu_eval_accuracy_international_law": 0.8461538461538461,
7434
+ "mmlu_eval_accuracy_jurisprudence": 0.45454545454545453,
7435
+ "mmlu_eval_accuracy_logical_fallacies": 0.6111111111111112,
7436
+ "mmlu_eval_accuracy_machine_learning": 0.36363636363636365,
7437
+ "mmlu_eval_accuracy_management": 0.6363636363636364,
7438
+ "mmlu_eval_accuracy_marketing": 0.84,
7439
+ "mmlu_eval_accuracy_medical_genetics": 0.9090909090909091,
7440
+ "mmlu_eval_accuracy_miscellaneous": 0.6744186046511628,
7441
+ "mmlu_eval_accuracy_moral_disputes": 0.47368421052631576,
7442
+ "mmlu_eval_accuracy_moral_scenarios": 0.24,
7443
+ "mmlu_eval_accuracy_nutrition": 0.6060606060606061,
7444
+ "mmlu_eval_accuracy_philosophy": 0.5588235294117647,
7445
+ "mmlu_eval_accuracy_prehistory": 0.5428571428571428,
7446
+ "mmlu_eval_accuracy_professional_accounting": 0.3548387096774194,
7447
+ "mmlu_eval_accuracy_professional_law": 0.32941176470588235,
7448
+ "mmlu_eval_accuracy_professional_medicine": 0.4838709677419355,
7449
+ "mmlu_eval_accuracy_professional_psychology": 0.5072463768115942,
7450
+ "mmlu_eval_accuracy_public_relations": 0.5,
7451
+ "mmlu_eval_accuracy_security_studies": 0.5185185185185185,
7452
+ "mmlu_eval_accuracy_sociology": 0.7272727272727273,
7453
+ "mmlu_eval_accuracy_us_foreign_policy": 0.8181818181818182,
7454
+ "mmlu_eval_accuracy_virology": 0.4444444444444444,
7455
+ "mmlu_eval_accuracy_world_religions": 0.7894736842105263,
7456
+ "mmlu_loss": 1.191394856214212,
7457
+ "step": 7800
7458
+ },
7459
+ {
7460
+ "epoch": 1.64,
7461
+ "learning_rate": 0.0002,
7462
+ "loss": 0.7018,
7463
+ "step": 7810
7464
+ },
7465
+ {
7466
+ "epoch": 1.64,
7467
+ "learning_rate": 0.0002,
7468
+ "loss": 0.6855,
7469
+ "step": 7820
7470
+ },
7471
+ {
7472
+ "epoch": 1.65,
7473
+ "learning_rate": 0.0002,
7474
+ "loss": 0.6327,
7475
+ "step": 7830
7476
+ },
7477
+ {
7478
+ "epoch": 1.65,
7479
+ "learning_rate": 0.0002,
7480
+ "loss": 0.6578,
7481
+ "step": 7840
7482
+ },
7483
+ {
7484
+ "epoch": 1.65,
7485
+ "learning_rate": 0.0002,
7486
+ "loss": 0.6678,
7487
+ "step": 7850
7488
+ },
7489
+ {
7490
+ "epoch": 1.65,
7491
+ "learning_rate": 0.0002,
7492
+ "loss": 0.67,
7493
+ "step": 7860
7494
+ },
7495
+ {
7496
+ "epoch": 1.65,
7497
+ "learning_rate": 0.0002,
7498
+ "loss": 0.6898,
7499
+ "step": 7870
7500
+ },
7501
+ {
7502
+ "epoch": 1.66,
7503
+ "learning_rate": 0.0002,
7504
+ "loss": 0.6412,
7505
+ "step": 7880
7506
+ },
7507
+ {
7508
+ "epoch": 1.66,
7509
+ "learning_rate": 0.0002,
7510
+ "loss": 0.6896,
7511
+ "step": 7890
7512
+ },
7513
+ {
7514
+ "epoch": 1.66,
7515
+ "learning_rate": 0.0002,
7516
+ "loss": 0.652,
7517
+ "step": 7900
7518
+ },
7519
+ {
7520
+ "epoch": 1.66,
7521
+ "learning_rate": 0.0002,
7522
+ "loss": 0.5984,
7523
+ "step": 7910
7524
+ },
7525
+ {
7526
+ "epoch": 1.66,
7527
+ "learning_rate": 0.0002,
7528
+ "loss": 0.6559,
7529
+ "step": 7920
7530
+ },
7531
+ {
7532
+ "epoch": 1.67,
7533
+ "learning_rate": 0.0002,
7534
+ "loss": 0.6171,
7535
+ "step": 7930
7536
+ },
7537
+ {
7538
+ "epoch": 1.67,
7539
+ "learning_rate": 0.0002,
7540
+ "loss": 0.7222,
7541
+ "step": 7940
7542
+ },
7543
+ {
7544
+ "epoch": 1.67,
7545
+ "learning_rate": 0.0002,
7546
+ "loss": 0.6233,
7547
+ "step": 7950
7548
+ },
7549
+ {
7550
+ "epoch": 1.67,
7551
+ "learning_rate": 0.0002,
7552
+ "loss": 0.6479,
7553
+ "step": 7960
7554
+ },
7555
+ {
7556
+ "epoch": 1.68,
7557
+ "learning_rate": 0.0002,
7558
+ "loss": 0.6898,
7559
+ "step": 7970
7560
+ },
7561
+ {
7562
+ "epoch": 1.68,
7563
+ "learning_rate": 0.0002,
7564
+ "loss": 0.6229,
7565
+ "step": 7980
7566
+ },
7567
+ {
7568
+ "epoch": 1.68,
7569
+ "learning_rate": 0.0002,
7570
+ "loss": 0.6232,
7571
+ "step": 7990
7572
+ },
7573
+ {
7574
+ "epoch": 1.68,
7575
+ "learning_rate": 0.0002,
7576
+ "loss": 0.6723,
7577
+ "step": 8000
7578
+ },
7579
+ {
7580
+ "epoch": 1.68,
7581
+ "eval_loss": 0.7203255295753479,
7582
+ "eval_runtime": 68.5826,
7583
+ "eval_samples_per_second": 14.581,
7584
+ "eval_steps_per_second": 7.29,
7585
+ "step": 8000
7586
+ },
7587
+ {
7588
+ "epoch": 1.68,
7589
+ "mmlu_eval_accuracy": 0.5020970296577588,
7590
+ "mmlu_eval_accuracy_abstract_algebra": 0.2727272727272727,
7591
+ "mmlu_eval_accuracy_anatomy": 0.6428571428571429,
7592
+ "mmlu_eval_accuracy_astronomy": 0.4375,
7593
+ "mmlu_eval_accuracy_business_ethics": 0.5454545454545454,
7594
+ "mmlu_eval_accuracy_clinical_knowledge": 0.5172413793103449,
7595
+ "mmlu_eval_accuracy_college_biology": 0.4375,
7596
+ "mmlu_eval_accuracy_college_chemistry": 0.25,
7597
+ "mmlu_eval_accuracy_college_computer_science": 0.36363636363636365,
7598
+ "mmlu_eval_accuracy_college_mathematics": 0.18181818181818182,
7599
+ "mmlu_eval_accuracy_college_medicine": 0.5,
7600
+ "mmlu_eval_accuracy_college_physics": 0.36363636363636365,
7601
+ "mmlu_eval_accuracy_computer_security": 0.5454545454545454,
7602
+ "mmlu_eval_accuracy_conceptual_physics": 0.38461538461538464,
7603
+ "mmlu_eval_accuracy_econometrics": 0.08333333333333333,
7604
+ "mmlu_eval_accuracy_electrical_engineering": 0.25,
7605
+ "mmlu_eval_accuracy_elementary_mathematics": 0.34146341463414637,
7606
+ "mmlu_eval_accuracy_formal_logic": 0.14285714285714285,
7607
+ "mmlu_eval_accuracy_global_facts": 0.5,
7608
+ "mmlu_eval_accuracy_high_school_biology": 0.4375,
7609
+ "mmlu_eval_accuracy_high_school_chemistry": 0.36363636363636365,
7610
+ "mmlu_eval_accuracy_high_school_computer_science": 0.5555555555555556,
7611
+ "mmlu_eval_accuracy_high_school_european_history": 0.5,
7612
+ "mmlu_eval_accuracy_high_school_geography": 0.8636363636363636,
7613
+ "mmlu_eval_accuracy_high_school_government_and_politics": 0.5238095238095238,
7614
+ "mmlu_eval_accuracy_high_school_macroeconomics": 0.46511627906976744,
7615
+ "mmlu_eval_accuracy_high_school_mathematics": 0.13793103448275862,
7616
+ "mmlu_eval_accuracy_high_school_microeconomics": 0.4230769230769231,
7617
+ "mmlu_eval_accuracy_high_school_physics": 0.35294117647058826,
7618
+ "mmlu_eval_accuracy_high_school_psychology": 0.8333333333333334,
7619
+ "mmlu_eval_accuracy_high_school_statistics": 0.43478260869565216,
7620
+ "mmlu_eval_accuracy_high_school_us_history": 0.5909090909090909,
7621
+ "mmlu_eval_accuracy_high_school_world_history": 0.6923076923076923,
7622
+ "mmlu_eval_accuracy_human_aging": 0.782608695652174,
7623
+ "mmlu_eval_accuracy_human_sexuality": 0.4166666666666667,
7624
+ "mmlu_eval_accuracy_international_law": 0.8461538461538461,
7625
+ "mmlu_eval_accuracy_jurisprudence": 0.45454545454545453,
7626
+ "mmlu_eval_accuracy_logical_fallacies": 0.6111111111111112,
7627
+ "mmlu_eval_accuracy_machine_learning": 0.36363636363636365,
7628
+ "mmlu_eval_accuracy_management": 0.7272727272727273,
7629
+ "mmlu_eval_accuracy_marketing": 0.84,
7630
+ "mmlu_eval_accuracy_medical_genetics": 0.9090909090909091,
7631
+ "mmlu_eval_accuracy_miscellaneous": 0.6744186046511628,
7632
+ "mmlu_eval_accuracy_moral_disputes": 0.42105263157894735,
7633
+ "mmlu_eval_accuracy_moral_scenarios": 0.29,
7634
+ "mmlu_eval_accuracy_nutrition": 0.6060606060606061,
7635
+ "mmlu_eval_accuracy_philosophy": 0.5588235294117647,
7636
+ "mmlu_eval_accuracy_prehistory": 0.5142857142857142,
7637
+ "mmlu_eval_accuracy_professional_accounting": 0.41935483870967744,
7638
+ "mmlu_eval_accuracy_professional_law": 0.3352941176470588,
7639
+ "mmlu_eval_accuracy_professional_medicine": 0.45161290322580644,
7640
+ "mmlu_eval_accuracy_professional_psychology": 0.5362318840579711,
7641
+ "mmlu_eval_accuracy_public_relations": 0.4166666666666667,
7642
+ "mmlu_eval_accuracy_security_studies": 0.5185185185185185,
7643
+ "mmlu_eval_accuracy_sociology": 0.7727272727272727,
7644
+ "mmlu_eval_accuracy_us_foreign_policy": 0.8181818181818182,
7645
+ "mmlu_eval_accuracy_virology": 0.6111111111111112,
7646
+ "mmlu_eval_accuracy_world_religions": 0.7894736842105263,
7647
+ "mmlu_loss": 1.179802088762388,
7648
+ "step": 8000
7649
+ },
7650
+ {
7651
+ "epoch": 1.68,
7652
+ "learning_rate": 0.0002,
7653
+ "loss": 0.7006,
7654
+ "step": 8010
7655
+ },
7656
+ {
7657
+ "epoch": 1.69,
7658
+ "learning_rate": 0.0002,
7659
+ "loss": 0.6832,
7660
+ "step": 8020
7661
+ },
7662
+ {
7663
+ "epoch": 1.69,
7664
+ "learning_rate": 0.0002,
7665
+ "loss": 0.6369,
7666
+ "step": 8030
7667
+ },
7668
+ {
7669
+ "epoch": 1.69,
7670
+ "learning_rate": 0.0002,
7671
+ "loss": 0.6072,
7672
+ "step": 8040
7673
+ },
7674
+ {
7675
+ "epoch": 1.69,
7676
+ "learning_rate": 0.0002,
7677
+ "loss": 0.7377,
7678
+ "step": 8050
7679
+ },
7680
+ {
7681
+ "epoch": 1.69,
7682
+ "learning_rate": 0.0002,
7683
+ "loss": 0.665,
7684
+ "step": 8060
7685
+ },
7686
+ {
7687
+ "epoch": 1.7,
7688
+ "learning_rate": 0.0002,
7689
+ "loss": 0.7071,
7690
+ "step": 8070
7691
+ },
7692
+ {
7693
+ "epoch": 1.7,
7694
+ "learning_rate": 0.0002,
7695
+ "loss": 0.6955,
7696
+ "step": 8080
7697
+ },
7698
+ {
7699
+ "epoch": 1.7,
7700
+ "learning_rate": 0.0002,
7701
+ "loss": 0.6439,
7702
+ "step": 8090
7703
+ },
7704
+ {
7705
+ "epoch": 1.7,
7706
+ "learning_rate": 0.0002,
7707
+ "loss": 0.6282,
7708
+ "step": 8100
7709
+ },
7710
+ {
7711
+ "epoch": 1.7,
7712
+ "learning_rate": 0.0002,
7713
+ "loss": 0.6891,
7714
+ "step": 8110
7715
+ },
7716
+ {
7717
+ "epoch": 1.71,
7718
+ "learning_rate": 0.0002,
7719
+ "loss": 0.6347,
7720
+ "step": 8120
7721
+ },
7722
+ {
7723
+ "epoch": 1.71,
7724
+ "learning_rate": 0.0002,
7725
+ "loss": 0.6353,
7726
+ "step": 8130
7727
+ },
7728
+ {
7729
+ "epoch": 1.71,
7730
+ "learning_rate": 0.0002,
7731
+ "loss": 0.6922,
7732
+ "step": 8140
7733
+ },
7734
+ {
7735
+ "epoch": 1.71,
7736
+ "learning_rate": 0.0002,
7737
+ "loss": 0.6889,
7738
+ "step": 8150
7739
+ },
7740
+ {
7741
+ "epoch": 1.72,
7742
+ "learning_rate": 0.0002,
7743
+ "loss": 0.6686,
7744
+ "step": 8160
7745
+ },
7746
+ {
7747
+ "epoch": 1.72,
7748
+ "learning_rate": 0.0002,
7749
+ "loss": 0.6963,
7750
+ "step": 8170
7751
+ },
7752
+ {
7753
+ "epoch": 1.72,
7754
+ "learning_rate": 0.0002,
7755
+ "loss": 0.6383,
7756
+ "step": 8180
7757
+ },
7758
+ {
7759
+ "epoch": 1.72,
7760
+ "learning_rate": 0.0002,
7761
+ "loss": 0.6832,
7762
+ "step": 8190
7763
+ },
7764
+ {
7765
+ "epoch": 1.72,
7766
+ "learning_rate": 0.0002,
7767
+ "loss": 0.6876,
7768
+ "step": 8200
7769
+ },
7770
+ {
7771
+ "epoch": 1.72,
7772
+ "eval_loss": 0.718351423740387,
7773
+ "eval_runtime": 68.543,
7774
+ "eval_samples_per_second": 14.589,
7775
+ "eval_steps_per_second": 7.295,
7776
+ "step": 8200
7777
+ },
7778
+ {
7779
+ "epoch": 1.72,
7780
+ "mmlu_eval_accuracy": 0.4837673768983555,
7781
+ "mmlu_eval_accuracy_abstract_algebra": 0.2727272727272727,
7782
+ "mmlu_eval_accuracy_anatomy": 0.6428571428571429,
7783
+ "mmlu_eval_accuracy_astronomy": 0.5,
7784
+ "mmlu_eval_accuracy_business_ethics": 0.5454545454545454,
7785
+ "mmlu_eval_accuracy_clinical_knowledge": 0.5172413793103449,
7786
+ "mmlu_eval_accuracy_college_biology": 0.4375,
7787
+ "mmlu_eval_accuracy_college_chemistry": 0.25,
7788
+ "mmlu_eval_accuracy_college_computer_science": 0.36363636363636365,
7789
+ "mmlu_eval_accuracy_college_mathematics": 0.18181818181818182,
7790
+ "mmlu_eval_accuracy_college_medicine": 0.36363636363636365,
7791
+ "mmlu_eval_accuracy_college_physics": 0.45454545454545453,
7792
+ "mmlu_eval_accuracy_computer_security": 0.45454545454545453,
7793
+ "mmlu_eval_accuracy_conceptual_physics": 0.4230769230769231,
7794
+ "mmlu_eval_accuracy_econometrics": 0.08333333333333333,
7795
+ "mmlu_eval_accuracy_electrical_engineering": 0.3125,
7796
+ "mmlu_eval_accuracy_elementary_mathematics": 0.34146341463414637,
7797
+ "mmlu_eval_accuracy_formal_logic": 0.21428571428571427,
7798
+ "mmlu_eval_accuracy_global_facts": 0.5,
7799
+ "mmlu_eval_accuracy_high_school_biology": 0.4375,
7800
+ "mmlu_eval_accuracy_high_school_chemistry": 0.3181818181818182,
7801
+ "mmlu_eval_accuracy_high_school_computer_science": 0.5555555555555556,
7802
+ "mmlu_eval_accuracy_high_school_european_history": 0.5,
7803
+ "mmlu_eval_accuracy_high_school_geography": 0.8181818181818182,
7804
+ "mmlu_eval_accuracy_high_school_government_and_politics": 0.5238095238095238,
7805
+ "mmlu_eval_accuracy_high_school_macroeconomics": 0.46511627906976744,
7806
+ "mmlu_eval_accuracy_high_school_mathematics": 0.13793103448275862,
7807
+ "mmlu_eval_accuracy_high_school_microeconomics": 0.34615384615384615,
7808
+ "mmlu_eval_accuracy_high_school_physics": 0.23529411764705882,
7809
+ "mmlu_eval_accuracy_high_school_psychology": 0.8,
7810
+ "mmlu_eval_accuracy_high_school_statistics": 0.30434782608695654,
7811
+ "mmlu_eval_accuracy_high_school_us_history": 0.5909090909090909,
7812
+ "mmlu_eval_accuracy_high_school_world_history": 0.6923076923076923,
7813
+ "mmlu_eval_accuracy_human_aging": 0.782608695652174,
7814
+ "mmlu_eval_accuracy_human_sexuality": 0.4166666666666667,
7815
+ "mmlu_eval_accuracy_international_law": 0.8461538461538461,
7816
+ "mmlu_eval_accuracy_jurisprudence": 0.45454545454545453,
7817
+ "mmlu_eval_accuracy_logical_fallacies": 0.5555555555555556,
7818
+ "mmlu_eval_accuracy_machine_learning": 0.36363636363636365,
7819
+ "mmlu_eval_accuracy_management": 0.5454545454545454,
7820
+ "mmlu_eval_accuracy_marketing": 0.76,
7821
+ "mmlu_eval_accuracy_medical_genetics": 0.9090909090909091,
7822
+ "mmlu_eval_accuracy_miscellaneous": 0.6627906976744186,
7823
+ "mmlu_eval_accuracy_moral_disputes": 0.4473684210526316,
7824
+ "mmlu_eval_accuracy_moral_scenarios": 0.27,
7825
+ "mmlu_eval_accuracy_nutrition": 0.5757575757575758,
7826
+ "mmlu_eval_accuracy_philosophy": 0.5294117647058824,
7827
+ "mmlu_eval_accuracy_prehistory": 0.4857142857142857,
7828
+ "mmlu_eval_accuracy_professional_accounting": 0.41935483870967744,
7829
+ "mmlu_eval_accuracy_professional_law": 0.3176470588235294,
7830
+ "mmlu_eval_accuracy_professional_medicine": 0.5161290322580645,
7831
+ "mmlu_eval_accuracy_professional_psychology": 0.5362318840579711,
7832
+ "mmlu_eval_accuracy_public_relations": 0.4166666666666667,
7833
+ "mmlu_eval_accuracy_security_studies": 0.48148148148148145,
7834
+ "mmlu_eval_accuracy_sociology": 0.7727272727272727,
7835
+ "mmlu_eval_accuracy_us_foreign_policy": 0.6363636363636364,
7836
+ "mmlu_eval_accuracy_virology": 0.5,
7837
+ "mmlu_eval_accuracy_world_religions": 0.7894736842105263,
7838
+ "mmlu_loss": 1.1326499803103605,
7839
+ "step": 8200
7840
+ },
7841
+ {
7842
+ "epoch": 1.73,
7843
+ "learning_rate": 0.0002,
7844
+ "loss": 0.6358,
7845
+ "step": 8210
7846
+ },
7847
+ {
7848
+ "epoch": 1.73,
7849
+ "learning_rate": 0.0002,
7850
+ "loss": 0.6971,
7851
+ "step": 8220
7852
+ },
7853
+ {
7854
+ "epoch": 1.73,
7855
+ "learning_rate": 0.0002,
7856
+ "loss": 0.659,
7857
+ "step": 8230
7858
+ },
7859
+ {
7860
+ "epoch": 1.73,
7861
+ "learning_rate": 0.0002,
7862
+ "loss": 0.6953,
7863
+ "step": 8240
7864
+ },
7865
+ {
7866
+ "epoch": 1.73,
7867
+ "learning_rate": 0.0002,
7868
+ "loss": 0.6798,
7869
+ "step": 8250
7870
+ },
7871
+ {
7872
+ "epoch": 1.74,
7873
+ "learning_rate": 0.0002,
7874
+ "loss": 0.7388,
7875
+ "step": 8260
7876
+ },
7877
+ {
7878
+ "epoch": 1.74,
7879
+ "learning_rate": 0.0002,
7880
+ "loss": 0.6326,
7881
+ "step": 8270
7882
+ },
7883
+ {
7884
+ "epoch": 1.74,
7885
+ "learning_rate": 0.0002,
7886
+ "loss": 0.6421,
7887
+ "step": 8280
7888
+ },
7889
+ {
7890
+ "epoch": 1.74,
7891
+ "learning_rate": 0.0002,
7892
+ "loss": 0.7034,
7893
+ "step": 8290
7894
+ },
7895
+ {
7896
+ "epoch": 1.74,
7897
+ "learning_rate": 0.0002,
7898
+ "loss": 0.6446,
7899
+ "step": 8300
7900
+ },
7901
+ {
7902
+ "epoch": 1.75,
7903
+ "learning_rate": 0.0002,
7904
+ "loss": 0.6453,
7905
+ "step": 8310
7906
+ },
7907
+ {
7908
+ "epoch": 1.75,
7909
+ "learning_rate": 0.0002,
7910
+ "loss": 0.6467,
7911
+ "step": 8320
7912
+ },
7913
+ {
7914
+ "epoch": 1.75,
7915
+ "learning_rate": 0.0002,
7916
+ "loss": 0.682,
7917
+ "step": 8330
7918
+ },
7919
+ {
7920
+ "epoch": 1.75,
7921
+ "learning_rate": 0.0002,
7922
+ "loss": 0.6742,
7923
+ "step": 8340
7924
+ },
7925
+ {
7926
+ "epoch": 1.75,
7927
+ "learning_rate": 0.0002,
7928
+ "loss": 0.6804,
7929
+ "step": 8350
7930
+ },
7931
+ {
7932
+ "epoch": 1.76,
7933
+ "learning_rate": 0.0002,
7934
+ "loss": 0.6372,
7935
+ "step": 8360
7936
+ },
7937
+ {
7938
+ "epoch": 1.76,
7939
+ "learning_rate": 0.0002,
7940
+ "loss": 0.6415,
7941
+ "step": 8370
7942
+ },
7943
+ {
7944
+ "epoch": 1.76,
7945
+ "learning_rate": 0.0002,
7946
+ "loss": 0.6883,
7947
+ "step": 8380
7948
+ },
7949
+ {
7950
+ "epoch": 1.76,
7951
+ "learning_rate": 0.0002,
7952
+ "loss": 0.6623,
7953
+ "step": 8390
7954
+ },
7955
+ {
7956
+ "epoch": 1.77,
7957
+ "learning_rate": 0.0002,
7958
+ "loss": 0.6615,
7959
+ "step": 8400
7960
+ },
7961
+ {
7962
+ "epoch": 1.77,
7963
+ "eval_loss": 0.7216736078262329,
7964
+ "eval_runtime": 69.5248,
7965
+ "eval_samples_per_second": 14.383,
7966
+ "eval_steps_per_second": 7.192,
7967
+ "step": 8400
7968
+ },
7969
+ {
7970
+ "epoch": 1.77,
7971
+ "mmlu_eval_accuracy": 0.4797165293839395,
7972
+ "mmlu_eval_accuracy_abstract_algebra": 0.18181818181818182,
7973
+ "mmlu_eval_accuracy_anatomy": 0.7142857142857143,
7974
+ "mmlu_eval_accuracy_astronomy": 0.5,
7975
+ "mmlu_eval_accuracy_business_ethics": 0.5454545454545454,
7976
+ "mmlu_eval_accuracy_clinical_knowledge": 0.5517241379310345,
7977
+ "mmlu_eval_accuracy_college_biology": 0.375,
7978
+ "mmlu_eval_accuracy_college_chemistry": 0.25,
7979
+ "mmlu_eval_accuracy_college_computer_science": 0.36363636363636365,
7980
+ "mmlu_eval_accuracy_college_mathematics": 0.18181818181818182,
7981
+ "mmlu_eval_accuracy_college_medicine": 0.4090909090909091,
7982
+ "mmlu_eval_accuracy_college_physics": 0.36363636363636365,
7983
+ "mmlu_eval_accuracy_computer_security": 0.2727272727272727,
7984
+ "mmlu_eval_accuracy_conceptual_physics": 0.38461538461538464,
7985
+ "mmlu_eval_accuracy_econometrics": 0.16666666666666666,
7986
+ "mmlu_eval_accuracy_electrical_engineering": 0.3125,
7987
+ "mmlu_eval_accuracy_elementary_mathematics": 0.2926829268292683,
7988
+ "mmlu_eval_accuracy_formal_logic": 0.2857142857142857,
7989
+ "mmlu_eval_accuracy_global_facts": 0.4,
7990
+ "mmlu_eval_accuracy_high_school_biology": 0.3125,
7991
+ "mmlu_eval_accuracy_high_school_chemistry": 0.2727272727272727,
7992
+ "mmlu_eval_accuracy_high_school_computer_science": 0.5555555555555556,
7993
+ "mmlu_eval_accuracy_high_school_european_history": 0.5555555555555556,
7994
+ "mmlu_eval_accuracy_high_school_geography": 0.8636363636363636,
7995
+ "mmlu_eval_accuracy_high_school_government_and_politics": 0.5714285714285714,
7996
+ "mmlu_eval_accuracy_high_school_macroeconomics": 0.46511627906976744,
7997
+ "mmlu_eval_accuracy_high_school_mathematics": 0.1724137931034483,
7998
+ "mmlu_eval_accuracy_high_school_microeconomics": 0.4230769230769231,
7999
+ "mmlu_eval_accuracy_high_school_physics": 0.11764705882352941,
8000
+ "mmlu_eval_accuracy_high_school_psychology": 0.8166666666666667,
8001
+ "mmlu_eval_accuracy_high_school_statistics": 0.30434782608695654,
8002
+ "mmlu_eval_accuracy_high_school_us_history": 0.5454545454545454,
8003
+ "mmlu_eval_accuracy_high_school_world_history": 0.6923076923076923,
8004
+ "mmlu_eval_accuracy_human_aging": 0.7391304347826086,
8005
+ "mmlu_eval_accuracy_human_sexuality": 0.4166666666666667,
8006
+ "mmlu_eval_accuracy_international_law": 0.8461538461538461,
8007
+ "mmlu_eval_accuracy_jurisprudence": 0.36363636363636365,
8008
+ "mmlu_eval_accuracy_logical_fallacies": 0.6111111111111112,
8009
+ "mmlu_eval_accuracy_machine_learning": 0.18181818181818182,
8010
+ "mmlu_eval_accuracy_management": 0.6363636363636364,
8011
+ "mmlu_eval_accuracy_marketing": 0.84,
8012
+ "mmlu_eval_accuracy_medical_genetics": 0.9090909090909091,
8013
+ "mmlu_eval_accuracy_miscellaneous": 0.6627906976744186,
8014
+ "mmlu_eval_accuracy_moral_disputes": 0.39473684210526316,
8015
+ "mmlu_eval_accuracy_moral_scenarios": 0.34,
8016
+ "mmlu_eval_accuracy_nutrition": 0.6060606060606061,
8017
+ "mmlu_eval_accuracy_philosophy": 0.5,
8018
+ "mmlu_eval_accuracy_prehistory": 0.45714285714285713,
8019
+ "mmlu_eval_accuracy_professional_accounting": 0.3870967741935484,
8020
+ "mmlu_eval_accuracy_professional_law": 0.3176470588235294,
8021
+ "mmlu_eval_accuracy_professional_medicine": 0.5161290322580645,
8022
+ "mmlu_eval_accuracy_professional_psychology": 0.5652173913043478,
8023
+ "mmlu_eval_accuracy_public_relations": 0.5,
8024
+ "mmlu_eval_accuracy_security_studies": 0.4074074074074074,
8025
+ "mmlu_eval_accuracy_sociology": 0.8181818181818182,
8026
+ "mmlu_eval_accuracy_us_foreign_policy": 0.8181818181818182,
8027
+ "mmlu_eval_accuracy_virology": 0.5,
8028
+ "mmlu_eval_accuracy_world_religions": 0.7894736842105263,
8029
+ "mmlu_loss": 1.0919515004007063,
8030
+ "step": 8400
8031
+ },
8032
+ {
8033
+ "epoch": 1.77,
8034
+ "learning_rate": 0.0002,
8035
+ "loss": 0.582,
8036
+ "step": 8410
8037
+ },
8038
+ {
8039
+ "epoch": 1.77,
8040
+ "learning_rate": 0.0002,
8041
+ "loss": 0.6839,
8042
+ "step": 8420
8043
+ },
8044
+ {
8045
+ "epoch": 1.77,
8046
+ "learning_rate": 0.0002,
8047
+ "loss": 0.6713,
8048
+ "step": 8430
8049
+ },
8050
+ {
8051
+ "epoch": 1.77,
8052
+ "learning_rate": 0.0002,
8053
+ "loss": 0.7167,
8054
+ "step": 8440
8055
+ },
8056
+ {
8057
+ "epoch": 1.78,
8058
+ "learning_rate": 0.0002,
8059
+ "loss": 0.5882,
8060
+ "step": 8450
8061
+ },
8062
+ {
8063
+ "epoch": 1.78,
8064
+ "learning_rate": 0.0002,
8065
+ "loss": 0.6557,
8066
+ "step": 8460
8067
+ },
8068
+ {
8069
+ "epoch": 1.78,
8070
+ "learning_rate": 0.0002,
8071
+ "loss": 0.6893,
8072
+ "step": 8470
8073
+ },
8074
+ {
8075
+ "epoch": 1.78,
8076
+ "learning_rate": 0.0002,
8077
+ "loss": 0.6681,
8078
+ "step": 8480
8079
+ },
8080
+ {
8081
+ "epoch": 1.78,
8082
+ "learning_rate": 0.0002,
8083
+ "loss": 0.6147,
8084
+ "step": 8490
8085
+ },
8086
+ {
8087
+ "epoch": 1.79,
8088
+ "learning_rate": 0.0002,
8089
+ "loss": 0.698,
8090
+ "step": 8500
8091
+ },
8092
+ {
8093
+ "epoch": 1.79,
8094
+ "learning_rate": 0.0002,
8095
+ "loss": 0.6464,
8096
+ "step": 8510
8097
+ },
8098
+ {
8099
+ "epoch": 1.79,
8100
+ "learning_rate": 0.0002,
8101
+ "loss": 0.7075,
8102
+ "step": 8520
8103
+ },
8104
+ {
8105
+ "epoch": 1.79,
8106
+ "learning_rate": 0.0002,
8107
+ "loss": 0.6394,
8108
+ "step": 8530
8109
+ },
8110
+ {
8111
+ "epoch": 1.79,
8112
+ "learning_rate": 0.0002,
8113
+ "loss": 0.6622,
8114
+ "step": 8540
8115
+ },
8116
+ {
8117
+ "epoch": 1.8,
8118
+ "learning_rate": 0.0002,
8119
+ "loss": 0.7166,
8120
+ "step": 8550
8121
+ },
8122
+ {
8123
+ "epoch": 1.8,
8124
+ "learning_rate": 0.0002,
8125
+ "loss": 0.6544,
8126
+ "step": 8560
8127
+ },
8128
+ {
8129
+ "epoch": 1.8,
8130
+ "learning_rate": 0.0002,
8131
+ "loss": 0.6436,
8132
+ "step": 8570
8133
+ },
8134
+ {
8135
+ "epoch": 1.8,
8136
+ "learning_rate": 0.0002,
8137
+ "loss": 0.688,
8138
+ "step": 8580
8139
+ },
8140
+ {
8141
+ "epoch": 1.81,
8142
+ "learning_rate": 0.0002,
8143
+ "loss": 0.6432,
8144
+ "step": 8590
8145
+ },
8146
+ {
8147
+ "epoch": 1.81,
8148
+ "learning_rate": 0.0002,
8149
+ "loss": 0.6629,
8150
+ "step": 8600
8151
+ },
8152
+ {
8153
+ "epoch": 1.81,
8154
+ "eval_loss": 0.7169492244720459,
8155
+ "eval_runtime": 68.6026,
8156
+ "eval_samples_per_second": 14.577,
8157
+ "eval_steps_per_second": 7.288,
8158
+ "step": 8600
8159
+ },
8160
+ {
8161
+ "epoch": 1.81,
8162
+ "mmlu_eval_accuracy": 0.48553671569218915,
8163
+ "mmlu_eval_accuracy_abstract_algebra": 0.18181818181818182,
8164
+ "mmlu_eval_accuracy_anatomy": 0.6428571428571429,
8165
+ "mmlu_eval_accuracy_astronomy": 0.5,
8166
+ "mmlu_eval_accuracy_business_ethics": 0.5454545454545454,
8167
+ "mmlu_eval_accuracy_clinical_knowledge": 0.5517241379310345,
8168
+ "mmlu_eval_accuracy_college_biology": 0.4375,
8169
+ "mmlu_eval_accuracy_college_chemistry": 0.25,
8170
+ "mmlu_eval_accuracy_college_computer_science": 0.36363636363636365,
8171
+ "mmlu_eval_accuracy_college_mathematics": 0.18181818181818182,
8172
+ "mmlu_eval_accuracy_college_medicine": 0.4090909090909091,
8173
+ "mmlu_eval_accuracy_college_physics": 0.36363636363636365,
8174
+ "mmlu_eval_accuracy_computer_security": 0.2727272727272727,
8175
+ "mmlu_eval_accuracy_conceptual_physics": 0.34615384615384615,
8176
+ "mmlu_eval_accuracy_econometrics": 0.16666666666666666,
8177
+ "mmlu_eval_accuracy_electrical_engineering": 0.3125,
8178
+ "mmlu_eval_accuracy_elementary_mathematics": 0.4146341463414634,
8179
+ "mmlu_eval_accuracy_formal_logic": 0.2857142857142857,
8180
+ "mmlu_eval_accuracy_global_facts": 0.4,
8181
+ "mmlu_eval_accuracy_high_school_biology": 0.34375,
8182
+ "mmlu_eval_accuracy_high_school_chemistry": 0.2727272727272727,
8183
+ "mmlu_eval_accuracy_high_school_computer_science": 0.5555555555555556,
8184
+ "mmlu_eval_accuracy_high_school_european_history": 0.5,
8185
+ "mmlu_eval_accuracy_high_school_geography": 0.8636363636363636,
8186
+ "mmlu_eval_accuracy_high_school_government_and_politics": 0.5714285714285714,
8187
+ "mmlu_eval_accuracy_high_school_macroeconomics": 0.4418604651162791,
8188
+ "mmlu_eval_accuracy_high_school_mathematics": 0.2413793103448276,
8189
+ "mmlu_eval_accuracy_high_school_microeconomics": 0.4230769230769231,
8190
+ "mmlu_eval_accuracy_high_school_physics": 0.17647058823529413,
8191
+ "mmlu_eval_accuracy_high_school_psychology": 0.7833333333333333,
8192
+ "mmlu_eval_accuracy_high_school_statistics": 0.30434782608695654,
8193
+ "mmlu_eval_accuracy_high_school_us_history": 0.5454545454545454,
8194
+ "mmlu_eval_accuracy_high_school_world_history": 0.6538461538461539,
8195
+ "mmlu_eval_accuracy_human_aging": 0.7391304347826086,
8196
+ "mmlu_eval_accuracy_human_sexuality": 0.4166666666666667,
8197
+ "mmlu_eval_accuracy_international_law": 0.8461538461538461,
8198
+ "mmlu_eval_accuracy_jurisprudence": 0.36363636363636365,
8199
+ "mmlu_eval_accuracy_logical_fallacies": 0.5555555555555556,
8200
+ "mmlu_eval_accuracy_machine_learning": 0.2727272727272727,
8201
+ "mmlu_eval_accuracy_management": 0.6363636363636364,
8202
+ "mmlu_eval_accuracy_marketing": 0.84,
8203
+ "mmlu_eval_accuracy_medical_genetics": 0.9090909090909091,
8204
+ "mmlu_eval_accuracy_miscellaneous": 0.6395348837209303,
8205
+ "mmlu_eval_accuracy_moral_disputes": 0.4473684210526316,
8206
+ "mmlu_eval_accuracy_moral_scenarios": 0.33,
8207
+ "mmlu_eval_accuracy_nutrition": 0.6060606060606061,
8208
+ "mmlu_eval_accuracy_philosophy": 0.6176470588235294,
8209
+ "mmlu_eval_accuracy_prehistory": 0.45714285714285713,
8210
+ "mmlu_eval_accuracy_professional_accounting": 0.3548387096774194,
8211
+ "mmlu_eval_accuracy_professional_law": 0.3058823529411765,
8212
+ "mmlu_eval_accuracy_professional_medicine": 0.5161290322580645,
8213
+ "mmlu_eval_accuracy_professional_psychology": 0.5362318840579711,
8214
+ "mmlu_eval_accuracy_public_relations": 0.5,
8215
+ "mmlu_eval_accuracy_security_studies": 0.4074074074074074,
8216
+ "mmlu_eval_accuracy_sociology": 0.8181818181818182,
8217
+ "mmlu_eval_accuracy_us_foreign_policy": 0.9090909090909091,
8218
+ "mmlu_eval_accuracy_virology": 0.6111111111111112,
8219
+ "mmlu_eval_accuracy_world_religions": 0.7368421052631579,
8220
+ "mmlu_loss": 1.1197648563716494,
8221
+ "step": 8600
8222
+ },
8223
+ {
8224
+ "epoch": 1.81,
8225
+ "learning_rate": 0.0002,
8226
+ "loss": 0.6186,
8227
+ "step": 8610
8228
+ },
8229
+ {
8230
+ "epoch": 1.81,
8231
+ "learning_rate": 0.0002,
8232
+ "loss": 0.6588,
8233
+ "step": 8620
8234
+ },
8235
+ {
8236
+ "epoch": 1.81,
8237
+ "learning_rate": 0.0002,
8238
+ "loss": 0.692,
8239
+ "step": 8630
8240
+ },
8241
+ {
8242
+ "epoch": 1.82,
8243
+ "learning_rate": 0.0002,
8244
+ "loss": 0.7083,
8245
+ "step": 8640
8246
+ },
8247
+ {
8248
+ "epoch": 1.82,
8249
+ "learning_rate": 0.0002,
8250
+ "loss": 0.7388,
8251
+ "step": 8650
8252
+ },
8253
+ {
8254
+ "epoch": 1.82,
8255
+ "learning_rate": 0.0002,
8256
+ "loss": 0.6371,
8257
+ "step": 8660
8258
+ },
8259
+ {
8260
+ "epoch": 1.82,
8261
+ "learning_rate": 0.0002,
8262
+ "loss": 0.7297,
8263
+ "step": 8670
8264
+ },
8265
+ {
8266
+ "epoch": 1.82,
8267
+ "learning_rate": 0.0002,
8268
+ "loss": 0.6437,
8269
+ "step": 8680
8270
+ },
8271
+ {
8272
+ "epoch": 1.83,
8273
+ "learning_rate": 0.0002,
8274
+ "loss": 0.6721,
8275
+ "step": 8690
8276
+ },
8277
+ {
8278
+ "epoch": 1.83,
8279
+ "learning_rate": 0.0002,
8280
+ "loss": 0.6727,
8281
+ "step": 8700
8282
+ },
8283
+ {
8284
+ "epoch": 1.83,
8285
+ "learning_rate": 0.0002,
8286
+ "loss": 0.6282,
8287
+ "step": 8710
8288
+ },
8289
+ {
8290
+ "epoch": 1.83,
8291
+ "learning_rate": 0.0002,
8292
+ "loss": 0.7015,
8293
+ "step": 8720
8294
+ },
8295
+ {
8296
+ "epoch": 1.83,
8297
+ "learning_rate": 0.0002,
8298
+ "loss": 0.7122,
8299
+ "step": 8730
8300
+ },
8301
+ {
8302
+ "epoch": 1.84,
8303
+ "learning_rate": 0.0002,
8304
+ "loss": 0.6743,
8305
+ "step": 8740
8306
+ },
8307
+ {
8308
+ "epoch": 1.84,
8309
+ "learning_rate": 0.0002,
8310
+ "loss": 0.6111,
8311
+ "step": 8750
8312
+ },
8313
+ {
8314
+ "epoch": 1.84,
8315
+ "learning_rate": 0.0002,
8316
+ "loss": 0.6789,
8317
+ "step": 8760
8318
+ },
8319
+ {
8320
+ "epoch": 1.84,
8321
+ "learning_rate": 0.0002,
8322
+ "loss": 0.6833,
8323
+ "step": 8770
8324
+ },
8325
+ {
8326
+ "epoch": 1.85,
8327
+ "learning_rate": 0.0002,
8328
+ "loss": 0.6558,
8329
+ "step": 8780
8330
+ },
8331
+ {
8332
+ "epoch": 1.85,
8333
+ "learning_rate": 0.0002,
8334
+ "loss": 0.65,
8335
+ "step": 8790
8336
+ },
8337
+ {
8338
+ "epoch": 1.85,
8339
+ "learning_rate": 0.0002,
8340
+ "loss": 0.7005,
8341
+ "step": 8800
8342
+ },
8343
+ {
8344
+ "epoch": 1.85,
8345
+ "eval_loss": 0.7162238359451294,
8346
+ "eval_runtime": 68.6771,
8347
+ "eval_samples_per_second": 14.561,
8348
+ "eval_steps_per_second": 7.28,
8349
+ "step": 8800
8350
+ },
8351
+ {
8352
+ "epoch": 1.85,
8353
+ "mmlu_eval_accuracy": 0.4885474743321033,
8354
+ "mmlu_eval_accuracy_abstract_algebra": 0.18181818181818182,
8355
+ "mmlu_eval_accuracy_anatomy": 0.6428571428571429,
8356
+ "mmlu_eval_accuracy_astronomy": 0.5,
8357
+ "mmlu_eval_accuracy_business_ethics": 0.5454545454545454,
8358
+ "mmlu_eval_accuracy_clinical_knowledge": 0.5862068965517241,
8359
+ "mmlu_eval_accuracy_college_biology": 0.375,
8360
+ "mmlu_eval_accuracy_college_chemistry": 0.25,
8361
+ "mmlu_eval_accuracy_college_computer_science": 0.36363636363636365,
8362
+ "mmlu_eval_accuracy_college_mathematics": 0.18181818181818182,
8363
+ "mmlu_eval_accuracy_college_medicine": 0.45454545454545453,
8364
+ "mmlu_eval_accuracy_college_physics": 0.36363636363636365,
8365
+ "mmlu_eval_accuracy_computer_security": 0.2727272727272727,
8366
+ "mmlu_eval_accuracy_conceptual_physics": 0.34615384615384615,
8367
+ "mmlu_eval_accuracy_econometrics": 0.25,
8368
+ "mmlu_eval_accuracy_electrical_engineering": 0.3125,
8369
+ "mmlu_eval_accuracy_elementary_mathematics": 0.36585365853658536,
8370
+ "mmlu_eval_accuracy_formal_logic": 0.2857142857142857,
8371
+ "mmlu_eval_accuracy_global_facts": 0.3,
8372
+ "mmlu_eval_accuracy_high_school_biology": 0.375,
8373
+ "mmlu_eval_accuracy_high_school_chemistry": 0.2727272727272727,
8374
+ "mmlu_eval_accuracy_high_school_computer_science": 0.5555555555555556,
8375
+ "mmlu_eval_accuracy_high_school_european_history": 0.5,
8376
+ "mmlu_eval_accuracy_high_school_geography": 0.8636363636363636,
8377
+ "mmlu_eval_accuracy_high_school_government_and_politics": 0.5714285714285714,
8378
+ "mmlu_eval_accuracy_high_school_macroeconomics": 0.4883720930232558,
8379
+ "mmlu_eval_accuracy_high_school_mathematics": 0.20689655172413793,
8380
+ "mmlu_eval_accuracy_high_school_microeconomics": 0.38461538461538464,
8381
+ "mmlu_eval_accuracy_high_school_physics": 0.17647058823529413,
8382
+ "mmlu_eval_accuracy_high_school_psychology": 0.8,
8383
+ "mmlu_eval_accuracy_high_school_statistics": 0.43478260869565216,
8384
+ "mmlu_eval_accuracy_high_school_us_history": 0.6363636363636364,
8385
+ "mmlu_eval_accuracy_high_school_world_history": 0.6538461538461539,
8386
+ "mmlu_eval_accuracy_human_aging": 0.7391304347826086,
8387
+ "mmlu_eval_accuracy_human_sexuality": 0.3333333333333333,
8388
+ "mmlu_eval_accuracy_international_law": 0.8461538461538461,
8389
+ "mmlu_eval_accuracy_jurisprudence": 0.36363636363636365,
8390
+ "mmlu_eval_accuracy_logical_fallacies": 0.5555555555555556,
8391
+ "mmlu_eval_accuracy_machine_learning": 0.36363636363636365,
8392
+ "mmlu_eval_accuracy_management": 0.6363636363636364,
8393
+ "mmlu_eval_accuracy_marketing": 0.84,
8394
+ "mmlu_eval_accuracy_medical_genetics": 0.9090909090909091,
8395
+ "mmlu_eval_accuracy_miscellaneous": 0.6627906976744186,
8396
+ "mmlu_eval_accuracy_moral_disputes": 0.47368421052631576,
8397
+ "mmlu_eval_accuracy_moral_scenarios": 0.28,
8398
+ "mmlu_eval_accuracy_nutrition": 0.6363636363636364,
8399
+ "mmlu_eval_accuracy_philosophy": 0.5588235294117647,
8400
+ "mmlu_eval_accuracy_prehistory": 0.4857142857142857,
8401
+ "mmlu_eval_accuracy_professional_accounting": 0.3225806451612903,
8402
+ "mmlu_eval_accuracy_professional_law": 0.3058823529411765,
8403
+ "mmlu_eval_accuracy_professional_medicine": 0.4838709677419355,
8404
+ "mmlu_eval_accuracy_professional_psychology": 0.5362318840579711,
8405
+ "mmlu_eval_accuracy_public_relations": 0.5,
8406
+ "mmlu_eval_accuracy_security_studies": 0.4444444444444444,
8407
+ "mmlu_eval_accuracy_sociology": 0.8181818181818182,
8408
+ "mmlu_eval_accuracy_us_foreign_policy": 0.9090909090909091,
8409
+ "mmlu_eval_accuracy_virology": 0.5555555555555556,
8410
+ "mmlu_eval_accuracy_world_religions": 0.7894736842105263,
8411
+ "mmlu_loss": 1.1086117945047333,
8412
+ "step": 8800
8413
+ },
8414
+ {
8415
+ "epoch": 1.85,
8416
+ "learning_rate": 0.0002,
8417
+ "loss": 0.6884,
8418
+ "step": 8810
8419
+ },
8420
+ {
8421
+ "epoch": 1.85,
8422
+ "learning_rate": 0.0002,
8423
+ "loss": 0.7064,
8424
+ "step": 8820
8425
+ },
8426
+ {
8427
+ "epoch": 1.86,
8428
+ "learning_rate": 0.0002,
8429
+ "loss": 0.6506,
8430
+ "step": 8830
8431
+ },
8432
+ {
8433
+ "epoch": 1.86,
8434
+ "learning_rate": 0.0002,
8435
+ "loss": 0.6503,
8436
+ "step": 8840
8437
+ },
8438
+ {
8439
+ "epoch": 1.86,
8440
+ "learning_rate": 0.0002,
8441
+ "loss": 0.6504,
8442
+ "step": 8850
8443
+ },
8444
+ {
8445
+ "epoch": 1.86,
8446
+ "learning_rate": 0.0002,
8447
+ "loss": 0.6786,
8448
+ "step": 8860
8449
+ },
8450
+ {
8451
+ "epoch": 1.86,
8452
+ "learning_rate": 0.0002,
8453
+ "loss": 0.6637,
8454
+ "step": 8870
8455
+ },
8456
+ {
8457
+ "epoch": 1.87,
8458
+ "learning_rate": 0.0002,
8459
+ "loss": 0.6356,
8460
+ "step": 8880
8461
+ },
8462
+ {
8463
+ "epoch": 1.87,
8464
+ "learning_rate": 0.0002,
8465
+ "loss": 0.69,
8466
+ "step": 8890
8467
+ },
8468
+ {
8469
+ "epoch": 1.87,
8470
+ "learning_rate": 0.0002,
8471
+ "loss": 0.6655,
8472
+ "step": 8900
8473
+ },
8474
+ {
8475
+ "epoch": 1.87,
8476
+ "learning_rate": 0.0002,
8477
+ "loss": 0.6908,
8478
+ "step": 8910
8479
+ },
8480
+ {
8481
+ "epoch": 1.87,
8482
+ "learning_rate": 0.0002,
8483
+ "loss": 0.6561,
8484
+ "step": 8920
8485
+ },
8486
+ {
8487
+ "epoch": 1.88,
8488
+ "learning_rate": 0.0002,
8489
+ "loss": 0.6508,
8490
+ "step": 8930
8491
+ },
8492
+ {
8493
+ "epoch": 1.88,
8494
+ "learning_rate": 0.0002,
8495
+ "loss": 0.6353,
8496
+ "step": 8940
8497
+ },
8498
+ {
8499
+ "epoch": 1.88,
8500
+ "learning_rate": 0.0002,
8501
+ "loss": 0.6181,
8502
+ "step": 8950
8503
+ },
8504
+ {
8505
+ "epoch": 1.88,
8506
+ "learning_rate": 0.0002,
8507
+ "loss": 0.6407,
8508
+ "step": 8960
8509
+ },
8510
+ {
8511
+ "epoch": 1.89,
8512
+ "learning_rate": 0.0002,
8513
+ "loss": 0.7321,
8514
+ "step": 8970
8515
+ },
8516
+ {
8517
+ "epoch": 1.89,
8518
+ "learning_rate": 0.0002,
8519
+ "loss": 0.6807,
8520
+ "step": 8980
8521
+ },
8522
+ {
8523
+ "epoch": 1.89,
8524
+ "learning_rate": 0.0002,
8525
+ "loss": 0.6902,
8526
+ "step": 8990
8527
+ },
8528
+ {
8529
+ "epoch": 1.89,
8530
+ "learning_rate": 0.0002,
8531
+ "loss": 0.6861,
8532
+ "step": 9000
8533
+ },
8534
+ {
8535
+ "epoch": 1.89,
8536
+ "eval_loss": 0.7179294228553772,
8537
+ "eval_runtime": 69.3479,
8538
+ "eval_samples_per_second": 14.42,
8539
+ "eval_steps_per_second": 7.21,
8540
+ "step": 9000
8541
+ },
8542
+ {
8543
+ "epoch": 1.89,
8544
+ "mmlu_eval_accuracy": 0.48011112874334316,
8545
+ "mmlu_eval_accuracy_abstract_algebra": 0.2727272727272727,
8546
+ "mmlu_eval_accuracy_anatomy": 0.6428571428571429,
8547
+ "mmlu_eval_accuracy_astronomy": 0.5,
8548
+ "mmlu_eval_accuracy_business_ethics": 0.5454545454545454,
8549
+ "mmlu_eval_accuracy_clinical_knowledge": 0.4827586206896552,
8550
+ "mmlu_eval_accuracy_college_biology": 0.3125,
8551
+ "mmlu_eval_accuracy_college_chemistry": 0.25,
8552
+ "mmlu_eval_accuracy_college_computer_science": 0.36363636363636365,
8553
+ "mmlu_eval_accuracy_college_mathematics": 0.2727272727272727,
8554
+ "mmlu_eval_accuracy_college_medicine": 0.5454545454545454,
8555
+ "mmlu_eval_accuracy_college_physics": 0.36363636363636365,
8556
+ "mmlu_eval_accuracy_computer_security": 0.36363636363636365,
8557
+ "mmlu_eval_accuracy_conceptual_physics": 0.34615384615384615,
8558
+ "mmlu_eval_accuracy_econometrics": 0.16666666666666666,
8559
+ "mmlu_eval_accuracy_electrical_engineering": 0.3125,
8560
+ "mmlu_eval_accuracy_elementary_mathematics": 0.3170731707317073,
8561
+ "mmlu_eval_accuracy_formal_logic": 0.21428571428571427,
8562
+ "mmlu_eval_accuracy_global_facts": 0.3,
8563
+ "mmlu_eval_accuracy_high_school_biology": 0.375,
8564
+ "mmlu_eval_accuracy_high_school_chemistry": 0.22727272727272727,
8565
+ "mmlu_eval_accuracy_high_school_computer_science": 0.5555555555555556,
8566
+ "mmlu_eval_accuracy_high_school_european_history": 0.4444444444444444,
8567
+ "mmlu_eval_accuracy_high_school_geography": 0.8636363636363636,
8568
+ "mmlu_eval_accuracy_high_school_government_and_politics": 0.5238095238095238,
8569
+ "mmlu_eval_accuracy_high_school_macroeconomics": 0.4883720930232558,
8570
+ "mmlu_eval_accuracy_high_school_mathematics": 0.27586206896551724,
8571
+ "mmlu_eval_accuracy_high_school_microeconomics": 0.4230769230769231,
8572
+ "mmlu_eval_accuracy_high_school_physics": 0.23529411764705882,
8573
+ "mmlu_eval_accuracy_high_school_psychology": 0.8,
8574
+ "mmlu_eval_accuracy_high_school_statistics": 0.30434782608695654,
8575
+ "mmlu_eval_accuracy_high_school_us_history": 0.6363636363636364,
8576
+ "mmlu_eval_accuracy_high_school_world_history": 0.6923076923076923,
8577
+ "mmlu_eval_accuracy_human_aging": 0.7391304347826086,
8578
+ "mmlu_eval_accuracy_human_sexuality": 0.3333333333333333,
8579
+ "mmlu_eval_accuracy_international_law": 0.8461538461538461,
8580
+ "mmlu_eval_accuracy_jurisprudence": 0.36363636363636365,
8581
+ "mmlu_eval_accuracy_logical_fallacies": 0.5555555555555556,
8582
+ "mmlu_eval_accuracy_machine_learning": 0.2727272727272727,
8583
+ "mmlu_eval_accuracy_management": 0.6363636363636364,
8584
+ "mmlu_eval_accuracy_marketing": 0.84,
8585
+ "mmlu_eval_accuracy_medical_genetics": 0.9090909090909091,
8586
+ "mmlu_eval_accuracy_miscellaneous": 0.6511627906976745,
8587
+ "mmlu_eval_accuracy_moral_disputes": 0.4473684210526316,
8588
+ "mmlu_eval_accuracy_moral_scenarios": 0.26,
8589
+ "mmlu_eval_accuracy_nutrition": 0.5757575757575758,
8590
+ "mmlu_eval_accuracy_philosophy": 0.5588235294117647,
8591
+ "mmlu_eval_accuracy_prehistory": 0.45714285714285713,
8592
+ "mmlu_eval_accuracy_professional_accounting": 0.3548387096774194,
8593
+ "mmlu_eval_accuracy_professional_law": 0.3058823529411765,
8594
+ "mmlu_eval_accuracy_professional_medicine": 0.5483870967741935,
8595
+ "mmlu_eval_accuracy_professional_psychology": 0.4927536231884058,
8596
+ "mmlu_eval_accuracy_public_relations": 0.5,
8597
+ "mmlu_eval_accuracy_security_studies": 0.4074074074074074,
8598
+ "mmlu_eval_accuracy_sociology": 0.7272727272727273,
8599
+ "mmlu_eval_accuracy_us_foreign_policy": 0.8181818181818182,
8600
+ "mmlu_eval_accuracy_virology": 0.6111111111111112,
8601
+ "mmlu_eval_accuracy_world_religions": 0.7368421052631579,
8602
+ "mmlu_loss": 1.2108071716429358,
8603
+ "step": 9000
8604
+ },
8605
+ {
8606
+ "epoch": 1.89,
8607
+ "learning_rate": 0.0002,
8608
+ "loss": 0.6864,
8609
+ "step": 9010
8610
+ },
8611
+ {
8612
+ "epoch": 1.9,
8613
+ "learning_rate": 0.0002,
8614
+ "loss": 0.6036,
8615
+ "step": 9020
8616
+ },
8617
+ {
8618
+ "epoch": 1.9,
8619
+ "learning_rate": 0.0002,
8620
+ "loss": 0.649,
8621
+ "step": 9030
8622
+ },
8623
+ {
8624
+ "epoch": 1.9,
8625
+ "learning_rate": 0.0002,
8626
+ "loss": 0.6724,
8627
+ "step": 9040
8628
+ },
8629
+ {
8630
+ "epoch": 1.9,
8631
+ "learning_rate": 0.0002,
8632
+ "loss": 0.6896,
8633
+ "step": 9050
8634
+ },
8635
+ {
8636
+ "epoch": 1.9,
8637
+ "learning_rate": 0.0002,
8638
+ "loss": 0.6677,
8639
+ "step": 9060
8640
+ },
8641
+ {
8642
+ "epoch": 1.91,
8643
+ "learning_rate": 0.0002,
8644
+ "loss": 0.6193,
8645
+ "step": 9070
8646
+ },
8647
+ {
8648
+ "epoch": 1.91,
8649
+ "learning_rate": 0.0002,
8650
+ "loss": 0.6674,
8651
+ "step": 9080
8652
+ },
8653
+ {
8654
+ "epoch": 1.91,
8655
+ "learning_rate": 0.0002,
8656
+ "loss": 0.6851,
8657
+ "step": 9090
8658
+ },
8659
+ {
8660
+ "epoch": 1.91,
8661
+ "learning_rate": 0.0002,
8662
+ "loss": 0.6879,
8663
+ "step": 9100
8664
+ },
8665
+ {
8666
+ "epoch": 1.91,
8667
+ "learning_rate": 0.0002,
8668
+ "loss": 0.6749,
8669
+ "step": 9110
8670
+ },
8671
+ {
8672
+ "epoch": 1.92,
8673
+ "learning_rate": 0.0002,
8674
+ "loss": 0.626,
8675
+ "step": 9120
8676
+ },
8677
+ {
8678
+ "epoch": 1.92,
8679
+ "learning_rate": 0.0002,
8680
+ "loss": 0.6636,
8681
+ "step": 9130
8682
+ },
8683
+ {
8684
+ "epoch": 1.92,
8685
+ "learning_rate": 0.0002,
8686
+ "loss": 0.6193,
8687
+ "step": 9140
8688
+ },
8689
+ {
8690
+ "epoch": 1.92,
8691
+ "learning_rate": 0.0002,
8692
+ "loss": 0.683,
8693
+ "step": 9150
8694
+ },
8695
+ {
8696
+ "epoch": 1.93,
8697
+ "learning_rate": 0.0002,
8698
+ "loss": 0.6418,
8699
+ "step": 9160
8700
+ },
8701
+ {
8702
+ "epoch": 1.93,
8703
+ "learning_rate": 0.0002,
8704
+ "loss": 0.6126,
8705
+ "step": 9170
8706
+ },
8707
+ {
8708
+ "epoch": 1.93,
8709
+ "learning_rate": 0.0002,
8710
+ "loss": 0.6308,
8711
+ "step": 9180
8712
+ },
8713
+ {
8714
+ "epoch": 1.93,
8715
+ "learning_rate": 0.0002,
8716
+ "loss": 0.623,
8717
+ "step": 9190
8718
+ },
8719
+ {
8720
+ "epoch": 1.93,
8721
+ "learning_rate": 0.0002,
8722
+ "loss": 0.697,
8723
+ "step": 9200
8724
+ },
8725
+ {
8726
+ "epoch": 1.93,
8727
+ "eval_loss": 0.7165359854698181,
8728
+ "eval_runtime": 68.5205,
8729
+ "eval_samples_per_second": 14.594,
8730
+ "eval_steps_per_second": 7.297,
8731
+ "step": 9200
8732
+ },
8733
+ {
8734
+ "epoch": 1.93,
8735
+ "mmlu_eval_accuracy": 0.47839592190982055,
8736
+ "mmlu_eval_accuracy_abstract_algebra": 0.2727272727272727,
8737
+ "mmlu_eval_accuracy_anatomy": 0.6428571428571429,
8738
+ "mmlu_eval_accuracy_astronomy": 0.5,
8739
+ "mmlu_eval_accuracy_business_ethics": 0.6363636363636364,
8740
+ "mmlu_eval_accuracy_clinical_knowledge": 0.4827586206896552,
8741
+ "mmlu_eval_accuracy_college_biology": 0.4375,
8742
+ "mmlu_eval_accuracy_college_chemistry": 0.25,
8743
+ "mmlu_eval_accuracy_college_computer_science": 0.36363636363636365,
8744
+ "mmlu_eval_accuracy_college_mathematics": 0.18181818181818182,
8745
+ "mmlu_eval_accuracy_college_medicine": 0.45454545454545453,
8746
+ "mmlu_eval_accuracy_college_physics": 0.2727272727272727,
8747
+ "mmlu_eval_accuracy_computer_security": 0.36363636363636365,
8748
+ "mmlu_eval_accuracy_conceptual_physics": 0.34615384615384615,
8749
+ "mmlu_eval_accuracy_econometrics": 0.16666666666666666,
8750
+ "mmlu_eval_accuracy_electrical_engineering": 0.3125,
8751
+ "mmlu_eval_accuracy_elementary_mathematics": 0.34146341463414637,
8752
+ "mmlu_eval_accuracy_formal_logic": 0.21428571428571427,
8753
+ "mmlu_eval_accuracy_global_facts": 0.3,
8754
+ "mmlu_eval_accuracy_high_school_biology": 0.3125,
8755
+ "mmlu_eval_accuracy_high_school_chemistry": 0.2727272727272727,
8756
+ "mmlu_eval_accuracy_high_school_computer_science": 0.5555555555555556,
8757
+ "mmlu_eval_accuracy_high_school_european_history": 0.4444444444444444,
8758
+ "mmlu_eval_accuracy_high_school_geography": 0.8636363636363636,
8759
+ "mmlu_eval_accuracy_high_school_government_and_politics": 0.5238095238095238,
8760
+ "mmlu_eval_accuracy_high_school_macroeconomics": 0.3953488372093023,
8761
+ "mmlu_eval_accuracy_high_school_mathematics": 0.1724137931034483,
8762
+ "mmlu_eval_accuracy_high_school_microeconomics": 0.4230769230769231,
8763
+ "mmlu_eval_accuracy_high_school_physics": 0.17647058823529413,
8764
+ "mmlu_eval_accuracy_high_school_psychology": 0.7833333333333333,
8765
+ "mmlu_eval_accuracy_high_school_statistics": 0.30434782608695654,
8766
+ "mmlu_eval_accuracy_high_school_us_history": 0.5,
8767
+ "mmlu_eval_accuracy_high_school_world_history": 0.6538461538461539,
8768
+ "mmlu_eval_accuracy_human_aging": 0.7391304347826086,
8769
+ "mmlu_eval_accuracy_human_sexuality": 0.4166666666666667,
8770
+ "mmlu_eval_accuracy_international_law": 0.8461538461538461,
8771
+ "mmlu_eval_accuracy_jurisprudence": 0.36363636363636365,
8772
+ "mmlu_eval_accuracy_logical_fallacies": 0.5555555555555556,
8773
+ "mmlu_eval_accuracy_machine_learning": 0.36363636363636365,
8774
+ "mmlu_eval_accuracy_management": 0.6363636363636364,
8775
+ "mmlu_eval_accuracy_marketing": 0.88,
8776
+ "mmlu_eval_accuracy_medical_genetics": 0.9090909090909091,
8777
+ "mmlu_eval_accuracy_miscellaneous": 0.686046511627907,
8778
+ "mmlu_eval_accuracy_moral_disputes": 0.4473684210526316,
8779
+ "mmlu_eval_accuracy_moral_scenarios": 0.3,
8780
+ "mmlu_eval_accuracy_nutrition": 0.6363636363636364,
8781
+ "mmlu_eval_accuracy_philosophy": 0.5882352941176471,
8782
+ "mmlu_eval_accuracy_prehistory": 0.42857142857142855,
8783
+ "mmlu_eval_accuracy_professional_accounting": 0.3548387096774194,
8784
+ "mmlu_eval_accuracy_professional_law": 0.28823529411764703,
8785
+ "mmlu_eval_accuracy_professional_medicine": 0.5161290322580645,
8786
+ "mmlu_eval_accuracy_professional_psychology": 0.5072463768115942,
8787
+ "mmlu_eval_accuracy_public_relations": 0.5833333333333334,
8788
+ "mmlu_eval_accuracy_security_studies": 0.4074074074074074,
8789
+ "mmlu_eval_accuracy_sociology": 0.7272727272727273,
8790
+ "mmlu_eval_accuracy_us_foreign_policy": 0.8181818181818182,
8791
+ "mmlu_eval_accuracy_virology": 0.6111111111111112,
8792
+ "mmlu_eval_accuracy_world_religions": 0.7368421052631579,
8793
+ "mmlu_loss": 1.1566752298071241,
8794
+ "step": 9200
8795
+ },
8796
+ {
8797
+ "epoch": 1.94,
8798
+ "learning_rate": 0.0002,
8799
+ "loss": 0.7178,
8800
+ "step": 9210
8801
+ },
8802
+ {
8803
+ "epoch": 1.94,
8804
+ "learning_rate": 0.0002,
8805
+ "loss": 0.7193,
8806
+ "step": 9220
8807
+ },
8808
+ {
8809
+ "epoch": 1.94,
8810
+ "learning_rate": 0.0002,
8811
+ "loss": 0.6901,
8812
+ "step": 9230
8813
+ },
8814
+ {
8815
+ "epoch": 1.94,
8816
+ "learning_rate": 0.0002,
8817
+ "loss": 0.6789,
8818
+ "step": 9240
8819
+ },
8820
+ {
8821
+ "epoch": 1.94,
8822
+ "learning_rate": 0.0002,
8823
+ "loss": 0.6611,
8824
+ "step": 9250
8825
+ },
8826
+ {
8827
+ "epoch": 1.95,
8828
+ "learning_rate": 0.0002,
8829
+ "loss": 0.7138,
8830
+ "step": 9260
8831
+ },
8832
+ {
8833
+ "epoch": 1.95,
8834
+ "learning_rate": 0.0002,
8835
+ "loss": 0.6949,
8836
+ "step": 9270
8837
+ },
8838
+ {
8839
+ "epoch": 1.95,
8840
+ "learning_rate": 0.0002,
8841
+ "loss": 0.6799,
8842
+ "step": 9280
8843
+ },
8844
+ {
8845
+ "epoch": 1.95,
8846
+ "learning_rate": 0.0002,
8847
+ "loss": 0.6668,
8848
+ "step": 9290
8849
+ },
8850
+ {
8851
+ "epoch": 1.95,
8852
+ "learning_rate": 0.0002,
8853
+ "loss": 0.6178,
8854
+ "step": 9300
8855
+ },
8856
+ {
8857
+ "epoch": 1.96,
8858
+ "learning_rate": 0.0002,
8859
+ "loss": 0.7063,
8860
+ "step": 9310
8861
+ },
8862
+ {
8863
+ "epoch": 1.96,
8864
+ "learning_rate": 0.0002,
8865
+ "loss": 0.7176,
8866
+ "step": 9320
8867
+ },
8868
+ {
8869
+ "epoch": 1.96,
8870
+ "learning_rate": 0.0002,
8871
+ "loss": 0.672,
8872
+ "step": 9330
8873
+ },
8874
+ {
8875
+ "epoch": 1.96,
8876
+ "learning_rate": 0.0002,
8877
+ "loss": 0.6902,
8878
+ "step": 9340
8879
+ },
8880
+ {
8881
+ "epoch": 1.97,
8882
+ "learning_rate": 0.0002,
8883
+ "loss": 0.689,
8884
+ "step": 9350
8885
+ },
8886
+ {
8887
+ "epoch": 1.97,
8888
+ "learning_rate": 0.0002,
8889
+ "loss": 0.6885,
8890
+ "step": 9360
8891
+ },
8892
+ {
8893
+ "epoch": 1.97,
8894
+ "learning_rate": 0.0002,
8895
+ "loss": 0.7043,
8896
+ "step": 9370
8897
+ },
8898
+ {
8899
+ "epoch": 1.97,
8900
+ "learning_rate": 0.0002,
8901
+ "loss": 0.6763,
8902
+ "step": 9380
8903
+ },
8904
+ {
8905
+ "epoch": 1.97,
8906
+ "learning_rate": 0.0002,
8907
+ "loss": 0.6469,
8908
+ "step": 9390
8909
+ },
8910
+ {
8911
+ "epoch": 1.98,
8912
+ "learning_rate": 0.0002,
8913
+ "loss": 0.6543,
8914
+ "step": 9400
8915
+ },
8916
+ {
8917
+ "epoch": 1.98,
8918
+ "eval_loss": 0.7109686732292175,
8919
+ "eval_runtime": 69.0283,
8920
+ "eval_samples_per_second": 14.487,
8921
+ "eval_steps_per_second": 7.243,
8922
+ "step": 9400
8923
+ },
8924
+ {
8925
+ "epoch": 1.98,
8926
+ "mmlu_eval_accuracy": 0.4702978654500649,
8927
+ "mmlu_eval_accuracy_abstract_algebra": 0.2727272727272727,
8928
+ "mmlu_eval_accuracy_anatomy": 0.5714285714285714,
8929
+ "mmlu_eval_accuracy_astronomy": 0.5,
8930
+ "mmlu_eval_accuracy_business_ethics": 0.6363636363636364,
8931
+ "mmlu_eval_accuracy_clinical_knowledge": 0.4827586206896552,
8932
+ "mmlu_eval_accuracy_college_biology": 0.375,
8933
+ "mmlu_eval_accuracy_college_chemistry": 0.25,
8934
+ "mmlu_eval_accuracy_college_computer_science": 0.36363636363636365,
8935
+ "mmlu_eval_accuracy_college_mathematics": 0.18181818181818182,
8936
+ "mmlu_eval_accuracy_college_medicine": 0.36363636363636365,
8937
+ "mmlu_eval_accuracy_college_physics": 0.2727272727272727,
8938
+ "mmlu_eval_accuracy_computer_security": 0.2727272727272727,
8939
+ "mmlu_eval_accuracy_conceptual_physics": 0.34615384615384615,
8940
+ "mmlu_eval_accuracy_econometrics": 0.16666666666666666,
8941
+ "mmlu_eval_accuracy_electrical_engineering": 0.3125,
8942
+ "mmlu_eval_accuracy_elementary_mathematics": 0.34146341463414637,
8943
+ "mmlu_eval_accuracy_formal_logic": 0.2857142857142857,
8944
+ "mmlu_eval_accuracy_global_facts": 0.2,
8945
+ "mmlu_eval_accuracy_high_school_biology": 0.3125,
8946
+ "mmlu_eval_accuracy_high_school_chemistry": 0.2727272727272727,
8947
+ "mmlu_eval_accuracy_high_school_computer_science": 0.5555555555555556,
8948
+ "mmlu_eval_accuracy_high_school_european_history": 0.4444444444444444,
8949
+ "mmlu_eval_accuracy_high_school_geography": 0.8636363636363636,
8950
+ "mmlu_eval_accuracy_high_school_government_and_politics": 0.5714285714285714,
8951
+ "mmlu_eval_accuracy_high_school_macroeconomics": 0.4418604651162791,
8952
+ "mmlu_eval_accuracy_high_school_mathematics": 0.2413793103448276,
8953
+ "mmlu_eval_accuracy_high_school_microeconomics": 0.34615384615384615,
8954
+ "mmlu_eval_accuracy_high_school_physics": 0.23529411764705882,
8955
+ "mmlu_eval_accuracy_high_school_psychology": 0.75,
8956
+ "mmlu_eval_accuracy_high_school_statistics": 0.30434782608695654,
8957
+ "mmlu_eval_accuracy_high_school_us_history": 0.5454545454545454,
8958
+ "mmlu_eval_accuracy_high_school_world_history": 0.6538461538461539,
8959
+ "mmlu_eval_accuracy_human_aging": 0.7391304347826086,
8960
+ "mmlu_eval_accuracy_human_sexuality": 0.3333333333333333,
8961
+ "mmlu_eval_accuracy_international_law": 0.8461538461538461,
8962
+ "mmlu_eval_accuracy_jurisprudence": 0.36363636363636365,
8963
+ "mmlu_eval_accuracy_logical_fallacies": 0.5555555555555556,
8964
+ "mmlu_eval_accuracy_machine_learning": 0.2727272727272727,
8965
+ "mmlu_eval_accuracy_management": 0.6363636363636364,
8966
+ "mmlu_eval_accuracy_marketing": 0.88,
8967
+ "mmlu_eval_accuracy_medical_genetics": 0.9090909090909091,
8968
+ "mmlu_eval_accuracy_miscellaneous": 0.6744186046511628,
8969
+ "mmlu_eval_accuracy_moral_disputes": 0.42105263157894735,
8970
+ "mmlu_eval_accuracy_moral_scenarios": 0.3,
8971
+ "mmlu_eval_accuracy_nutrition": 0.6060606060606061,
8972
+ "mmlu_eval_accuracy_philosophy": 0.5882352941176471,
8973
+ "mmlu_eval_accuracy_prehistory": 0.4857142857142857,
8974
+ "mmlu_eval_accuracy_professional_accounting": 0.3548387096774194,
8975
+ "mmlu_eval_accuracy_professional_law": 0.28823529411764703,
8976
+ "mmlu_eval_accuracy_professional_medicine": 0.5161290322580645,
8977
+ "mmlu_eval_accuracy_professional_psychology": 0.5217391304347826,
8978
+ "mmlu_eval_accuracy_public_relations": 0.5,
8979
+ "mmlu_eval_accuracy_security_studies": 0.4074074074074074,
8980
+ "mmlu_eval_accuracy_sociology": 0.7272727272727273,
8981
+ "mmlu_eval_accuracy_us_foreign_policy": 0.9090909090909091,
8982
+ "mmlu_eval_accuracy_virology": 0.5,
8983
+ "mmlu_eval_accuracy_world_religions": 0.7368421052631579,
8984
+ "mmlu_loss": 1.1730008564169663,
8985
+ "step": 9400
8986
+ },
8987
+ {
8988
+ "epoch": 1.98,
8989
+ "learning_rate": 0.0002,
8990
+ "loss": 0.6513,
8991
+ "step": 9410
8992
+ },
8993
+ {
8994
+ "epoch": 1.98,
8995
+ "learning_rate": 0.0002,
8996
+ "loss": 0.6653,
8997
+ "step": 9420
8998
+ },
8999
+ {
9000
+ "epoch": 1.98,
9001
+ "learning_rate": 0.0002,
9002
+ "loss": 0.7043,
9003
+ "step": 9430
9004
+ },
9005
+ {
9006
+ "epoch": 1.98,
9007
+ "learning_rate": 0.0002,
9008
+ "loss": 0.6812,
9009
+ "step": 9440
9010
+ },
9011
+ {
9012
+ "epoch": 1.99,
9013
+ "learning_rate": 0.0002,
9014
+ "loss": 0.6883,
9015
+ "step": 9450
9016
+ },
9017
+ {
9018
+ "epoch": 1.99,
9019
+ "learning_rate": 0.0002,
9020
+ "loss": 0.6441,
9021
+ "step": 9460
9022
+ },
9023
+ {
9024
+ "epoch": 1.99,
9025
+ "learning_rate": 0.0002,
9026
+ "loss": 0.6676,
9027
+ "step": 9470
9028
+ },
9029
+ {
9030
+ "epoch": 1.99,
9031
+ "learning_rate": 0.0002,
9032
+ "loss": 0.7076,
9033
+ "step": 9480
9034
+ },
9035
+ {
9036
+ "epoch": 1.99,
9037
+ "learning_rate": 0.0002,
9038
+ "loss": 0.6783,
9039
+ "step": 9490
9040
+ },
9041
+ {
9042
+ "epoch": 2.0,
9043
+ "learning_rate": 0.0002,
9044
+ "loss": 0.7126,
9045
+ "step": 9500
9046
+ },
9047
+ {
9048
+ "epoch": 2.0,
9049
+ "learning_rate": 0.0002,
9050
+ "loss": 0.628,
9051
+ "step": 9510
9052
+ },
9053
+ {
9054
+ "epoch": 2.0,
9055
+ "learning_rate": 0.0002,
9056
+ "loss": 0.6724,
9057
+ "step": 9520
9058
+ },
9059
+ {
9060
+ "epoch": 2.0,
9061
+ "learning_rate": 0.0002,
9062
+ "loss": 0.5755,
9063
+ "step": 9530
9064
+ },
9065
+ {
9066
+ "epoch": 2.01,
9067
+ "learning_rate": 0.0002,
9068
+ "loss": 0.5531,
9069
+ "step": 9540
9070
+ },
9071
+ {
9072
+ "epoch": 2.01,
9073
+ "learning_rate": 0.0002,
9074
+ "loss": 0.5657,
9075
+ "step": 9550
9076
+ },
9077
+ {
9078
+ "epoch": 2.01,
9079
+ "learning_rate": 0.0002,
9080
+ "loss": 0.5821,
9081
+ "step": 9560
9082
+ },
9083
+ {
9084
+ "epoch": 2.01,
9085
+ "learning_rate": 0.0002,
9086
+ "loss": 0.5723,
9087
+ "step": 9570
9088
+ },
9089
+ {
9090
+ "epoch": 2.01,
9091
+ "learning_rate": 0.0002,
9092
+ "loss": 0.6019,
9093
+ "step": 9580
9094
+ },
9095
+ {
9096
+ "epoch": 2.02,
9097
+ "learning_rate": 0.0002,
9098
+ "loss": 0.596,
9099
+ "step": 9590
9100
+ },
9101
+ {
9102
+ "epoch": 2.02,
9103
+ "learning_rate": 0.0002,
9104
+ "loss": 0.5986,
9105
+ "step": 9600
9106
+ },
9107
+ {
9108
+ "epoch": 2.02,
9109
+ "eval_loss": 0.7332544922828674,
9110
+ "eval_runtime": 68.8204,
9111
+ "eval_samples_per_second": 14.531,
9112
+ "eval_steps_per_second": 7.265,
9113
+ "step": 9600
9114
+ },
9115
+ {
9116
+ "epoch": 2.02,
9117
+ "mmlu_eval_accuracy": 0.48391071516098805,
9118
+ "mmlu_eval_accuracy_abstract_algebra": 0.18181818181818182,
9119
+ "mmlu_eval_accuracy_anatomy": 0.5714285714285714,
9120
+ "mmlu_eval_accuracy_astronomy": 0.5,
9121
+ "mmlu_eval_accuracy_business_ethics": 0.6363636363636364,
9122
+ "mmlu_eval_accuracy_clinical_knowledge": 0.4827586206896552,
9123
+ "mmlu_eval_accuracy_college_biology": 0.4375,
9124
+ "mmlu_eval_accuracy_college_chemistry": 0.25,
9125
+ "mmlu_eval_accuracy_college_computer_science": 0.45454545454545453,
9126
+ "mmlu_eval_accuracy_college_mathematics": 0.09090909090909091,
9127
+ "mmlu_eval_accuracy_college_medicine": 0.4090909090909091,
9128
+ "mmlu_eval_accuracy_college_physics": 0.18181818181818182,
9129
+ "mmlu_eval_accuracy_computer_security": 0.36363636363636365,
9130
+ "mmlu_eval_accuracy_conceptual_physics": 0.38461538461538464,
9131
+ "mmlu_eval_accuracy_econometrics": 0.25,
9132
+ "mmlu_eval_accuracy_electrical_engineering": 0.375,
9133
+ "mmlu_eval_accuracy_elementary_mathematics": 0.3170731707317073,
9134
+ "mmlu_eval_accuracy_formal_logic": 0.35714285714285715,
9135
+ "mmlu_eval_accuracy_global_facts": 0.3,
9136
+ "mmlu_eval_accuracy_high_school_biology": 0.34375,
9137
+ "mmlu_eval_accuracy_high_school_chemistry": 0.2727272727272727,
9138
+ "mmlu_eval_accuracy_high_school_computer_science": 0.5555555555555556,
9139
+ "mmlu_eval_accuracy_high_school_european_history": 0.5,
9140
+ "mmlu_eval_accuracy_high_school_geography": 0.8636363636363636,
9141
+ "mmlu_eval_accuracy_high_school_government_and_politics": 0.5238095238095238,
9142
+ "mmlu_eval_accuracy_high_school_macroeconomics": 0.4418604651162791,
9143
+ "mmlu_eval_accuracy_high_school_mathematics": 0.2413793103448276,
9144
+ "mmlu_eval_accuracy_high_school_microeconomics": 0.34615384615384615,
9145
+ "mmlu_eval_accuracy_high_school_physics": 0.29411764705882354,
9146
+ "mmlu_eval_accuracy_high_school_psychology": 0.7833333333333333,
9147
+ "mmlu_eval_accuracy_high_school_statistics": 0.391304347826087,
9148
+ "mmlu_eval_accuracy_high_school_us_history": 0.5,
9149
+ "mmlu_eval_accuracy_high_school_world_history": 0.6538461538461539,
9150
+ "mmlu_eval_accuracy_human_aging": 0.782608695652174,
9151
+ "mmlu_eval_accuracy_human_sexuality": 0.4166666666666667,
9152
+ "mmlu_eval_accuracy_international_law": 0.8461538461538461,
9153
+ "mmlu_eval_accuracy_jurisprudence": 0.45454545454545453,
9154
+ "mmlu_eval_accuracy_logical_fallacies": 0.5,
9155
+ "mmlu_eval_accuracy_machine_learning": 0.2727272727272727,
9156
+ "mmlu_eval_accuracy_management": 0.7272727272727273,
9157
+ "mmlu_eval_accuracy_marketing": 0.88,
9158
+ "mmlu_eval_accuracy_medical_genetics": 0.9090909090909091,
9159
+ "mmlu_eval_accuracy_miscellaneous": 0.6744186046511628,
9160
+ "mmlu_eval_accuracy_moral_disputes": 0.5,
9161
+ "mmlu_eval_accuracy_moral_scenarios": 0.22,
9162
+ "mmlu_eval_accuracy_nutrition": 0.6363636363636364,
9163
+ "mmlu_eval_accuracy_philosophy": 0.5882352941176471,
9164
+ "mmlu_eval_accuracy_prehistory": 0.4857142857142857,
9165
+ "mmlu_eval_accuracy_professional_accounting": 0.3870967741935484,
9166
+ "mmlu_eval_accuracy_professional_law": 0.29411764705882354,
9167
+ "mmlu_eval_accuracy_professional_medicine": 0.45161290322580644,
9168
+ "mmlu_eval_accuracy_professional_psychology": 0.5217391304347826,
9169
+ "mmlu_eval_accuracy_public_relations": 0.5,
9170
+ "mmlu_eval_accuracy_security_studies": 0.4444444444444444,
9171
+ "mmlu_eval_accuracy_sociology": 0.7272727272727273,
9172
+ "mmlu_eval_accuracy_us_foreign_policy": 0.8181818181818182,
9173
+ "mmlu_eval_accuracy_virology": 0.5,
9174
+ "mmlu_eval_accuracy_world_religions": 0.7894736842105263,
9175
+ "mmlu_loss": 1.2143672528450544,
9176
+ "step": 9600
9177
+ },
9178
+ {
9179
+ "epoch": 2.02,
9180
+ "learning_rate": 0.0002,
9181
+ "loss": 0.5681,
9182
+ "step": 9610
9183
+ },
9184
+ {
9185
+ "epoch": 2.02,
9186
+ "learning_rate": 0.0002,
9187
+ "loss": 0.5552,
9188
+ "step": 9620
9189
+ },
9190
+ {
9191
+ "epoch": 2.02,
9192
+ "learning_rate": 0.0002,
9193
+ "loss": 0.5512,
9194
+ "step": 9630
9195
+ },
9196
+ {
9197
+ "epoch": 2.03,
9198
+ "learning_rate": 0.0002,
9199
+ "loss": 0.5521,
9200
+ "step": 9640
9201
+ },
9202
+ {
9203
+ "epoch": 2.03,
9204
+ "learning_rate": 0.0002,
9205
+ "loss": 0.5549,
9206
+ "step": 9650
9207
+ },
9208
+ {
9209
+ "epoch": 2.03,
9210
+ "learning_rate": 0.0002,
9211
+ "loss": 0.5885,
9212
+ "step": 9660
9213
+ },
9214
+ {
9215
+ "epoch": 2.03,
9216
+ "learning_rate": 0.0002,
9217
+ "loss": 0.6067,
9218
+ "step": 9670
9219
+ },
9220
+ {
9221
+ "epoch": 2.03,
9222
+ "learning_rate": 0.0002,
9223
+ "loss": 0.6036,
9224
+ "step": 9680
9225
+ },
9226
+ {
9227
+ "epoch": 2.04,
9228
+ "learning_rate": 0.0002,
9229
+ "loss": 0.568,
9230
+ "step": 9690
9231
+ },
9232
+ {
9233
+ "epoch": 2.04,
9234
+ "learning_rate": 0.0002,
9235
+ "loss": 0.5804,
9236
+ "step": 9700
9237
+ },
9238
+ {
9239
+ "epoch": 2.04,
9240
+ "learning_rate": 0.0002,
9241
+ "loss": 0.5598,
9242
+ "step": 9710
9243
+ },
9244
+ {
9245
+ "epoch": 2.04,
9246
+ "learning_rate": 0.0002,
9247
+ "loss": 0.6057,
9248
+ "step": 9720
9249
+ },
9250
+ {
9251
+ "epoch": 2.04,
9252
+ "learning_rate": 0.0002,
9253
+ "loss": 0.5918,
9254
+ "step": 9730
9255
+ },
9256
+ {
9257
+ "epoch": 2.05,
9258
+ "learning_rate": 0.0002,
9259
+ "loss": 0.5472,
9260
+ "step": 9740
9261
+ },
9262
+ {
9263
+ "epoch": 2.05,
9264
+ "learning_rate": 0.0002,
9265
+ "loss": 0.6174,
9266
+ "step": 9750
9267
+ },
9268
+ {
9269
+ "epoch": 2.05,
9270
+ "learning_rate": 0.0002,
9271
+ "loss": 0.5732,
9272
+ "step": 9760
9273
+ },
9274
+ {
9275
+ "epoch": 2.05,
9276
+ "learning_rate": 0.0002,
9277
+ "loss": 0.5516,
9278
+ "step": 9770
9279
+ },
9280
+ {
9281
+ "epoch": 2.06,
9282
+ "learning_rate": 0.0002,
9283
+ "loss": 0.5365,
9284
+ "step": 9780
9285
+ },
9286
+ {
9287
+ "epoch": 2.06,
9288
+ "learning_rate": 0.0002,
9289
+ "loss": 0.5752,
9290
+ "step": 9790
9291
+ },
9292
+ {
9293
+ "epoch": 2.06,
9294
+ "learning_rate": 0.0002,
9295
+ "loss": 0.5389,
9296
+ "step": 9800
9297
+ },
9298
+ {
9299
+ "epoch": 2.06,
9300
+ "eval_loss": 0.7387669682502747,
9301
+ "eval_runtime": 68.5167,
9302
+ "eval_samples_per_second": 14.595,
9303
+ "eval_steps_per_second": 7.297,
9304
+ "step": 9800
9305
+ },
9306
+ {
9307
+ "epoch": 2.06,
9308
+ "mmlu_eval_accuracy": 0.4896363915300104,
9309
+ "mmlu_eval_accuracy_abstract_algebra": 0.2727272727272727,
9310
+ "mmlu_eval_accuracy_anatomy": 0.5714285714285714,
9311
+ "mmlu_eval_accuracy_astronomy": 0.5,
9312
+ "mmlu_eval_accuracy_business_ethics": 0.6363636363636364,
9313
+ "mmlu_eval_accuracy_clinical_knowledge": 0.5172413793103449,
9314
+ "mmlu_eval_accuracy_college_biology": 0.4375,
9315
+ "mmlu_eval_accuracy_college_chemistry": 0.25,
9316
+ "mmlu_eval_accuracy_college_computer_science": 0.45454545454545453,
9317
+ "mmlu_eval_accuracy_college_mathematics": 0.18181818181818182,
9318
+ "mmlu_eval_accuracy_college_medicine": 0.36363636363636365,
9319
+ "mmlu_eval_accuracy_college_physics": 0.36363636363636365,
9320
+ "mmlu_eval_accuracy_computer_security": 0.45454545454545453,
9321
+ "mmlu_eval_accuracy_conceptual_physics": 0.34615384615384615,
9322
+ "mmlu_eval_accuracy_econometrics": 0.16666666666666666,
9323
+ "mmlu_eval_accuracy_electrical_engineering": 0.3125,
9324
+ "mmlu_eval_accuracy_elementary_mathematics": 0.34146341463414637,
9325
+ "mmlu_eval_accuracy_formal_logic": 0.2857142857142857,
9326
+ "mmlu_eval_accuracy_global_facts": 0.2,
9327
+ "mmlu_eval_accuracy_high_school_biology": 0.34375,
9328
+ "mmlu_eval_accuracy_high_school_chemistry": 0.3181818181818182,
9329
+ "mmlu_eval_accuracy_high_school_computer_science": 0.5555555555555556,
9330
+ "mmlu_eval_accuracy_high_school_european_history": 0.5,
9331
+ "mmlu_eval_accuracy_high_school_geography": 0.8181818181818182,
9332
+ "mmlu_eval_accuracy_high_school_government_and_politics": 0.5238095238095238,
9333
+ "mmlu_eval_accuracy_high_school_macroeconomics": 0.46511627906976744,
9334
+ "mmlu_eval_accuracy_high_school_mathematics": 0.20689655172413793,
9335
+ "mmlu_eval_accuracy_high_school_microeconomics": 0.38461538461538464,
9336
+ "mmlu_eval_accuracy_high_school_physics": 0.23529411764705882,
9337
+ "mmlu_eval_accuracy_high_school_psychology": 0.8,
9338
+ "mmlu_eval_accuracy_high_school_statistics": 0.43478260869565216,
9339
+ "mmlu_eval_accuracy_high_school_us_history": 0.5,
9340
+ "mmlu_eval_accuracy_high_school_world_history": 0.6538461538461539,
9341
+ "mmlu_eval_accuracy_human_aging": 0.7391304347826086,
9342
+ "mmlu_eval_accuracy_human_sexuality": 0.4166666666666667,
9343
+ "mmlu_eval_accuracy_international_law": 0.8461538461538461,
9344
+ "mmlu_eval_accuracy_jurisprudence": 0.45454545454545453,
9345
+ "mmlu_eval_accuracy_logical_fallacies": 0.6666666666666666,
9346
+ "mmlu_eval_accuracy_machine_learning": 0.36363636363636365,
9347
+ "mmlu_eval_accuracy_management": 0.6363636363636364,
9348
+ "mmlu_eval_accuracy_marketing": 0.88,
9349
+ "mmlu_eval_accuracy_medical_genetics": 0.9090909090909091,
9350
+ "mmlu_eval_accuracy_miscellaneous": 0.6744186046511628,
9351
+ "mmlu_eval_accuracy_moral_disputes": 0.47368421052631576,
9352
+ "mmlu_eval_accuracy_moral_scenarios": 0.25,
9353
+ "mmlu_eval_accuracy_nutrition": 0.6363636363636364,
9354
+ "mmlu_eval_accuracy_philosophy": 0.5588235294117647,
9355
+ "mmlu_eval_accuracy_prehistory": 0.5428571428571428,
9356
+ "mmlu_eval_accuracy_professional_accounting": 0.3548387096774194,
9357
+ "mmlu_eval_accuracy_professional_law": 0.3,
9358
+ "mmlu_eval_accuracy_professional_medicine": 0.5161290322580645,
9359
+ "mmlu_eval_accuracy_professional_psychology": 0.5217391304347826,
9360
+ "mmlu_eval_accuracy_public_relations": 0.5,
9361
+ "mmlu_eval_accuracy_security_studies": 0.4444444444444444,
9362
+ "mmlu_eval_accuracy_sociology": 0.6818181818181818,
9363
+ "mmlu_eval_accuracy_us_foreign_policy": 0.9090909090909091,
9364
+ "mmlu_eval_accuracy_virology": 0.5,
9365
+ "mmlu_eval_accuracy_world_religions": 0.7368421052631579,
9366
+ "mmlu_loss": 1.1208905377291511,
9367
+ "step": 9800
9368
  }
9369
  ],
9370
  "max_steps": 10000,
9371
  "num_train_epochs": 3,
9372
+ "total_flos": 1.5644657169559388e+18,
9373
  "trial_name": null,
9374
  "trial_params": null
9375
  }
{checkpoint-7600 β†’ checkpoint-9800}/training_args.bin RENAMED
File without changes