DylanJHJ commited on
Commit
7fa7ec1
·
1 Parent(s): 6ee7f39

update models

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +20 -0
  2. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/config.json +45 -0
  3. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/model.safetensors +3 -0
  4. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/optimizer.pt +3 -0
  5. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/rng_state_0.pth +3 -0
  6. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/rng_state_1.pth +3 -0
  7. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/rng_state_2.pth +3 -0
  8. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/rng_state_3.pth +3 -0
  9. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/scheduler.pt +3 -0
  10. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/trainer_state.json +799 -0
  11. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/training_args.bin +3 -0
  12. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/config.json +45 -0
  13. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/model.safetensors +3 -0
  14. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/optimizer.pt +3 -0
  15. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/rng_state_0.pth +3 -0
  16. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/rng_state_1.pth +3 -0
  17. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/rng_state_2.pth +3 -0
  18. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/rng_state_3.pth +3 -0
  19. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/scheduler.pt +3 -0
  20. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/trainer_state.json +1564 -0
  21. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/training_args.bin +3 -0
  22. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/config.json +45 -0
  23. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/model.safetensors +3 -0
  24. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/optimizer.pt +3 -0
  25. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/rng_state_0.pth +3 -0
  26. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/rng_state_1.pth +3 -0
  27. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/rng_state_2.pth +3 -0
  28. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/rng_state_3.pth +3 -0
  29. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/scheduler.pt +3 -0
  30. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/trainer_state.json +2329 -0
  31. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/training_args.bin +3 -0
  32. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/config.json +45 -0
  33. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/model.safetensors +3 -0
  34. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/optimizer.pt +3 -0
  35. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/rng_state_0.pth +3 -0
  36. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/rng_state_1.pth +3 -0
  37. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/rng_state_2.pth +3 -0
  38. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/rng_state_3.pth +3 -0
  39. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/scheduler.pt +3 -0
  40. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/trainer_state.json +3094 -0
  41. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/training_args.bin +3 -0
  42. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-5000/removed +0 -0
  43. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/config.json +45 -0
  44. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/model.safetensors +3 -0
  45. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/optimizer.pt +3 -0
  46. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/rng_state_0.pth +3 -0
  47. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/rng_state_1.pth +3 -0
  48. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/rng_state_2.pth +3 -0
  49. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/rng_state_3.pth +3 -0
  50. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/runs/Dec10_21-48-05_nid005066/events.out.tfevents.1765396268.nid005066.34579.0 +3 -0
.gitattributes CHANGED
@@ -33,3 +33,23 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/optimizer.pt filter=lfs diff=lfs merge=lfs -text
37
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
38
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/model.safetensors filter=lfs diff=lfs merge=lfs -text
39
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/model.safetensors filter=lfs diff=lfs merge=lfs -text
40
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
41
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/model.safetensors filter=lfs diff=lfs merge=lfs -text
42
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
43
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
44
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
45
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/model.safetensors filter=lfs diff=lfs merge=lfs -text
46
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco/optimizer.pt filter=lfs diff=lfs merge=lfs -text
47
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco/checkpoint-4000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
48
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco/checkpoint-4000/model.safetensors filter=lfs diff=lfs merge=lfs -text
49
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco/model.safetensors filter=lfs diff=lfs merge=lfs -text
50
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco/checkpoint-2000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
51
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco/checkpoint-2000/model.safetensors filter=lfs diff=lfs merge=lfs -text
52
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco/checkpoint-3000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
53
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
54
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco/checkpoint-1000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
55
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco/checkpoint-1000/model.safetensors filter=lfs diff=lfs merge=lfs -text
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "gradient_checkpointing": false,
20
+ "hidden_activation": "gelu",
21
+ "hidden_size": 768,
22
+ "initializer_cutoff_factor": 2.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 1152,
25
+ "layer_norm_eps": 1e-05,
26
+ "local_attention": 128,
27
+ "local_rope_theta": 10000.0,
28
+ "max_position_embeddings": 8192,
29
+ "mlp_bias": false,
30
+ "mlp_dropout": 0.0,
31
+ "model_type": "modernbert",
32
+ "norm_bias": false,
33
+ "norm_eps": 1e-05,
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 22,
36
+ "pad_token_id": 50283,
37
+ "position_embedding_type": "absolute",
38
+ "repad_logits_with_grad": false,
39
+ "sep_token_id": 50282,
40
+ "sparse_pred_ignore_index": -100,
41
+ "sparse_prediction": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.53.0",
44
+ "vocab_size": 50368
45
+ }
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f8cea23e59027c2c5b36168d940ec4d3844b1f29c814cb8d6202d62fa45f9c0
3
+ size 298041696
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4444b588f984289c2834e5eed337f71344ab7374e6ca84c4ebf27f527fdfb31e
3
+ size 596170443
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3044a568f2c48f55afab6e63cf0c046d3586784e77e39b1af825abecfb2cdba8
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b50176e6a2e04d27f26849df53fc28c50b0357c3f1bc26ca6b6bb8848991144
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbda571d0d08add9abed7e3b3b11bb38b8d472a92fd8409d00660f44b646db10
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5ce74a6352b774e1ca5b1ca2e8b3e684a2cc29987576c968c710783e2b33305
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:053b181ac2db5ceae48310494587c492212456802136bc3852c62ccc7b9ff38c
3
+ size 1465
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/trainer_state.json ADDED
@@ -0,0 +1,799 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 0.7849293563579278,
6
+ "eval_steps": 1000.0,
7
+ "global_step": 1000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0,
14
+ "eval/nano_beir.arguana": 0.4400549254974839,
15
+ "step": 0
16
+ },
17
+ {
18
+ "epoch": 0,
19
+ "eval/nano_beir.climate_fever": 0.22880053057901228,
20
+ "step": 0
21
+ },
22
+ {
23
+ "epoch": 0,
24
+ "eval/nano_beir.dbpedia_entity": 0.30026716538531234,
25
+ "step": 0
26
+ },
27
+ {
28
+ "epoch": 0,
29
+ "eval/nano_beir.fever": 0.6419178788751609,
30
+ "step": 0
31
+ },
32
+ {
33
+ "epoch": 0,
34
+ "eval/nano_beir.fiqa": 0.4370593595651809,
35
+ "step": 0
36
+ },
37
+ {
38
+ "epoch": 0,
39
+ "eval/nano_beir.hotpotqa": 0.6723047477224562,
40
+ "step": 0
41
+ },
42
+ {
43
+ "epoch": 0,
44
+ "eval/nano_beir.nfcorpus": 0.3437446614516137,
45
+ "step": 0
46
+ },
47
+ {
48
+ "epoch": 0,
49
+ "eval/nano_beir.nq": 0.436711543163819,
50
+ "step": 0
51
+ },
52
+ {
53
+ "epoch": 0,
54
+ "eval/nano_beir.quora": 0.8439997970269911,
55
+ "step": 0
56
+ },
57
+ {
58
+ "epoch": 0,
59
+ "eval/nano_beir.scidocs": 0.26304331393922176,
60
+ "step": 0
61
+ },
62
+ {
63
+ "epoch": 0,
64
+ "eval/nano_beir.scifact": 0.7599341408912704,
65
+ "step": 0
66
+ },
67
+ {
68
+ "epoch": 0,
69
+ "eval/nano_beir.webis_touche2020": 0.29696490432942557,
70
+ "step": 0
71
+ },
72
+ {
73
+ "epoch": 0,
74
+ "eval/avg": 0.47206691403557904,
75
+ "step": 0
76
+ },
77
+ {
78
+ "epoch": 0.007849293563579277,
79
+ "grad_norm": 9.3125,
80
+ "learning_rate": 1.8e-06,
81
+ "loss": 0.8676,
82
+ "step": 10
83
+ },
84
+ {
85
+ "epoch": 0.015698587127158554,
86
+ "grad_norm": 7.46875,
87
+ "learning_rate": 3.8e-06,
88
+ "loss": 0.9307,
89
+ "step": 20
90
+ },
91
+ {
92
+ "epoch": 0.023547880690737835,
93
+ "grad_norm": 9.0,
94
+ "learning_rate": 5.8e-06,
95
+ "loss": 0.8335,
96
+ "step": 30
97
+ },
98
+ {
99
+ "epoch": 0.03139717425431711,
100
+ "grad_norm": 7.15625,
101
+ "learning_rate": 7.8e-06,
102
+ "loss": 0.8277,
103
+ "step": 40
104
+ },
105
+ {
106
+ "epoch": 0.03924646781789639,
107
+ "grad_norm": 7.65625,
108
+ "learning_rate": 9.800000000000001e-06,
109
+ "loss": 0.8458,
110
+ "step": 50
111
+ },
112
+ {
113
+ "epoch": 0.04709576138147567,
114
+ "grad_norm": 9.125,
115
+ "learning_rate": 1.18e-05,
116
+ "loss": 0.8074,
117
+ "step": 60
118
+ },
119
+ {
120
+ "epoch": 0.054945054945054944,
121
+ "grad_norm": 8.1875,
122
+ "learning_rate": 1.3800000000000002e-05,
123
+ "loss": 0.7318,
124
+ "step": 70
125
+ },
126
+ {
127
+ "epoch": 0.06279434850863422,
128
+ "grad_norm": 7.65625,
129
+ "learning_rate": 1.58e-05,
130
+ "loss": 0.8056,
131
+ "step": 80
132
+ },
133
+ {
134
+ "epoch": 0.0706436420722135,
135
+ "grad_norm": 8.625,
136
+ "learning_rate": 1.78e-05,
137
+ "loss": 0.8091,
138
+ "step": 90
139
+ },
140
+ {
141
+ "epoch": 0.07849293563579278,
142
+ "grad_norm": 7.78125,
143
+ "learning_rate": 1.9800000000000004e-05,
144
+ "loss": 0.8001,
145
+ "step": 100
146
+ },
147
+ {
148
+ "epoch": 0.08634222919937205,
149
+ "grad_norm": 8.0625,
150
+ "learning_rate": 2.18e-05,
151
+ "loss": 0.9097,
152
+ "step": 110
153
+ },
154
+ {
155
+ "epoch": 0.09419152276295134,
156
+ "grad_norm": 5.78125,
157
+ "learning_rate": 2.38e-05,
158
+ "loss": 0.7487,
159
+ "step": 120
160
+ },
161
+ {
162
+ "epoch": 0.10204081632653061,
163
+ "grad_norm": 6.5625,
164
+ "learning_rate": 2.58e-05,
165
+ "loss": 0.7602,
166
+ "step": 130
167
+ },
168
+ {
169
+ "epoch": 0.10989010989010989,
170
+ "grad_norm": 6.40625,
171
+ "learning_rate": 2.7800000000000005e-05,
172
+ "loss": 0.7309,
173
+ "step": 140
174
+ },
175
+ {
176
+ "epoch": 0.11773940345368916,
177
+ "grad_norm": 6.75,
178
+ "learning_rate": 2.98e-05,
179
+ "loss": 0.7629,
180
+ "step": 150
181
+ },
182
+ {
183
+ "epoch": 0.12558869701726844,
184
+ "grad_norm": 6.46875,
185
+ "learning_rate": 3.18e-05,
186
+ "loss": 0.8294,
187
+ "step": 160
188
+ },
189
+ {
190
+ "epoch": 0.13343799058084774,
191
+ "grad_norm": 6.53125,
192
+ "learning_rate": 3.38e-05,
193
+ "loss": 0.6725,
194
+ "step": 170
195
+ },
196
+ {
197
+ "epoch": 0.141287284144427,
198
+ "grad_norm": 6.59375,
199
+ "learning_rate": 3.58e-05,
200
+ "loss": 0.6788,
201
+ "step": 180
202
+ },
203
+ {
204
+ "epoch": 0.14913657770800628,
205
+ "grad_norm": 6.125,
206
+ "learning_rate": 3.7800000000000004e-05,
207
+ "loss": 0.7473,
208
+ "step": 190
209
+ },
210
+ {
211
+ "epoch": 0.15698587127158556,
212
+ "grad_norm": 6.5625,
213
+ "learning_rate": 3.9800000000000005e-05,
214
+ "loss": 0.7462,
215
+ "step": 200
216
+ },
217
+ {
218
+ "epoch": 0.16483516483516483,
219
+ "grad_norm": 6.53125,
220
+ "learning_rate": 4.18e-05,
221
+ "loss": 0.6637,
222
+ "step": 210
223
+ },
224
+ {
225
+ "epoch": 0.1726844583987441,
226
+ "grad_norm": 6.96875,
227
+ "learning_rate": 4.38e-05,
228
+ "loss": 0.7876,
229
+ "step": 220
230
+ },
231
+ {
232
+ "epoch": 0.18053375196232338,
233
+ "grad_norm": 8.0,
234
+ "learning_rate": 4.58e-05,
235
+ "loss": 0.7846,
236
+ "step": 230
237
+ },
238
+ {
239
+ "epoch": 0.18838304552590268,
240
+ "grad_norm": 9.875,
241
+ "learning_rate": 4.78e-05,
242
+ "loss": 0.7945,
243
+ "step": 240
244
+ },
245
+ {
246
+ "epoch": 0.19623233908948196,
247
+ "grad_norm": 7.625,
248
+ "learning_rate": 4.9800000000000004e-05,
249
+ "loss": 0.7045,
250
+ "step": 250
251
+ },
252
+ {
253
+ "epoch": 0.20408163265306123,
254
+ "grad_norm": 5.1875,
255
+ "learning_rate": 5.1800000000000005e-05,
256
+ "loss": 0.7334,
257
+ "step": 260
258
+ },
259
+ {
260
+ "epoch": 0.2119309262166405,
261
+ "grad_norm": 5.625,
262
+ "learning_rate": 5.380000000000001e-05,
263
+ "loss": 0.7552,
264
+ "step": 270
265
+ },
266
+ {
267
+ "epoch": 0.21978021978021978,
268
+ "grad_norm": 4.125,
269
+ "learning_rate": 5.580000000000001e-05,
270
+ "loss": 0.7312,
271
+ "step": 280
272
+ },
273
+ {
274
+ "epoch": 0.22762951334379905,
275
+ "grad_norm": 6.1875,
276
+ "learning_rate": 5.7799999999999995e-05,
277
+ "loss": 0.6843,
278
+ "step": 290
279
+ },
280
+ {
281
+ "epoch": 0.23547880690737832,
282
+ "grad_norm": 5.9375,
283
+ "learning_rate": 5.9800000000000003e-05,
284
+ "loss": 0.7139,
285
+ "step": 300
286
+ },
287
+ {
288
+ "epoch": 0.24332810047095763,
289
+ "grad_norm": 4.90625,
290
+ "learning_rate": 6.18e-05,
291
+ "loss": 0.6777,
292
+ "step": 310
293
+ },
294
+ {
295
+ "epoch": 0.25117739403453687,
296
+ "grad_norm": 5.25,
297
+ "learning_rate": 6.38e-05,
298
+ "loss": 0.7767,
299
+ "step": 320
300
+ },
301
+ {
302
+ "epoch": 0.25902668759811615,
303
+ "grad_norm": 5.71875,
304
+ "learning_rate": 6.58e-05,
305
+ "loss": 0.7373,
306
+ "step": 330
307
+ },
308
+ {
309
+ "epoch": 0.2668759811616955,
310
+ "grad_norm": 6.0,
311
+ "learning_rate": 6.780000000000001e-05,
312
+ "loss": 0.8043,
313
+ "step": 340
314
+ },
315
+ {
316
+ "epoch": 0.27472527472527475,
317
+ "grad_norm": 5.6875,
318
+ "learning_rate": 6.98e-05,
319
+ "loss": 0.748,
320
+ "step": 350
321
+ },
322
+ {
323
+ "epoch": 0.282574568288854,
324
+ "grad_norm": 5.21875,
325
+ "learning_rate": 7.18e-05,
326
+ "loss": 0.6968,
327
+ "step": 360
328
+ },
329
+ {
330
+ "epoch": 0.2904238618524333,
331
+ "grad_norm": 5.65625,
332
+ "learning_rate": 7.38e-05,
333
+ "loss": 0.6794,
334
+ "step": 370
335
+ },
336
+ {
337
+ "epoch": 0.29827315541601257,
338
+ "grad_norm": 5.71875,
339
+ "learning_rate": 7.58e-05,
340
+ "loss": 0.6401,
341
+ "step": 380
342
+ },
343
+ {
344
+ "epoch": 0.30612244897959184,
345
+ "grad_norm": 6.40625,
346
+ "learning_rate": 7.780000000000001e-05,
347
+ "loss": 0.8027,
348
+ "step": 390
349
+ },
350
+ {
351
+ "epoch": 0.3139717425431711,
352
+ "grad_norm": 6.9375,
353
+ "learning_rate": 7.98e-05,
354
+ "loss": 0.6472,
355
+ "step": 400
356
+ },
357
+ {
358
+ "epoch": 0.3218210361067504,
359
+ "grad_norm": 5.5625,
360
+ "learning_rate": 8.18e-05,
361
+ "loss": 0.7221,
362
+ "step": 410
363
+ },
364
+ {
365
+ "epoch": 0.32967032967032966,
366
+ "grad_norm": 5.5,
367
+ "learning_rate": 8.38e-05,
368
+ "loss": 0.668,
369
+ "step": 420
370
+ },
371
+ {
372
+ "epoch": 0.33751962323390894,
373
+ "grad_norm": 6.09375,
374
+ "learning_rate": 8.58e-05,
375
+ "loss": 0.6923,
376
+ "step": 430
377
+ },
378
+ {
379
+ "epoch": 0.3453689167974882,
380
+ "grad_norm": 7.09375,
381
+ "learning_rate": 8.78e-05,
382
+ "loss": 0.6982,
383
+ "step": 440
384
+ },
385
+ {
386
+ "epoch": 0.3532182103610675,
387
+ "grad_norm": 5.5,
388
+ "learning_rate": 8.98e-05,
389
+ "loss": 0.6671,
390
+ "step": 450
391
+ },
392
+ {
393
+ "epoch": 0.36106750392464676,
394
+ "grad_norm": 6.1875,
395
+ "learning_rate": 9.180000000000001e-05,
396
+ "loss": 0.6487,
397
+ "step": 460
398
+ },
399
+ {
400
+ "epoch": 0.36891679748822603,
401
+ "grad_norm": 6.875,
402
+ "learning_rate": 9.38e-05,
403
+ "loss": 0.7471,
404
+ "step": 470
405
+ },
406
+ {
407
+ "epoch": 0.37676609105180536,
408
+ "grad_norm": 5.71875,
409
+ "learning_rate": 9.58e-05,
410
+ "loss": 0.6445,
411
+ "step": 480
412
+ },
413
+ {
414
+ "epoch": 0.38461538461538464,
415
+ "grad_norm": 5.125,
416
+ "learning_rate": 9.78e-05,
417
+ "loss": 0.659,
418
+ "step": 490
419
+ },
420
+ {
421
+ "epoch": 0.3924646781789639,
422
+ "grad_norm": 7.96875,
423
+ "learning_rate": 9.98e-05,
424
+ "loss": 0.6469,
425
+ "step": 500
426
+ },
427
+ {
428
+ "epoch": 0.4003139717425432,
429
+ "grad_norm": 6.5625,
430
+ "learning_rate": 9.999901304280685e-05,
431
+ "loss": 0.6404,
432
+ "step": 510
433
+ },
434
+ {
435
+ "epoch": 0.40816326530612246,
436
+ "grad_norm": 7.5625,
437
+ "learning_rate": 9.999560138895238e-05,
438
+ "loss": 0.7289,
439
+ "step": 520
440
+ },
441
+ {
442
+ "epoch": 0.41601255886970173,
443
+ "grad_norm": 5.5625,
444
+ "learning_rate": 9.99897530200195e-05,
445
+ "loss": 0.6332,
446
+ "step": 530
447
+ },
448
+ {
449
+ "epoch": 0.423861852433281,
450
+ "grad_norm": 5.09375,
451
+ "learning_rate": 9.998146822104943e-05,
452
+ "loss": 0.7495,
453
+ "step": 540
454
+ },
455
+ {
456
+ "epoch": 0.4317111459968603,
457
+ "grad_norm": 4.625,
458
+ "learning_rate": 9.997074739583162e-05,
459
+ "loss": 0.6474,
460
+ "step": 550
461
+ },
462
+ {
463
+ "epoch": 0.43956043956043955,
464
+ "grad_norm": 5.4375,
465
+ "learning_rate": 9.995759106688393e-05,
466
+ "loss": 0.6788,
467
+ "step": 560
468
+ },
469
+ {
470
+ "epoch": 0.4474097331240188,
471
+ "grad_norm": 5.5,
472
+ "learning_rate": 9.99419998754273e-05,
473
+ "loss": 0.6584,
474
+ "step": 570
475
+ },
476
+ {
477
+ "epoch": 0.4552590266875981,
478
+ "grad_norm": 5.125,
479
+ "learning_rate": 9.992397458135438e-05,
480
+ "loss": 0.7894,
481
+ "step": 580
482
+ },
483
+ {
484
+ "epoch": 0.4631083202511774,
485
+ "grad_norm": 5.71875,
486
+ "learning_rate": 9.990351606319261e-05,
487
+ "loss": 0.7216,
488
+ "step": 590
489
+ },
490
+ {
491
+ "epoch": 0.47095761381475665,
492
+ "grad_norm": 5.40625,
493
+ "learning_rate": 9.988062531806126e-05,
494
+ "loss": 0.6335,
495
+ "step": 600
496
+ },
497
+ {
498
+ "epoch": 0.478806907378336,
499
+ "grad_norm": 7.0,
500
+ "learning_rate": 9.985530346162299e-05,
501
+ "loss": 0.6338,
502
+ "step": 610
503
+ },
504
+ {
505
+ "epoch": 0.48665620094191525,
506
+ "grad_norm": 5.375,
507
+ "learning_rate": 9.982755172802933e-05,
508
+ "loss": 0.6607,
509
+ "step": 620
510
+ },
511
+ {
512
+ "epoch": 0.4945054945054945,
513
+ "grad_norm": 5.59375,
514
+ "learning_rate": 9.979737146986064e-05,
515
+ "loss": 0.6775,
516
+ "step": 630
517
+ },
518
+ {
519
+ "epoch": 0.5023547880690737,
520
+ "grad_norm": 9.0,
521
+ "learning_rate": 9.976476415806013e-05,
522
+ "loss": 0.642,
523
+ "step": 640
524
+ },
525
+ {
526
+ "epoch": 0.5102040816326531,
527
+ "grad_norm": 7.46875,
528
+ "learning_rate": 9.972973138186217e-05,
529
+ "loss": 0.7126,
530
+ "step": 650
531
+ },
532
+ {
533
+ "epoch": 0.5180533751962323,
534
+ "grad_norm": 4.15625,
535
+ "learning_rate": 9.969227484871484e-05,
536
+ "loss": 0.6095,
537
+ "step": 660
538
+ },
539
+ {
540
+ "epoch": 0.5259026687598116,
541
+ "grad_norm": 5.03125,
542
+ "learning_rate": 9.965239638419672e-05,
543
+ "loss": 0.7593,
544
+ "step": 670
545
+ },
546
+ {
547
+ "epoch": 0.533751962323391,
548
+ "grad_norm": 5.1875,
549
+ "learning_rate": 9.961009793192792e-05,
550
+ "loss": 0.7108,
551
+ "step": 680
552
+ },
553
+ {
554
+ "epoch": 0.5416012558869702,
555
+ "grad_norm": 6.0,
556
+ "learning_rate": 9.956538155347534e-05,
557
+ "loss": 0.7327,
558
+ "step": 690
559
+ },
560
+ {
561
+ "epoch": 0.5494505494505495,
562
+ "grad_norm": 4.8125,
563
+ "learning_rate": 9.951824942825215e-05,
564
+ "loss": 0.677,
565
+ "step": 700
566
+ },
567
+ {
568
+ "epoch": 0.5572998430141287,
569
+ "grad_norm": 4.6875,
570
+ "learning_rate": 9.946870385341167e-05,
571
+ "loss": 0.6671,
572
+ "step": 710
573
+ },
574
+ {
575
+ "epoch": 0.565149136577708,
576
+ "grad_norm": 4.8125,
577
+ "learning_rate": 9.941674724373531e-05,
578
+ "loss": 0.604,
579
+ "step": 720
580
+ },
581
+ {
582
+ "epoch": 0.5729984301412873,
583
+ "grad_norm": 5.84375,
584
+ "learning_rate": 9.936238213151491e-05,
585
+ "loss": 0.6487,
586
+ "step": 730
587
+ },
588
+ {
589
+ "epoch": 0.5808477237048666,
590
+ "grad_norm": 5.90625,
591
+ "learning_rate": 9.930561116642935e-05,
592
+ "loss": 0.6417,
593
+ "step": 740
594
+ },
595
+ {
596
+ "epoch": 0.5886970172684458,
597
+ "grad_norm": 6.03125,
598
+ "learning_rate": 9.924643711541539e-05,
599
+ "loss": 0.8061,
600
+ "step": 750
601
+ },
602
+ {
603
+ "epoch": 0.5965463108320251,
604
+ "grad_norm": 4.90625,
605
+ "learning_rate": 9.918486286253279e-05,
606
+ "loss": 0.7204,
607
+ "step": 760
608
+ },
609
+ {
610
+ "epoch": 0.6043956043956044,
611
+ "grad_norm": 4.5,
612
+ "learning_rate": 9.912089140882376e-05,
613
+ "loss": 0.6467,
614
+ "step": 770
615
+ },
616
+ {
617
+ "epoch": 0.6122448979591837,
618
+ "grad_norm": 6.5,
619
+ "learning_rate": 9.90545258721667e-05,
620
+ "loss": 0.6307,
621
+ "step": 780
622
+ },
623
+ {
624
+ "epoch": 0.6200941915227629,
625
+ "grad_norm": 6.71875,
626
+ "learning_rate": 9.898576948712426e-05,
627
+ "loss": 0.8016,
628
+ "step": 790
629
+ },
630
+ {
631
+ "epoch": 0.6279434850863422,
632
+ "grad_norm": 8.0,
633
+ "learning_rate": 9.891462560478562e-05,
634
+ "loss": 0.7312,
635
+ "step": 800
636
+ },
637
+ {
638
+ "epoch": 0.6357927786499215,
639
+ "grad_norm": 6.03125,
640
+ "learning_rate": 9.884109769260325e-05,
641
+ "loss": 0.7053,
642
+ "step": 810
643
+ },
644
+ {
645
+ "epoch": 0.6436420722135008,
646
+ "grad_norm": 4.21875,
647
+ "learning_rate": 9.876518933422386e-05,
648
+ "loss": 0.6855,
649
+ "step": 820
650
+ },
651
+ {
652
+ "epoch": 0.6514913657770801,
653
+ "grad_norm": 5.28125,
654
+ "learning_rate": 9.868690422931372e-05,
655
+ "loss": 0.7619,
656
+ "step": 830
657
+ },
658
+ {
659
+ "epoch": 0.6593406593406593,
660
+ "grad_norm": 8.25,
661
+ "learning_rate": 9.860624619337844e-05,
662
+ "loss": 0.7078,
663
+ "step": 840
664
+ },
665
+ {
666
+ "epoch": 0.6671899529042387,
667
+ "grad_norm": 5.3125,
668
+ "learning_rate": 9.852321915757687e-05,
669
+ "loss": 0.6612,
670
+ "step": 850
671
+ },
672
+ {
673
+ "epoch": 0.6750392464678179,
674
+ "grad_norm": 5.8125,
675
+ "learning_rate": 9.843782716852963e-05,
676
+ "loss": 0.7664,
677
+ "step": 860
678
+ },
679
+ {
680
+ "epoch": 0.6828885400313972,
681
+ "grad_norm": 4.90625,
682
+ "learning_rate": 9.835007438812177e-05,
683
+ "loss": 0.7502,
684
+ "step": 870
685
+ },
686
+ {
687
+ "epoch": 0.6907378335949764,
688
+ "grad_norm": 4.84375,
689
+ "learning_rate": 9.82599650933e-05,
690
+ "loss": 0.6931,
691
+ "step": 880
692
+ },
693
+ {
694
+ "epoch": 0.6985871271585558,
695
+ "grad_norm": 4.78125,
696
+ "learning_rate": 9.816750367586425e-05,
697
+ "loss": 0.7217,
698
+ "step": 890
699
+ },
700
+ {
701
+ "epoch": 0.706436420722135,
702
+ "grad_norm": 5.4375,
703
+ "learning_rate": 9.807269464225355e-05,
704
+ "loss": 0.6607,
705
+ "step": 900
706
+ },
707
+ {
708
+ "epoch": 0.7142857142857143,
709
+ "grad_norm": 6.40625,
710
+ "learning_rate": 9.797554261332645e-05,
711
+ "loss": 0.7778,
712
+ "step": 910
713
+ },
714
+ {
715
+ "epoch": 0.7221350078492935,
716
+ "grad_norm": 6.90625,
717
+ "learning_rate": 9.787605232413574e-05,
718
+ "loss": 0.7161,
719
+ "step": 920
720
+ },
721
+ {
722
+ "epoch": 0.7299843014128728,
723
+ "grad_norm": 6.28125,
724
+ "learning_rate": 9.777422862369783e-05,
725
+ "loss": 0.6461,
726
+ "step": 930
727
+ },
728
+ {
729
+ "epoch": 0.7378335949764521,
730
+ "grad_norm": 6.125,
731
+ "learning_rate": 9.767007647475617e-05,
732
+ "loss": 0.6601,
733
+ "step": 940
734
+ },
735
+ {
736
+ "epoch": 0.7456828885400314,
737
+ "grad_norm": 5.375,
738
+ "learning_rate": 9.756360095353956e-05,
739
+ "loss": 0.6062,
740
+ "step": 950
741
+ },
742
+ {
743
+ "epoch": 0.7535321821036107,
744
+ "grad_norm": 7.03125,
745
+ "learning_rate": 9.745480724951473e-05,
746
+ "loss": 0.7328,
747
+ "step": 960
748
+ },
749
+ {
750
+ "epoch": 0.7613814756671899,
751
+ "grad_norm": 5.34375,
752
+ "learning_rate": 9.73437006651333e-05,
753
+ "loss": 0.6816,
754
+ "step": 970
755
+ },
756
+ {
757
+ "epoch": 0.7692307692307693,
758
+ "grad_norm": 4.90625,
759
+ "learning_rate": 9.723028661557345e-05,
760
+ "loss": 0.7171,
761
+ "step": 980
762
+ },
763
+ {
764
+ "epoch": 0.7770800627943485,
765
+ "grad_norm": 3.828125,
766
+ "learning_rate": 9.711457062847595e-05,
767
+ "loss": 0.6988,
768
+ "step": 990
769
+ },
770
+ {
771
+ "epoch": 0.7849293563579278,
772
+ "grad_norm": 5.09375,
773
+ "learning_rate": 9.699655834367478e-05,
774
+ "loss": 0.6331,
775
+ "step": 1000
776
+ }
777
+ ],
778
+ "logging_steps": 10,
779
+ "max_steps": 5000,
780
+ "num_input_tokens_seen": 0,
781
+ "num_train_epochs": 4,
782
+ "save_steps": 1000,
783
+ "stateful_callbacks": {
784
+ "TrainerControl": {
785
+ "args": {
786
+ "should_epoch_stop": false,
787
+ "should_evaluate": false,
788
+ "should_log": false,
789
+ "should_save": true,
790
+ "should_training_stop": false
791
+ },
792
+ "attributes": {}
793
+ }
794
+ },
795
+ "total_flos": 0.0,
796
+ "train_batch_size": 16,
797
+ "trial_name": null,
798
+ "trial_params": null
799
+ }
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-1000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2ed821ed7ef0074e4b13618254be9ea2c38d68708aace5456b6baa256883072
3
+ size 6097
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "gradient_checkpointing": false,
20
+ "hidden_activation": "gelu",
21
+ "hidden_size": 768,
22
+ "initializer_cutoff_factor": 2.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 1152,
25
+ "layer_norm_eps": 1e-05,
26
+ "local_attention": 128,
27
+ "local_rope_theta": 10000.0,
28
+ "max_position_embeddings": 8192,
29
+ "mlp_bias": false,
30
+ "mlp_dropout": 0.0,
31
+ "model_type": "modernbert",
32
+ "norm_bias": false,
33
+ "norm_eps": 1e-05,
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 22,
36
+ "pad_token_id": 50283,
37
+ "position_embedding_type": "absolute",
38
+ "repad_logits_with_grad": false,
39
+ "sep_token_id": 50282,
40
+ "sparse_pred_ignore_index": -100,
41
+ "sparse_prediction": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.53.0",
44
+ "vocab_size": 50368
45
+ }
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ce892ac499c726f9f0ffc58f21eaed98059a99787e8a5e0b47b09519dd56c03
3
+ size 298041696
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9816537f9d6d804d6866451e1583a3857ffd0af4e789dd38f7361227c01b1b62
3
+ size 596170443
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18bb6b6e7da6250d22c25b95c64287be4aea598d9d97ef67b0fd69b9be869ed7
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3cff16289615ba210b601be2162794d1949cb1d62132099a4cf62330c43649a5
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6caa3980c911491b74ec2e96e48a78cd7c3365dd32116c61642a4b6839a2da1
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e93d0d6464a92d227e9235c134b2224cf28ea45ba375cf82e6949c632e2b0d5
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff65ca8b6e225d1d82cc63887ee464e8c896624115a4dc693eae1fff20815699
3
+ size 1465
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/trainer_state.json ADDED
@@ -0,0 +1,1564 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 1.5698587127158556,
6
+ "eval_steps": 1000.0,
7
+ "global_step": 2000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0,
14
+ "eval/nano_beir.arguana": 0.4400549254974839,
15
+ "step": 0
16
+ },
17
+ {
18
+ "epoch": 0,
19
+ "eval/nano_beir.climate_fever": 0.22880053057901228,
20
+ "step": 0
21
+ },
22
+ {
23
+ "epoch": 0,
24
+ "eval/nano_beir.dbpedia_entity": 0.30026716538531234,
25
+ "step": 0
26
+ },
27
+ {
28
+ "epoch": 0,
29
+ "eval/nano_beir.fever": 0.6419178788751609,
30
+ "step": 0
31
+ },
32
+ {
33
+ "epoch": 0,
34
+ "eval/nano_beir.fiqa": 0.4370593595651809,
35
+ "step": 0
36
+ },
37
+ {
38
+ "epoch": 0,
39
+ "eval/nano_beir.hotpotqa": 0.6723047477224562,
40
+ "step": 0
41
+ },
42
+ {
43
+ "epoch": 0,
44
+ "eval/nano_beir.nfcorpus": 0.3437446614516137,
45
+ "step": 0
46
+ },
47
+ {
48
+ "epoch": 0,
49
+ "eval/nano_beir.nq": 0.436711543163819,
50
+ "step": 0
51
+ },
52
+ {
53
+ "epoch": 0,
54
+ "eval/nano_beir.quora": 0.8439997970269911,
55
+ "step": 0
56
+ },
57
+ {
58
+ "epoch": 0,
59
+ "eval/nano_beir.scidocs": 0.26304331393922176,
60
+ "step": 0
61
+ },
62
+ {
63
+ "epoch": 0,
64
+ "eval/nano_beir.scifact": 0.7599341408912704,
65
+ "step": 0
66
+ },
67
+ {
68
+ "epoch": 0,
69
+ "eval/nano_beir.webis_touche2020": 0.29696490432942557,
70
+ "step": 0
71
+ },
72
+ {
73
+ "epoch": 0,
74
+ "eval/avg": 0.47206691403557904,
75
+ "step": 0
76
+ },
77
+ {
78
+ "epoch": 0.007849293563579277,
79
+ "grad_norm": 9.3125,
80
+ "learning_rate": 1.8e-06,
81
+ "loss": 0.8676,
82
+ "step": 10
83
+ },
84
+ {
85
+ "epoch": 0.015698587127158554,
86
+ "grad_norm": 7.46875,
87
+ "learning_rate": 3.8e-06,
88
+ "loss": 0.9307,
89
+ "step": 20
90
+ },
91
+ {
92
+ "epoch": 0.023547880690737835,
93
+ "grad_norm": 9.0,
94
+ "learning_rate": 5.8e-06,
95
+ "loss": 0.8335,
96
+ "step": 30
97
+ },
98
+ {
99
+ "epoch": 0.03139717425431711,
100
+ "grad_norm": 7.15625,
101
+ "learning_rate": 7.8e-06,
102
+ "loss": 0.8277,
103
+ "step": 40
104
+ },
105
+ {
106
+ "epoch": 0.03924646781789639,
107
+ "grad_norm": 7.65625,
108
+ "learning_rate": 9.800000000000001e-06,
109
+ "loss": 0.8458,
110
+ "step": 50
111
+ },
112
+ {
113
+ "epoch": 0.04709576138147567,
114
+ "grad_norm": 9.125,
115
+ "learning_rate": 1.18e-05,
116
+ "loss": 0.8074,
117
+ "step": 60
118
+ },
119
+ {
120
+ "epoch": 0.054945054945054944,
121
+ "grad_norm": 8.1875,
122
+ "learning_rate": 1.3800000000000002e-05,
123
+ "loss": 0.7318,
124
+ "step": 70
125
+ },
126
+ {
127
+ "epoch": 0.06279434850863422,
128
+ "grad_norm": 7.65625,
129
+ "learning_rate": 1.58e-05,
130
+ "loss": 0.8056,
131
+ "step": 80
132
+ },
133
+ {
134
+ "epoch": 0.0706436420722135,
135
+ "grad_norm": 8.625,
136
+ "learning_rate": 1.78e-05,
137
+ "loss": 0.8091,
138
+ "step": 90
139
+ },
140
+ {
141
+ "epoch": 0.07849293563579278,
142
+ "grad_norm": 7.78125,
143
+ "learning_rate": 1.9800000000000004e-05,
144
+ "loss": 0.8001,
145
+ "step": 100
146
+ },
147
+ {
148
+ "epoch": 0.08634222919937205,
149
+ "grad_norm": 8.0625,
150
+ "learning_rate": 2.18e-05,
151
+ "loss": 0.9097,
152
+ "step": 110
153
+ },
154
+ {
155
+ "epoch": 0.09419152276295134,
156
+ "grad_norm": 5.78125,
157
+ "learning_rate": 2.38e-05,
158
+ "loss": 0.7487,
159
+ "step": 120
160
+ },
161
+ {
162
+ "epoch": 0.10204081632653061,
163
+ "grad_norm": 6.5625,
164
+ "learning_rate": 2.58e-05,
165
+ "loss": 0.7602,
166
+ "step": 130
167
+ },
168
+ {
169
+ "epoch": 0.10989010989010989,
170
+ "grad_norm": 6.40625,
171
+ "learning_rate": 2.7800000000000005e-05,
172
+ "loss": 0.7309,
173
+ "step": 140
174
+ },
175
+ {
176
+ "epoch": 0.11773940345368916,
177
+ "grad_norm": 6.75,
178
+ "learning_rate": 2.98e-05,
179
+ "loss": 0.7629,
180
+ "step": 150
181
+ },
182
+ {
183
+ "epoch": 0.12558869701726844,
184
+ "grad_norm": 6.46875,
185
+ "learning_rate": 3.18e-05,
186
+ "loss": 0.8294,
187
+ "step": 160
188
+ },
189
+ {
190
+ "epoch": 0.13343799058084774,
191
+ "grad_norm": 6.53125,
192
+ "learning_rate": 3.38e-05,
193
+ "loss": 0.6725,
194
+ "step": 170
195
+ },
196
+ {
197
+ "epoch": 0.141287284144427,
198
+ "grad_norm": 6.59375,
199
+ "learning_rate": 3.58e-05,
200
+ "loss": 0.6788,
201
+ "step": 180
202
+ },
203
+ {
204
+ "epoch": 0.14913657770800628,
205
+ "grad_norm": 6.125,
206
+ "learning_rate": 3.7800000000000004e-05,
207
+ "loss": 0.7473,
208
+ "step": 190
209
+ },
210
+ {
211
+ "epoch": 0.15698587127158556,
212
+ "grad_norm": 6.5625,
213
+ "learning_rate": 3.9800000000000005e-05,
214
+ "loss": 0.7462,
215
+ "step": 200
216
+ },
217
+ {
218
+ "epoch": 0.16483516483516483,
219
+ "grad_norm": 6.53125,
220
+ "learning_rate": 4.18e-05,
221
+ "loss": 0.6637,
222
+ "step": 210
223
+ },
224
+ {
225
+ "epoch": 0.1726844583987441,
226
+ "grad_norm": 6.96875,
227
+ "learning_rate": 4.38e-05,
228
+ "loss": 0.7876,
229
+ "step": 220
230
+ },
231
+ {
232
+ "epoch": 0.18053375196232338,
233
+ "grad_norm": 8.0,
234
+ "learning_rate": 4.58e-05,
235
+ "loss": 0.7846,
236
+ "step": 230
237
+ },
238
+ {
239
+ "epoch": 0.18838304552590268,
240
+ "grad_norm": 9.875,
241
+ "learning_rate": 4.78e-05,
242
+ "loss": 0.7945,
243
+ "step": 240
244
+ },
245
+ {
246
+ "epoch": 0.19623233908948196,
247
+ "grad_norm": 7.625,
248
+ "learning_rate": 4.9800000000000004e-05,
249
+ "loss": 0.7045,
250
+ "step": 250
251
+ },
252
+ {
253
+ "epoch": 0.20408163265306123,
254
+ "grad_norm": 5.1875,
255
+ "learning_rate": 5.1800000000000005e-05,
256
+ "loss": 0.7334,
257
+ "step": 260
258
+ },
259
+ {
260
+ "epoch": 0.2119309262166405,
261
+ "grad_norm": 5.625,
262
+ "learning_rate": 5.380000000000001e-05,
263
+ "loss": 0.7552,
264
+ "step": 270
265
+ },
266
+ {
267
+ "epoch": 0.21978021978021978,
268
+ "grad_norm": 4.125,
269
+ "learning_rate": 5.580000000000001e-05,
270
+ "loss": 0.7312,
271
+ "step": 280
272
+ },
273
+ {
274
+ "epoch": 0.22762951334379905,
275
+ "grad_norm": 6.1875,
276
+ "learning_rate": 5.7799999999999995e-05,
277
+ "loss": 0.6843,
278
+ "step": 290
279
+ },
280
+ {
281
+ "epoch": 0.23547880690737832,
282
+ "grad_norm": 5.9375,
283
+ "learning_rate": 5.9800000000000003e-05,
284
+ "loss": 0.7139,
285
+ "step": 300
286
+ },
287
+ {
288
+ "epoch": 0.24332810047095763,
289
+ "grad_norm": 4.90625,
290
+ "learning_rate": 6.18e-05,
291
+ "loss": 0.6777,
292
+ "step": 310
293
+ },
294
+ {
295
+ "epoch": 0.25117739403453687,
296
+ "grad_norm": 5.25,
297
+ "learning_rate": 6.38e-05,
298
+ "loss": 0.7767,
299
+ "step": 320
300
+ },
301
+ {
302
+ "epoch": 0.25902668759811615,
303
+ "grad_norm": 5.71875,
304
+ "learning_rate": 6.58e-05,
305
+ "loss": 0.7373,
306
+ "step": 330
307
+ },
308
+ {
309
+ "epoch": 0.2668759811616955,
310
+ "grad_norm": 6.0,
311
+ "learning_rate": 6.780000000000001e-05,
312
+ "loss": 0.8043,
313
+ "step": 340
314
+ },
315
+ {
316
+ "epoch": 0.27472527472527475,
317
+ "grad_norm": 5.6875,
318
+ "learning_rate": 6.98e-05,
319
+ "loss": 0.748,
320
+ "step": 350
321
+ },
322
+ {
323
+ "epoch": 0.282574568288854,
324
+ "grad_norm": 5.21875,
325
+ "learning_rate": 7.18e-05,
326
+ "loss": 0.6968,
327
+ "step": 360
328
+ },
329
+ {
330
+ "epoch": 0.2904238618524333,
331
+ "grad_norm": 5.65625,
332
+ "learning_rate": 7.38e-05,
333
+ "loss": 0.6794,
334
+ "step": 370
335
+ },
336
+ {
337
+ "epoch": 0.29827315541601257,
338
+ "grad_norm": 5.71875,
339
+ "learning_rate": 7.58e-05,
340
+ "loss": 0.6401,
341
+ "step": 380
342
+ },
343
+ {
344
+ "epoch": 0.30612244897959184,
345
+ "grad_norm": 6.40625,
346
+ "learning_rate": 7.780000000000001e-05,
347
+ "loss": 0.8027,
348
+ "step": 390
349
+ },
350
+ {
351
+ "epoch": 0.3139717425431711,
352
+ "grad_norm": 6.9375,
353
+ "learning_rate": 7.98e-05,
354
+ "loss": 0.6472,
355
+ "step": 400
356
+ },
357
+ {
358
+ "epoch": 0.3218210361067504,
359
+ "grad_norm": 5.5625,
360
+ "learning_rate": 8.18e-05,
361
+ "loss": 0.7221,
362
+ "step": 410
363
+ },
364
+ {
365
+ "epoch": 0.32967032967032966,
366
+ "grad_norm": 5.5,
367
+ "learning_rate": 8.38e-05,
368
+ "loss": 0.668,
369
+ "step": 420
370
+ },
371
+ {
372
+ "epoch": 0.33751962323390894,
373
+ "grad_norm": 6.09375,
374
+ "learning_rate": 8.58e-05,
375
+ "loss": 0.6923,
376
+ "step": 430
377
+ },
378
+ {
379
+ "epoch": 0.3453689167974882,
380
+ "grad_norm": 7.09375,
381
+ "learning_rate": 8.78e-05,
382
+ "loss": 0.6982,
383
+ "step": 440
384
+ },
385
+ {
386
+ "epoch": 0.3532182103610675,
387
+ "grad_norm": 5.5,
388
+ "learning_rate": 8.98e-05,
389
+ "loss": 0.6671,
390
+ "step": 450
391
+ },
392
+ {
393
+ "epoch": 0.36106750392464676,
394
+ "grad_norm": 6.1875,
395
+ "learning_rate": 9.180000000000001e-05,
396
+ "loss": 0.6487,
397
+ "step": 460
398
+ },
399
+ {
400
+ "epoch": 0.36891679748822603,
401
+ "grad_norm": 6.875,
402
+ "learning_rate": 9.38e-05,
403
+ "loss": 0.7471,
404
+ "step": 470
405
+ },
406
+ {
407
+ "epoch": 0.37676609105180536,
408
+ "grad_norm": 5.71875,
409
+ "learning_rate": 9.58e-05,
410
+ "loss": 0.6445,
411
+ "step": 480
412
+ },
413
+ {
414
+ "epoch": 0.38461538461538464,
415
+ "grad_norm": 5.125,
416
+ "learning_rate": 9.78e-05,
417
+ "loss": 0.659,
418
+ "step": 490
419
+ },
420
+ {
421
+ "epoch": 0.3924646781789639,
422
+ "grad_norm": 7.96875,
423
+ "learning_rate": 9.98e-05,
424
+ "loss": 0.6469,
425
+ "step": 500
426
+ },
427
+ {
428
+ "epoch": 0.4003139717425432,
429
+ "grad_norm": 6.5625,
430
+ "learning_rate": 9.999901304280685e-05,
431
+ "loss": 0.6404,
432
+ "step": 510
433
+ },
434
+ {
435
+ "epoch": 0.40816326530612246,
436
+ "grad_norm": 7.5625,
437
+ "learning_rate": 9.999560138895238e-05,
438
+ "loss": 0.7289,
439
+ "step": 520
440
+ },
441
+ {
442
+ "epoch": 0.41601255886970173,
443
+ "grad_norm": 5.5625,
444
+ "learning_rate": 9.99897530200195e-05,
445
+ "loss": 0.6332,
446
+ "step": 530
447
+ },
448
+ {
449
+ "epoch": 0.423861852433281,
450
+ "grad_norm": 5.09375,
451
+ "learning_rate": 9.998146822104943e-05,
452
+ "loss": 0.7495,
453
+ "step": 540
454
+ },
455
+ {
456
+ "epoch": 0.4317111459968603,
457
+ "grad_norm": 4.625,
458
+ "learning_rate": 9.997074739583162e-05,
459
+ "loss": 0.6474,
460
+ "step": 550
461
+ },
462
+ {
463
+ "epoch": 0.43956043956043955,
464
+ "grad_norm": 5.4375,
465
+ "learning_rate": 9.995759106688393e-05,
466
+ "loss": 0.6788,
467
+ "step": 560
468
+ },
469
+ {
470
+ "epoch": 0.4474097331240188,
471
+ "grad_norm": 5.5,
472
+ "learning_rate": 9.99419998754273e-05,
473
+ "loss": 0.6584,
474
+ "step": 570
475
+ },
476
+ {
477
+ "epoch": 0.4552590266875981,
478
+ "grad_norm": 5.125,
479
+ "learning_rate": 9.992397458135438e-05,
480
+ "loss": 0.7894,
481
+ "step": 580
482
+ },
483
+ {
484
+ "epoch": 0.4631083202511774,
485
+ "grad_norm": 5.71875,
486
+ "learning_rate": 9.990351606319261e-05,
487
+ "loss": 0.7216,
488
+ "step": 590
489
+ },
490
+ {
491
+ "epoch": 0.47095761381475665,
492
+ "grad_norm": 5.40625,
493
+ "learning_rate": 9.988062531806126e-05,
494
+ "loss": 0.6335,
495
+ "step": 600
496
+ },
497
+ {
498
+ "epoch": 0.478806907378336,
499
+ "grad_norm": 7.0,
500
+ "learning_rate": 9.985530346162299e-05,
501
+ "loss": 0.6338,
502
+ "step": 610
503
+ },
504
+ {
505
+ "epoch": 0.48665620094191525,
506
+ "grad_norm": 5.375,
507
+ "learning_rate": 9.982755172802933e-05,
508
+ "loss": 0.6607,
509
+ "step": 620
510
+ },
511
+ {
512
+ "epoch": 0.4945054945054945,
513
+ "grad_norm": 5.59375,
514
+ "learning_rate": 9.979737146986064e-05,
515
+ "loss": 0.6775,
516
+ "step": 630
517
+ },
518
+ {
519
+ "epoch": 0.5023547880690737,
520
+ "grad_norm": 9.0,
521
+ "learning_rate": 9.976476415806013e-05,
522
+ "loss": 0.642,
523
+ "step": 640
524
+ },
525
+ {
526
+ "epoch": 0.5102040816326531,
527
+ "grad_norm": 7.46875,
528
+ "learning_rate": 9.972973138186217e-05,
529
+ "loss": 0.7126,
530
+ "step": 650
531
+ },
532
+ {
533
+ "epoch": 0.5180533751962323,
534
+ "grad_norm": 4.15625,
535
+ "learning_rate": 9.969227484871484e-05,
536
+ "loss": 0.6095,
537
+ "step": 660
538
+ },
539
+ {
540
+ "epoch": 0.5259026687598116,
541
+ "grad_norm": 5.03125,
542
+ "learning_rate": 9.965239638419672e-05,
543
+ "loss": 0.7593,
544
+ "step": 670
545
+ },
546
+ {
547
+ "epoch": 0.533751962323391,
548
+ "grad_norm": 5.1875,
549
+ "learning_rate": 9.961009793192792e-05,
550
+ "loss": 0.7108,
551
+ "step": 680
552
+ },
553
+ {
554
+ "epoch": 0.5416012558869702,
555
+ "grad_norm": 6.0,
556
+ "learning_rate": 9.956538155347534e-05,
557
+ "loss": 0.7327,
558
+ "step": 690
559
+ },
560
+ {
561
+ "epoch": 0.5494505494505495,
562
+ "grad_norm": 4.8125,
563
+ "learning_rate": 9.951824942825215e-05,
564
+ "loss": 0.677,
565
+ "step": 700
566
+ },
567
+ {
568
+ "epoch": 0.5572998430141287,
569
+ "grad_norm": 4.6875,
570
+ "learning_rate": 9.946870385341167e-05,
571
+ "loss": 0.6671,
572
+ "step": 710
573
+ },
574
+ {
575
+ "epoch": 0.565149136577708,
576
+ "grad_norm": 4.8125,
577
+ "learning_rate": 9.941674724373531e-05,
578
+ "loss": 0.604,
579
+ "step": 720
580
+ },
581
+ {
582
+ "epoch": 0.5729984301412873,
583
+ "grad_norm": 5.84375,
584
+ "learning_rate": 9.936238213151491e-05,
585
+ "loss": 0.6487,
586
+ "step": 730
587
+ },
588
+ {
589
+ "epoch": 0.5808477237048666,
590
+ "grad_norm": 5.90625,
591
+ "learning_rate": 9.930561116642935e-05,
592
+ "loss": 0.6417,
593
+ "step": 740
594
+ },
595
+ {
596
+ "epoch": 0.5886970172684458,
597
+ "grad_norm": 6.03125,
598
+ "learning_rate": 9.924643711541539e-05,
599
+ "loss": 0.8061,
600
+ "step": 750
601
+ },
602
+ {
603
+ "epoch": 0.5965463108320251,
604
+ "grad_norm": 4.90625,
605
+ "learning_rate": 9.918486286253279e-05,
606
+ "loss": 0.7204,
607
+ "step": 760
608
+ },
609
+ {
610
+ "epoch": 0.6043956043956044,
611
+ "grad_norm": 4.5,
612
+ "learning_rate": 9.912089140882376e-05,
613
+ "loss": 0.6467,
614
+ "step": 770
615
+ },
616
+ {
617
+ "epoch": 0.6122448979591837,
618
+ "grad_norm": 6.5,
619
+ "learning_rate": 9.90545258721667e-05,
620
+ "loss": 0.6307,
621
+ "step": 780
622
+ },
623
+ {
624
+ "epoch": 0.6200941915227629,
625
+ "grad_norm": 6.71875,
626
+ "learning_rate": 9.898576948712426e-05,
627
+ "loss": 0.8016,
628
+ "step": 790
629
+ },
630
+ {
631
+ "epoch": 0.6279434850863422,
632
+ "grad_norm": 8.0,
633
+ "learning_rate": 9.891462560478562e-05,
634
+ "loss": 0.7312,
635
+ "step": 800
636
+ },
637
+ {
638
+ "epoch": 0.6357927786499215,
639
+ "grad_norm": 6.03125,
640
+ "learning_rate": 9.884109769260325e-05,
641
+ "loss": 0.7053,
642
+ "step": 810
643
+ },
644
+ {
645
+ "epoch": 0.6436420722135008,
646
+ "grad_norm": 4.21875,
647
+ "learning_rate": 9.876518933422386e-05,
648
+ "loss": 0.6855,
649
+ "step": 820
650
+ },
651
+ {
652
+ "epoch": 0.6514913657770801,
653
+ "grad_norm": 5.28125,
654
+ "learning_rate": 9.868690422931372e-05,
655
+ "loss": 0.7619,
656
+ "step": 830
657
+ },
658
+ {
659
+ "epoch": 0.6593406593406593,
660
+ "grad_norm": 8.25,
661
+ "learning_rate": 9.860624619337844e-05,
662
+ "loss": 0.7078,
663
+ "step": 840
664
+ },
665
+ {
666
+ "epoch": 0.6671899529042387,
667
+ "grad_norm": 5.3125,
668
+ "learning_rate": 9.852321915757687e-05,
669
+ "loss": 0.6612,
670
+ "step": 850
671
+ },
672
+ {
673
+ "epoch": 0.6750392464678179,
674
+ "grad_norm": 5.8125,
675
+ "learning_rate": 9.843782716852963e-05,
676
+ "loss": 0.7664,
677
+ "step": 860
678
+ },
679
+ {
680
+ "epoch": 0.6828885400313972,
681
+ "grad_norm": 4.90625,
682
+ "learning_rate": 9.835007438812177e-05,
683
+ "loss": 0.7502,
684
+ "step": 870
685
+ },
686
+ {
687
+ "epoch": 0.6907378335949764,
688
+ "grad_norm": 4.84375,
689
+ "learning_rate": 9.82599650933e-05,
690
+ "loss": 0.6931,
691
+ "step": 880
692
+ },
693
+ {
694
+ "epoch": 0.6985871271585558,
695
+ "grad_norm": 4.78125,
696
+ "learning_rate": 9.816750367586425e-05,
697
+ "loss": 0.7217,
698
+ "step": 890
699
+ },
700
+ {
701
+ "epoch": 0.706436420722135,
702
+ "grad_norm": 5.4375,
703
+ "learning_rate": 9.807269464225355e-05,
704
+ "loss": 0.6607,
705
+ "step": 900
706
+ },
707
+ {
708
+ "epoch": 0.7142857142857143,
709
+ "grad_norm": 6.40625,
710
+ "learning_rate": 9.797554261332645e-05,
711
+ "loss": 0.7778,
712
+ "step": 910
713
+ },
714
+ {
715
+ "epoch": 0.7221350078492935,
716
+ "grad_norm": 6.90625,
717
+ "learning_rate": 9.787605232413574e-05,
718
+ "loss": 0.7161,
719
+ "step": 920
720
+ },
721
+ {
722
+ "epoch": 0.7299843014128728,
723
+ "grad_norm": 6.28125,
724
+ "learning_rate": 9.777422862369783e-05,
725
+ "loss": 0.6461,
726
+ "step": 930
727
+ },
728
+ {
729
+ "epoch": 0.7378335949764521,
730
+ "grad_norm": 6.125,
731
+ "learning_rate": 9.767007647475617e-05,
732
+ "loss": 0.6601,
733
+ "step": 940
734
+ },
735
+ {
736
+ "epoch": 0.7456828885400314,
737
+ "grad_norm": 5.375,
738
+ "learning_rate": 9.756360095353956e-05,
739
+ "loss": 0.6062,
740
+ "step": 950
741
+ },
742
+ {
743
+ "epoch": 0.7535321821036107,
744
+ "grad_norm": 7.03125,
745
+ "learning_rate": 9.745480724951473e-05,
746
+ "loss": 0.7328,
747
+ "step": 960
748
+ },
749
+ {
750
+ "epoch": 0.7613814756671899,
751
+ "grad_norm": 5.34375,
752
+ "learning_rate": 9.73437006651333e-05,
753
+ "loss": 0.6816,
754
+ "step": 970
755
+ },
756
+ {
757
+ "epoch": 0.7692307692307693,
758
+ "grad_norm": 4.90625,
759
+ "learning_rate": 9.723028661557345e-05,
760
+ "loss": 0.7171,
761
+ "step": 980
762
+ },
763
+ {
764
+ "epoch": 0.7770800627943485,
765
+ "grad_norm": 3.828125,
766
+ "learning_rate": 9.711457062847595e-05,
767
+ "loss": 0.6988,
768
+ "step": 990
769
+ },
770
+ {
771
+ "epoch": 0.7849293563579278,
772
+ "grad_norm": 5.09375,
773
+ "learning_rate": 9.699655834367478e-05,
774
+ "loss": 0.6331,
775
+ "step": 1000
776
+ },
777
+ {
778
+ "epoch": 0.7849293563579278,
779
+ "eval/nano_beir.arguana": 0.4803257504428309,
780
+ "step": 1000
781
+ },
782
+ {
783
+ "epoch": 0.7849293563579278,
784
+ "eval/nano_beir.climate_fever": 0.25852105142356685,
785
+ "step": 1000
786
+ },
787
+ {
788
+ "epoch": 0.7849293563579278,
789
+ "eval/nano_beir.dbpedia_entity": 0.33382012080541146,
790
+ "step": 1000
791
+ },
792
+ {
793
+ "epoch": 0.7849293563579278,
794
+ "eval/nano_beir.fever": 0.6414665409400568,
795
+ "step": 1000
796
+ },
797
+ {
798
+ "epoch": 0.7849293563579278,
799
+ "eval/nano_beir.fiqa": 0.4530048021812019,
800
+ "step": 1000
801
+ },
802
+ {
803
+ "epoch": 0.7849293563579278,
804
+ "eval/nano_beir.hotpotqa": 0.6827182548147874,
805
+ "step": 1000
806
+ },
807
+ {
808
+ "epoch": 0.7849293563579278,
809
+ "eval/nano_beir.nfcorpus": 0.3587781171330431,
810
+ "step": 1000
811
+ },
812
+ {
813
+ "epoch": 0.7849293563579278,
814
+ "eval/nano_beir.nq": 0.426613822215205,
815
+ "step": 1000
816
+ },
817
+ {
818
+ "epoch": 0.7849293563579278,
819
+ "eval/nano_beir.quora": 0.8451303146637551,
820
+ "step": 1000
821
+ },
822
+ {
823
+ "epoch": 0.7849293563579278,
824
+ "eval/nano_beir.scidocs": 0.24604303501015984,
825
+ "step": 1000
826
+ },
827
+ {
828
+ "epoch": 0.7849293563579278,
829
+ "eval/nano_beir.scifact": 0.7460449406252556,
830
+ "step": 1000
831
+ },
832
+ {
833
+ "epoch": 0.7849293563579278,
834
+ "eval/nano_beir.webis_touche2020": 0.2860074344347365,
835
+ "step": 1000
836
+ },
837
+ {
838
+ "epoch": 0.7849293563579278,
839
+ "eval/avg": 0.47987284872416747,
840
+ "step": 1000
841
+ },
842
+ {
843
+ "epoch": 0.792778649921507,
844
+ "grad_norm": 5.71875,
845
+ "learning_rate": 9.68762555129222e-05,
846
+ "loss": 0.6397,
847
+ "step": 1010
848
+ },
849
+ {
850
+ "epoch": 0.8006279434850864,
851
+ "grad_norm": 5.5625,
852
+ "learning_rate": 9.675366799960841e-05,
853
+ "loss": 0.8047,
854
+ "step": 1020
855
+ },
856
+ {
857
+ "epoch": 0.8084772370486656,
858
+ "grad_norm": 6.0625,
859
+ "learning_rate": 9.662880177847594e-05,
860
+ "loss": 0.6725,
861
+ "step": 1030
862
+ },
863
+ {
864
+ "epoch": 0.8163265306122449,
865
+ "grad_norm": 5.0,
866
+ "learning_rate": 9.65016629353282e-05,
867
+ "loss": 0.6879,
868
+ "step": 1040
869
+ },
870
+ {
871
+ "epoch": 0.8241758241758241,
872
+ "grad_norm": 4.90625,
873
+ "learning_rate": 9.637225766673307e-05,
874
+ "loss": 0.7164,
875
+ "step": 1050
876
+ },
877
+ {
878
+ "epoch": 0.8320251177394035,
879
+ "grad_norm": 5.40625,
880
+ "learning_rate": 9.624059227972075e-05,
881
+ "loss": 0.7116,
882
+ "step": 1060
883
+ },
884
+ {
885
+ "epoch": 0.8398744113029827,
886
+ "grad_norm": 5.96875,
887
+ "learning_rate": 9.610667319147647e-05,
888
+ "loss": 0.7343,
889
+ "step": 1070
890
+ },
891
+ {
892
+ "epoch": 0.847723704866562,
893
+ "grad_norm": 5.6875,
894
+ "learning_rate": 9.597050692902765e-05,
895
+ "loss": 0.6432,
896
+ "step": 1080
897
+ },
898
+ {
899
+ "epoch": 0.8555729984301413,
900
+ "grad_norm": 4.90625,
901
+ "learning_rate": 9.583210012892582e-05,
902
+ "loss": 0.6632,
903
+ "step": 1090
904
+ },
905
+ {
906
+ "epoch": 0.8634222919937206,
907
+ "grad_norm": 6.40625,
908
+ "learning_rate": 9.569145953692315e-05,
909
+ "loss": 0.7413,
910
+ "step": 1100
911
+ },
912
+ {
913
+ "epoch": 0.8712715855572999,
914
+ "grad_norm": 4.40625,
915
+ "learning_rate": 9.55485920076437e-05,
916
+ "loss": 0.6581,
917
+ "step": 1110
918
+ },
919
+ {
920
+ "epoch": 0.8791208791208791,
921
+ "grad_norm": 5.625,
922
+ "learning_rate": 9.540350450424926e-05,
923
+ "loss": 0.7424,
924
+ "step": 1120
925
+ },
926
+ {
927
+ "epoch": 0.8869701726844584,
928
+ "grad_norm": 5.4375,
929
+ "learning_rate": 9.525620409810007e-05,
930
+ "loss": 0.5684,
931
+ "step": 1130
932
+ },
933
+ {
934
+ "epoch": 0.8948194662480377,
935
+ "grad_norm": 6.15625,
936
+ "learning_rate": 9.510669796841014e-05,
937
+ "loss": 0.6588,
938
+ "step": 1140
939
+ },
940
+ {
941
+ "epoch": 0.902668759811617,
942
+ "grad_norm": 5.34375,
943
+ "learning_rate": 9.495499340189728e-05,
944
+ "loss": 0.6428,
945
+ "step": 1150
946
+ },
947
+ {
948
+ "epoch": 0.9105180533751962,
949
+ "grad_norm": 4.84375,
950
+ "learning_rate": 9.480109779242805e-05,
951
+ "loss": 0.6236,
952
+ "step": 1160
953
+ },
954
+ {
955
+ "epoch": 0.9183673469387755,
956
+ "grad_norm": 5.5,
957
+ "learning_rate": 9.464501864065735e-05,
958
+ "loss": 0.7007,
959
+ "step": 1170
960
+ },
961
+ {
962
+ "epoch": 0.9262166405023547,
963
+ "grad_norm": 4.75,
964
+ "learning_rate": 9.448676355366282e-05,
965
+ "loss": 0.732,
966
+ "step": 1180
967
+ },
968
+ {
969
+ "epoch": 0.9340659340659341,
970
+ "grad_norm": 4.0,
971
+ "learning_rate": 9.432634024457414e-05,
972
+ "loss": 0.759,
973
+ "step": 1190
974
+ },
975
+ {
976
+ "epoch": 0.9419152276295133,
977
+ "grad_norm": 6.4375,
978
+ "learning_rate": 9.416375653219709e-05,
979
+ "loss": 0.6467,
980
+ "step": 1200
981
+ },
982
+ {
983
+ "epoch": 0.9497645211930926,
984
+ "grad_norm": 4.53125,
985
+ "learning_rate": 9.399902034063242e-05,
986
+ "loss": 0.6971,
987
+ "step": 1210
988
+ },
989
+ {
990
+ "epoch": 0.957613814756672,
991
+ "grad_norm": 4.5,
992
+ "learning_rate": 9.383213969888972e-05,
993
+ "loss": 0.6034,
994
+ "step": 1220
995
+ },
996
+ {
997
+ "epoch": 0.9654631083202512,
998
+ "grad_norm": 4.90625,
999
+ "learning_rate": 9.366312274049602e-05,
1000
+ "loss": 0.7138,
1001
+ "step": 1230
1002
+ },
1003
+ {
1004
+ "epoch": 0.9733124018838305,
1005
+ "grad_norm": 4.59375,
1006
+ "learning_rate": 9.34919777030994e-05,
1007
+ "loss": 0.7296,
1008
+ "step": 1240
1009
+ },
1010
+ {
1011
+ "epoch": 0.9811616954474097,
1012
+ "grad_norm": 4.6875,
1013
+ "learning_rate": 9.331871292806759e-05,
1014
+ "loss": 0.6068,
1015
+ "step": 1250
1016
+ },
1017
+ {
1018
+ "epoch": 0.989010989010989,
1019
+ "grad_norm": 6.09375,
1020
+ "learning_rate": 9.314333686008125e-05,
1021
+ "loss": 0.6281,
1022
+ "step": 1260
1023
+ },
1024
+ {
1025
+ "epoch": 0.9968602825745683,
1026
+ "grad_norm": 5.71875,
1027
+ "learning_rate": 9.296585804672252e-05,
1028
+ "loss": 0.5406,
1029
+ "step": 1270
1030
+ },
1031
+ {
1032
+ "epoch": 1.0047095761381475,
1033
+ "grad_norm": 5.3125,
1034
+ "learning_rate": 9.278628513805838e-05,
1035
+ "loss": 0.6783,
1036
+ "step": 1280
1037
+ },
1038
+ {
1039
+ "epoch": 1.012558869701727,
1040
+ "grad_norm": 5.15625,
1041
+ "learning_rate": 9.260462688621905e-05,
1042
+ "loss": 0.7021,
1043
+ "step": 1290
1044
+ },
1045
+ {
1046
+ "epoch": 1.0204081632653061,
1047
+ "grad_norm": 4.84375,
1048
+ "learning_rate": 9.242089214497145e-05,
1049
+ "loss": 0.6157,
1050
+ "step": 1300
1051
+ },
1052
+ {
1053
+ "epoch": 1.0282574568288854,
1054
+ "grad_norm": 4.8125,
1055
+ "learning_rate": 9.223508986928766e-05,
1056
+ "loss": 0.5672,
1057
+ "step": 1310
1058
+ },
1059
+ {
1060
+ "epoch": 1.0361067503924646,
1061
+ "grad_norm": 4.96875,
1062
+ "learning_rate": 9.204722911490846e-05,
1063
+ "loss": 0.711,
1064
+ "step": 1320
1065
+ },
1066
+ {
1067
+ "epoch": 1.043956043956044,
1068
+ "grad_norm": 5.84375,
1069
+ "learning_rate": 9.1857319037902e-05,
1070
+ "loss": 0.5718,
1071
+ "step": 1330
1072
+ },
1073
+ {
1074
+ "epoch": 1.0518053375196232,
1075
+ "grad_norm": 6.03125,
1076
+ "learning_rate": 9.16653688942175e-05,
1077
+ "loss": 0.6359,
1078
+ "step": 1340
1079
+ },
1080
+ {
1081
+ "epoch": 1.0596546310832025,
1082
+ "grad_norm": 6.125,
1083
+ "learning_rate": 9.147138803923416e-05,
1084
+ "loss": 0.6593,
1085
+ "step": 1350
1086
+ },
1087
+ {
1088
+ "epoch": 1.0675039246467817,
1089
+ "grad_norm": 7.78125,
1090
+ "learning_rate": 9.127538592730519e-05,
1091
+ "loss": 0.7177,
1092
+ "step": 1360
1093
+ },
1094
+ {
1095
+ "epoch": 1.0753532182103611,
1096
+ "grad_norm": 4.59375,
1097
+ "learning_rate": 9.107737211129701e-05,
1098
+ "loss": 0.6882,
1099
+ "step": 1370
1100
+ },
1101
+ {
1102
+ "epoch": 1.0832025117739403,
1103
+ "grad_norm": 5.59375,
1104
+ "learning_rate": 9.087735624212365e-05,
1105
+ "loss": 0.5998,
1106
+ "step": 1380
1107
+ },
1108
+ {
1109
+ "epoch": 1.0910518053375196,
1110
+ "grad_norm": 5.21875,
1111
+ "learning_rate": 9.06753480682764e-05,
1112
+ "loss": 0.7154,
1113
+ "step": 1390
1114
+ },
1115
+ {
1116
+ "epoch": 1.098901098901099,
1117
+ "grad_norm": 6.28125,
1118
+ "learning_rate": 9.047135743534868e-05,
1119
+ "loss": 0.6295,
1120
+ "step": 1400
1121
+ },
1122
+ {
1123
+ "epoch": 1.1067503924646782,
1124
+ "grad_norm": 6.25,
1125
+ "learning_rate": 9.02653942855561e-05,
1126
+ "loss": 0.6978,
1127
+ "step": 1410
1128
+ },
1129
+ {
1130
+ "epoch": 1.1145996860282574,
1131
+ "grad_norm": 4.6875,
1132
+ "learning_rate": 9.005746865725207e-05,
1133
+ "loss": 0.6089,
1134
+ "step": 1420
1135
+ },
1136
+ {
1137
+ "epoch": 1.1224489795918366,
1138
+ "grad_norm": 7.78125,
1139
+ "learning_rate": 8.984759068443831e-05,
1140
+ "loss": 0.5945,
1141
+ "step": 1430
1142
+ },
1143
+ {
1144
+ "epoch": 1.130298273155416,
1145
+ "grad_norm": 7.34375,
1146
+ "learning_rate": 8.963577059627118e-05,
1147
+ "loss": 0.6959,
1148
+ "step": 1440
1149
+ },
1150
+ {
1151
+ "epoch": 1.1381475667189953,
1152
+ "grad_norm": 5.3125,
1153
+ "learning_rate": 8.942201871656291e-05,
1154
+ "loss": 0.6495,
1155
+ "step": 1450
1156
+ },
1157
+ {
1158
+ "epoch": 1.1459968602825745,
1159
+ "grad_norm": 5.84375,
1160
+ "learning_rate": 8.920634546327858e-05,
1161
+ "loss": 0.6512,
1162
+ "step": 1460
1163
+ },
1164
+ {
1165
+ "epoch": 1.1538461538461537,
1166
+ "grad_norm": 4.6875,
1167
+ "learning_rate": 8.898876134802826e-05,
1168
+ "loss": 0.6301,
1169
+ "step": 1470
1170
+ },
1171
+ {
1172
+ "epoch": 1.1616954474097332,
1173
+ "grad_norm": 7.0625,
1174
+ "learning_rate": 8.87692769755548e-05,
1175
+ "loss": 0.5665,
1176
+ "step": 1480
1177
+ },
1178
+ {
1179
+ "epoch": 1.1695447409733124,
1180
+ "grad_norm": 5.71875,
1181
+ "learning_rate": 8.854790304321681e-05,
1182
+ "loss": 0.6361,
1183
+ "step": 1490
1184
+ },
1185
+ {
1186
+ "epoch": 1.1773940345368916,
1187
+ "grad_norm": 5.75,
1188
+ "learning_rate": 8.832465034046749e-05,
1189
+ "loss": 0.6687,
1190
+ "step": 1500
1191
+ },
1192
+ {
1193
+ "epoch": 1.185243328100471,
1194
+ "grad_norm": 6.3125,
1195
+ "learning_rate": 8.80995297483286e-05,
1196
+ "loss": 0.6106,
1197
+ "step": 1510
1198
+ },
1199
+ {
1200
+ "epoch": 1.1930926216640503,
1201
+ "grad_norm": 6.53125,
1202
+ "learning_rate": 8.787255223886019e-05,
1203
+ "loss": 0.6526,
1204
+ "step": 1520
1205
+ },
1206
+ {
1207
+ "epoch": 1.2009419152276295,
1208
+ "grad_norm": 5.6875,
1209
+ "learning_rate": 8.764372887462586e-05,
1210
+ "loss": 0.6845,
1211
+ "step": 1530
1212
+ },
1213
+ {
1214
+ "epoch": 1.2087912087912087,
1215
+ "grad_norm": 5.78125,
1216
+ "learning_rate": 8.741307080815357e-05,
1217
+ "loss": 0.6695,
1218
+ "step": 1540
1219
+ },
1220
+ {
1221
+ "epoch": 1.2166405023547882,
1222
+ "grad_norm": 4.15625,
1223
+ "learning_rate": 8.718058928139205e-05,
1224
+ "loss": 0.6021,
1225
+ "step": 1550
1226
+ },
1227
+ {
1228
+ "epoch": 1.2244897959183674,
1229
+ "grad_norm": 5.5625,
1230
+ "learning_rate": 8.694629562516294e-05,
1231
+ "loss": 0.6173,
1232
+ "step": 1560
1233
+ },
1234
+ {
1235
+ "epoch": 1.2323390894819466,
1236
+ "grad_norm": 4.125,
1237
+ "learning_rate": 8.671020125860851e-05,
1238
+ "loss": 0.5407,
1239
+ "step": 1570
1240
+ },
1241
+ {
1242
+ "epoch": 1.2401883830455258,
1243
+ "grad_norm": 6.28125,
1244
+ "learning_rate": 8.647231768863512e-05,
1245
+ "loss": 0.6177,
1246
+ "step": 1580
1247
+ },
1248
+ {
1249
+ "epoch": 1.2480376766091053,
1250
+ "grad_norm": 5.03125,
1251
+ "learning_rate": 8.623265650935234e-05,
1252
+ "loss": 0.6132,
1253
+ "step": 1590
1254
+ },
1255
+ {
1256
+ "epoch": 1.2558869701726845,
1257
+ "grad_norm": 5.9375,
1258
+ "learning_rate": 8.599122940150795e-05,
1259
+ "loss": 0.5913,
1260
+ "step": 1600
1261
+ },
1262
+ {
1263
+ "epoch": 1.2637362637362637,
1264
+ "grad_norm": 6.28125,
1265
+ "learning_rate": 8.574804813191858e-05,
1266
+ "loss": 0.7089,
1267
+ "step": 1610
1268
+ },
1269
+ {
1270
+ "epoch": 1.2715855572998431,
1271
+ "grad_norm": 7.3125,
1272
+ "learning_rate": 8.550312455289625e-05,
1273
+ "loss": 0.5426,
1274
+ "step": 1620
1275
+ },
1276
+ {
1277
+ "epoch": 1.2794348508634223,
1278
+ "grad_norm": 4.59375,
1279
+ "learning_rate": 8.525647060167063e-05,
1280
+ "loss": 0.5678,
1281
+ "step": 1630
1282
+ },
1283
+ {
1284
+ "epoch": 1.2872841444270016,
1285
+ "grad_norm": 6.34375,
1286
+ "learning_rate": 8.500809829980734e-05,
1287
+ "loss": 0.6343,
1288
+ "step": 1640
1289
+ },
1290
+ {
1291
+ "epoch": 1.2951334379905808,
1292
+ "grad_norm": 6.5625,
1293
+ "learning_rate": 8.4758019752622e-05,
1294
+ "loss": 0.7523,
1295
+ "step": 1650
1296
+ },
1297
+ {
1298
+ "epoch": 1.30298273155416,
1299
+ "grad_norm": 4.5,
1300
+ "learning_rate": 8.450624714859016e-05,
1301
+ "loss": 0.699,
1302
+ "step": 1660
1303
+ },
1304
+ {
1305
+ "epoch": 1.3108320251177394,
1306
+ "grad_norm": 5.09375,
1307
+ "learning_rate": 8.425279275875336e-05,
1308
+ "loss": 0.6517,
1309
+ "step": 1670
1310
+ },
1311
+ {
1312
+ "epoch": 1.3186813186813187,
1313
+ "grad_norm": 4.21875,
1314
+ "learning_rate": 8.399766893612096e-05,
1315
+ "loss": 0.563,
1316
+ "step": 1680
1317
+ },
1318
+ {
1319
+ "epoch": 1.3265306122448979,
1320
+ "grad_norm": 5.5625,
1321
+ "learning_rate": 8.374088811506819e-05,
1322
+ "loss": 0.6381,
1323
+ "step": 1690
1324
+ },
1325
+ {
1326
+ "epoch": 1.3343799058084773,
1327
+ "grad_norm": 4.6875,
1328
+ "learning_rate": 8.348246281072998e-05,
1329
+ "loss": 0.6794,
1330
+ "step": 1700
1331
+ },
1332
+ {
1333
+ "epoch": 1.3422291993720565,
1334
+ "grad_norm": 5.1875,
1335
+ "learning_rate": 8.322240561839109e-05,
1336
+ "loss": 0.6494,
1337
+ "step": 1710
1338
+ },
1339
+ {
1340
+ "epoch": 1.3500784929356358,
1341
+ "grad_norm": 5.09375,
1342
+ "learning_rate": 8.296072921287217e-05,
1343
+ "loss": 0.587,
1344
+ "step": 1720
1345
+ },
1346
+ {
1347
+ "epoch": 1.3579277864992152,
1348
+ "grad_norm": 6.34375,
1349
+ "learning_rate": 8.269744634791208e-05,
1350
+ "loss": 0.6888,
1351
+ "step": 1730
1352
+ },
1353
+ {
1354
+ "epoch": 1.3657770800627944,
1355
+ "grad_norm": 4.40625,
1356
+ "learning_rate": 8.243256985554621e-05,
1357
+ "loss": 0.6629,
1358
+ "step": 1740
1359
+ },
1360
+ {
1361
+ "epoch": 1.3736263736263736,
1362
+ "grad_norm": 5.4375,
1363
+ "learning_rate": 8.21661126454811e-05,
1364
+ "loss": 0.6633,
1365
+ "step": 1750
1366
+ },
1367
+ {
1368
+ "epoch": 1.3814756671899528,
1369
+ "grad_norm": 5.03125,
1370
+ "learning_rate": 8.189808770446528e-05,
1371
+ "loss": 0.633,
1372
+ "step": 1760
1373
+ },
1374
+ {
1375
+ "epoch": 1.389324960753532,
1376
+ "grad_norm": 4.59375,
1377
+ "learning_rate": 8.162850809565623e-05,
1378
+ "loss": 0.6725,
1379
+ "step": 1770
1380
+ },
1381
+ {
1382
+ "epoch": 1.3971742543171115,
1383
+ "grad_norm": 5.71875,
1384
+ "learning_rate": 8.135738695798376e-05,
1385
+ "loss": 0.6739,
1386
+ "step": 1780
1387
+ },
1388
+ {
1389
+ "epoch": 1.4050235478806907,
1390
+ "grad_norm": 5.34375,
1391
+ "learning_rate": 8.108473750550965e-05,
1392
+ "loss": 0.6547,
1393
+ "step": 1790
1394
+ },
1395
+ {
1396
+ "epoch": 1.41287284144427,
1397
+ "grad_norm": 6.0,
1398
+ "learning_rate": 8.081057302678352e-05,
1399
+ "loss": 0.5921,
1400
+ "step": 1800
1401
+ },
1402
+ {
1403
+ "epoch": 1.4207221350078494,
1404
+ "grad_norm": 6.5625,
1405
+ "learning_rate": 8.053490688419533e-05,
1406
+ "loss": 0.6639,
1407
+ "step": 1810
1408
+ },
1409
+ {
1410
+ "epoch": 1.4285714285714286,
1411
+ "grad_norm": 6.90625,
1412
+ "learning_rate": 8.02577525133239e-05,
1413
+ "loss": 0.6445,
1414
+ "step": 1820
1415
+ },
1416
+ {
1417
+ "epoch": 1.4364207221350078,
1418
+ "grad_norm": 6.375,
1419
+ "learning_rate": 7.997912342228232e-05,
1420
+ "loss": 0.6453,
1421
+ "step": 1830
1422
+ },
1423
+ {
1424
+ "epoch": 1.4442700156985873,
1425
+ "grad_norm": 6.34375,
1426
+ "learning_rate": 7.969903319105935e-05,
1427
+ "loss": 0.6649,
1428
+ "step": 1840
1429
+ },
1430
+ {
1431
+ "epoch": 1.4521193092621665,
1432
+ "grad_norm": 5.875,
1433
+ "learning_rate": 7.941749547085777e-05,
1434
+ "loss": 0.7093,
1435
+ "step": 1850
1436
+ },
1437
+ {
1438
+ "epoch": 1.4599686028257457,
1439
+ "grad_norm": 5.71875,
1440
+ "learning_rate": 7.913452398342881e-05,
1441
+ "loss": 0.5864,
1442
+ "step": 1860
1443
+ },
1444
+ {
1445
+ "epoch": 1.467817896389325,
1446
+ "grad_norm": 4.875,
1447
+ "learning_rate": 7.885013252040359e-05,
1448
+ "loss": 0.6377,
1449
+ "step": 1870
1450
+ },
1451
+ {
1452
+ "epoch": 1.4756671899529041,
1453
+ "grad_norm": 3.625,
1454
+ "learning_rate": 7.856433494262077e-05,
1455
+ "loss": 0.6205,
1456
+ "step": 1880
1457
+ },
1458
+ {
1459
+ "epoch": 1.4835164835164836,
1460
+ "grad_norm": 5.15625,
1461
+ "learning_rate": 7.827714517945115e-05,
1462
+ "loss": 0.6009,
1463
+ "step": 1890
1464
+ },
1465
+ {
1466
+ "epoch": 1.4913657770800628,
1467
+ "grad_norm": 3.890625,
1468
+ "learning_rate": 7.798857722811856e-05,
1469
+ "loss": 0.5702,
1470
+ "step": 1900
1471
+ },
1472
+ {
1473
+ "epoch": 1.499215070643642,
1474
+ "grad_norm": 5.03125,
1475
+ "learning_rate": 7.769864515301787e-05,
1476
+ "loss": 0.5561,
1477
+ "step": 1910
1478
+ },
1479
+ {
1480
+ "epoch": 1.5070643642072215,
1481
+ "grad_norm": 6.0625,
1482
+ "learning_rate": 7.740736308502938e-05,
1483
+ "loss": 0.6446,
1484
+ "step": 1920
1485
+ },
1486
+ {
1487
+ "epoch": 1.5149136577708007,
1488
+ "grad_norm": 6.34375,
1489
+ "learning_rate": 7.711474522083015e-05,
1490
+ "loss": 0.5983,
1491
+ "step": 1930
1492
+ },
1493
+ {
1494
+ "epoch": 1.5227629513343799,
1495
+ "grad_norm": 6.1875,
1496
+ "learning_rate": 7.682080582220206e-05,
1497
+ "loss": 0.5358,
1498
+ "step": 1940
1499
+ },
1500
+ {
1501
+ "epoch": 1.5306122448979593,
1502
+ "grad_norm": 6.5,
1503
+ "learning_rate": 7.65255592153367e-05,
1504
+ "loss": 0.7394,
1505
+ "step": 1950
1506
+ },
1507
+ {
1508
+ "epoch": 1.5384615384615383,
1509
+ "grad_norm": 4.8125,
1510
+ "learning_rate": 7.622901979013717e-05,
1511
+ "loss": 0.5742,
1512
+ "step": 1960
1513
+ },
1514
+ {
1515
+ "epoch": 1.5463108320251178,
1516
+ "grad_norm": 4.5,
1517
+ "learning_rate": 7.593120199951671e-05,
1518
+ "loss": 0.6425,
1519
+ "step": 1970
1520
+ },
1521
+ {
1522
+ "epoch": 1.554160125588697,
1523
+ "grad_norm": 6.03125,
1524
+ "learning_rate": 7.563212035869425e-05,
1525
+ "loss": 0.6593,
1526
+ "step": 1980
1527
+ },
1528
+ {
1529
+ "epoch": 1.5620094191522762,
1530
+ "grad_norm": 7.03125,
1531
+ "learning_rate": 7.533178944448705e-05,
1532
+ "loss": 0.5948,
1533
+ "step": 1990
1534
+ },
1535
+ {
1536
+ "epoch": 1.5698587127158556,
1537
+ "grad_norm": 6.0,
1538
+ "learning_rate": 7.503022389460014e-05,
1539
+ "loss": 0.625,
1540
+ "step": 2000
1541
+ }
1542
+ ],
1543
+ "logging_steps": 10,
1544
+ "max_steps": 5000,
1545
+ "num_input_tokens_seen": 0,
1546
+ "num_train_epochs": 4,
1547
+ "save_steps": 1000,
1548
+ "stateful_callbacks": {
1549
+ "TrainerControl": {
1550
+ "args": {
1551
+ "should_epoch_stop": false,
1552
+ "should_evaluate": false,
1553
+ "should_log": false,
1554
+ "should_save": true,
1555
+ "should_training_stop": false
1556
+ },
1557
+ "attributes": {}
1558
+ }
1559
+ },
1560
+ "total_flos": 0.0,
1561
+ "train_batch_size": 16,
1562
+ "trial_name": null,
1563
+ "trial_params": null
1564
+ }
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-2000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2ed821ed7ef0074e4b13618254be9ea2c38d68708aace5456b6baa256883072
3
+ size 6097
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "gradient_checkpointing": false,
20
+ "hidden_activation": "gelu",
21
+ "hidden_size": 768,
22
+ "initializer_cutoff_factor": 2.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 1152,
25
+ "layer_norm_eps": 1e-05,
26
+ "local_attention": 128,
27
+ "local_rope_theta": 10000.0,
28
+ "max_position_embeddings": 8192,
29
+ "mlp_bias": false,
30
+ "mlp_dropout": 0.0,
31
+ "model_type": "modernbert",
32
+ "norm_bias": false,
33
+ "norm_eps": 1e-05,
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 22,
36
+ "pad_token_id": 50283,
37
+ "position_embedding_type": "absolute",
38
+ "repad_logits_with_grad": false,
39
+ "sep_token_id": 50282,
40
+ "sparse_pred_ignore_index": -100,
41
+ "sparse_prediction": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.53.0",
44
+ "vocab_size": 50368
45
+ }
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d41fde859da34b9501b484af69475f50a37de92a28b6fa47a87c5caff3703ae
3
+ size 298041696
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ba44b14f4a98062eaecb16a59796513bafcff78abd007a3fba0d5e20e889ea8
3
+ size 596170443
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac0e770c61e3acef7bd144efee65a450ed487b244ed293a8dd801394ffcae775
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21354c9980d6ba769d430eb9962ffba457eb04a66cb90e33b0934d3157fae7cf
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d574a8567b1636a399633d2b20273fdb3be4b888e37f0d331cd695b973f10463
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cca2b80df90b78b3ba9d9d2951c5b7aa80fc0a44b4512b495de6de40396fcb6e
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01a8f419d963b829163d14296852a07ad341e5e63e49d44de7c0a10c9369d217
3
+ size 1465
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/trainer_state.json ADDED
@@ -0,0 +1,2329 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 2.3547880690737832,
6
+ "eval_steps": 1000.0,
7
+ "global_step": 3000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0,
14
+ "eval/nano_beir.arguana": 0.4400549254974839,
15
+ "step": 0
16
+ },
17
+ {
18
+ "epoch": 0,
19
+ "eval/nano_beir.climate_fever": 0.22880053057901228,
20
+ "step": 0
21
+ },
22
+ {
23
+ "epoch": 0,
24
+ "eval/nano_beir.dbpedia_entity": 0.30026716538531234,
25
+ "step": 0
26
+ },
27
+ {
28
+ "epoch": 0,
29
+ "eval/nano_beir.fever": 0.6419178788751609,
30
+ "step": 0
31
+ },
32
+ {
33
+ "epoch": 0,
34
+ "eval/nano_beir.fiqa": 0.4370593595651809,
35
+ "step": 0
36
+ },
37
+ {
38
+ "epoch": 0,
39
+ "eval/nano_beir.hotpotqa": 0.6723047477224562,
40
+ "step": 0
41
+ },
42
+ {
43
+ "epoch": 0,
44
+ "eval/nano_beir.nfcorpus": 0.3437446614516137,
45
+ "step": 0
46
+ },
47
+ {
48
+ "epoch": 0,
49
+ "eval/nano_beir.nq": 0.436711543163819,
50
+ "step": 0
51
+ },
52
+ {
53
+ "epoch": 0,
54
+ "eval/nano_beir.quora": 0.8439997970269911,
55
+ "step": 0
56
+ },
57
+ {
58
+ "epoch": 0,
59
+ "eval/nano_beir.scidocs": 0.26304331393922176,
60
+ "step": 0
61
+ },
62
+ {
63
+ "epoch": 0,
64
+ "eval/nano_beir.scifact": 0.7599341408912704,
65
+ "step": 0
66
+ },
67
+ {
68
+ "epoch": 0,
69
+ "eval/nano_beir.webis_touche2020": 0.29696490432942557,
70
+ "step": 0
71
+ },
72
+ {
73
+ "epoch": 0,
74
+ "eval/avg": 0.47206691403557904,
75
+ "step": 0
76
+ },
77
+ {
78
+ "epoch": 0.007849293563579277,
79
+ "grad_norm": 9.3125,
80
+ "learning_rate": 1.8e-06,
81
+ "loss": 0.8676,
82
+ "step": 10
83
+ },
84
+ {
85
+ "epoch": 0.015698587127158554,
86
+ "grad_norm": 7.46875,
87
+ "learning_rate": 3.8e-06,
88
+ "loss": 0.9307,
89
+ "step": 20
90
+ },
91
+ {
92
+ "epoch": 0.023547880690737835,
93
+ "grad_norm": 9.0,
94
+ "learning_rate": 5.8e-06,
95
+ "loss": 0.8335,
96
+ "step": 30
97
+ },
98
+ {
99
+ "epoch": 0.03139717425431711,
100
+ "grad_norm": 7.15625,
101
+ "learning_rate": 7.8e-06,
102
+ "loss": 0.8277,
103
+ "step": 40
104
+ },
105
+ {
106
+ "epoch": 0.03924646781789639,
107
+ "grad_norm": 7.65625,
108
+ "learning_rate": 9.800000000000001e-06,
109
+ "loss": 0.8458,
110
+ "step": 50
111
+ },
112
+ {
113
+ "epoch": 0.04709576138147567,
114
+ "grad_norm": 9.125,
115
+ "learning_rate": 1.18e-05,
116
+ "loss": 0.8074,
117
+ "step": 60
118
+ },
119
+ {
120
+ "epoch": 0.054945054945054944,
121
+ "grad_norm": 8.1875,
122
+ "learning_rate": 1.3800000000000002e-05,
123
+ "loss": 0.7318,
124
+ "step": 70
125
+ },
126
+ {
127
+ "epoch": 0.06279434850863422,
128
+ "grad_norm": 7.65625,
129
+ "learning_rate": 1.58e-05,
130
+ "loss": 0.8056,
131
+ "step": 80
132
+ },
133
+ {
134
+ "epoch": 0.0706436420722135,
135
+ "grad_norm": 8.625,
136
+ "learning_rate": 1.78e-05,
137
+ "loss": 0.8091,
138
+ "step": 90
139
+ },
140
+ {
141
+ "epoch": 0.07849293563579278,
142
+ "grad_norm": 7.78125,
143
+ "learning_rate": 1.9800000000000004e-05,
144
+ "loss": 0.8001,
145
+ "step": 100
146
+ },
147
+ {
148
+ "epoch": 0.08634222919937205,
149
+ "grad_norm": 8.0625,
150
+ "learning_rate": 2.18e-05,
151
+ "loss": 0.9097,
152
+ "step": 110
153
+ },
154
+ {
155
+ "epoch": 0.09419152276295134,
156
+ "grad_norm": 5.78125,
157
+ "learning_rate": 2.38e-05,
158
+ "loss": 0.7487,
159
+ "step": 120
160
+ },
161
+ {
162
+ "epoch": 0.10204081632653061,
163
+ "grad_norm": 6.5625,
164
+ "learning_rate": 2.58e-05,
165
+ "loss": 0.7602,
166
+ "step": 130
167
+ },
168
+ {
169
+ "epoch": 0.10989010989010989,
170
+ "grad_norm": 6.40625,
171
+ "learning_rate": 2.7800000000000005e-05,
172
+ "loss": 0.7309,
173
+ "step": 140
174
+ },
175
+ {
176
+ "epoch": 0.11773940345368916,
177
+ "grad_norm": 6.75,
178
+ "learning_rate": 2.98e-05,
179
+ "loss": 0.7629,
180
+ "step": 150
181
+ },
182
+ {
183
+ "epoch": 0.12558869701726844,
184
+ "grad_norm": 6.46875,
185
+ "learning_rate": 3.18e-05,
186
+ "loss": 0.8294,
187
+ "step": 160
188
+ },
189
+ {
190
+ "epoch": 0.13343799058084774,
191
+ "grad_norm": 6.53125,
192
+ "learning_rate": 3.38e-05,
193
+ "loss": 0.6725,
194
+ "step": 170
195
+ },
196
+ {
197
+ "epoch": 0.141287284144427,
198
+ "grad_norm": 6.59375,
199
+ "learning_rate": 3.58e-05,
200
+ "loss": 0.6788,
201
+ "step": 180
202
+ },
203
+ {
204
+ "epoch": 0.14913657770800628,
205
+ "grad_norm": 6.125,
206
+ "learning_rate": 3.7800000000000004e-05,
207
+ "loss": 0.7473,
208
+ "step": 190
209
+ },
210
+ {
211
+ "epoch": 0.15698587127158556,
212
+ "grad_norm": 6.5625,
213
+ "learning_rate": 3.9800000000000005e-05,
214
+ "loss": 0.7462,
215
+ "step": 200
216
+ },
217
+ {
218
+ "epoch": 0.16483516483516483,
219
+ "grad_norm": 6.53125,
220
+ "learning_rate": 4.18e-05,
221
+ "loss": 0.6637,
222
+ "step": 210
223
+ },
224
+ {
225
+ "epoch": 0.1726844583987441,
226
+ "grad_norm": 6.96875,
227
+ "learning_rate": 4.38e-05,
228
+ "loss": 0.7876,
229
+ "step": 220
230
+ },
231
+ {
232
+ "epoch": 0.18053375196232338,
233
+ "grad_norm": 8.0,
234
+ "learning_rate": 4.58e-05,
235
+ "loss": 0.7846,
236
+ "step": 230
237
+ },
238
+ {
239
+ "epoch": 0.18838304552590268,
240
+ "grad_norm": 9.875,
241
+ "learning_rate": 4.78e-05,
242
+ "loss": 0.7945,
243
+ "step": 240
244
+ },
245
+ {
246
+ "epoch": 0.19623233908948196,
247
+ "grad_norm": 7.625,
248
+ "learning_rate": 4.9800000000000004e-05,
249
+ "loss": 0.7045,
250
+ "step": 250
251
+ },
252
+ {
253
+ "epoch": 0.20408163265306123,
254
+ "grad_norm": 5.1875,
255
+ "learning_rate": 5.1800000000000005e-05,
256
+ "loss": 0.7334,
257
+ "step": 260
258
+ },
259
+ {
260
+ "epoch": 0.2119309262166405,
261
+ "grad_norm": 5.625,
262
+ "learning_rate": 5.380000000000001e-05,
263
+ "loss": 0.7552,
264
+ "step": 270
265
+ },
266
+ {
267
+ "epoch": 0.21978021978021978,
268
+ "grad_norm": 4.125,
269
+ "learning_rate": 5.580000000000001e-05,
270
+ "loss": 0.7312,
271
+ "step": 280
272
+ },
273
+ {
274
+ "epoch": 0.22762951334379905,
275
+ "grad_norm": 6.1875,
276
+ "learning_rate": 5.7799999999999995e-05,
277
+ "loss": 0.6843,
278
+ "step": 290
279
+ },
280
+ {
281
+ "epoch": 0.23547880690737832,
282
+ "grad_norm": 5.9375,
283
+ "learning_rate": 5.9800000000000003e-05,
284
+ "loss": 0.7139,
285
+ "step": 300
286
+ },
287
+ {
288
+ "epoch": 0.24332810047095763,
289
+ "grad_norm": 4.90625,
290
+ "learning_rate": 6.18e-05,
291
+ "loss": 0.6777,
292
+ "step": 310
293
+ },
294
+ {
295
+ "epoch": 0.25117739403453687,
296
+ "grad_norm": 5.25,
297
+ "learning_rate": 6.38e-05,
298
+ "loss": 0.7767,
299
+ "step": 320
300
+ },
301
+ {
302
+ "epoch": 0.25902668759811615,
303
+ "grad_norm": 5.71875,
304
+ "learning_rate": 6.58e-05,
305
+ "loss": 0.7373,
306
+ "step": 330
307
+ },
308
+ {
309
+ "epoch": 0.2668759811616955,
310
+ "grad_norm": 6.0,
311
+ "learning_rate": 6.780000000000001e-05,
312
+ "loss": 0.8043,
313
+ "step": 340
314
+ },
315
+ {
316
+ "epoch": 0.27472527472527475,
317
+ "grad_norm": 5.6875,
318
+ "learning_rate": 6.98e-05,
319
+ "loss": 0.748,
320
+ "step": 350
321
+ },
322
+ {
323
+ "epoch": 0.282574568288854,
324
+ "grad_norm": 5.21875,
325
+ "learning_rate": 7.18e-05,
326
+ "loss": 0.6968,
327
+ "step": 360
328
+ },
329
+ {
330
+ "epoch": 0.2904238618524333,
331
+ "grad_norm": 5.65625,
332
+ "learning_rate": 7.38e-05,
333
+ "loss": 0.6794,
334
+ "step": 370
335
+ },
336
+ {
337
+ "epoch": 0.29827315541601257,
338
+ "grad_norm": 5.71875,
339
+ "learning_rate": 7.58e-05,
340
+ "loss": 0.6401,
341
+ "step": 380
342
+ },
343
+ {
344
+ "epoch": 0.30612244897959184,
345
+ "grad_norm": 6.40625,
346
+ "learning_rate": 7.780000000000001e-05,
347
+ "loss": 0.8027,
348
+ "step": 390
349
+ },
350
+ {
351
+ "epoch": 0.3139717425431711,
352
+ "grad_norm": 6.9375,
353
+ "learning_rate": 7.98e-05,
354
+ "loss": 0.6472,
355
+ "step": 400
356
+ },
357
+ {
358
+ "epoch": 0.3218210361067504,
359
+ "grad_norm": 5.5625,
360
+ "learning_rate": 8.18e-05,
361
+ "loss": 0.7221,
362
+ "step": 410
363
+ },
364
+ {
365
+ "epoch": 0.32967032967032966,
366
+ "grad_norm": 5.5,
367
+ "learning_rate": 8.38e-05,
368
+ "loss": 0.668,
369
+ "step": 420
370
+ },
371
+ {
372
+ "epoch": 0.33751962323390894,
373
+ "grad_norm": 6.09375,
374
+ "learning_rate": 8.58e-05,
375
+ "loss": 0.6923,
376
+ "step": 430
377
+ },
378
+ {
379
+ "epoch": 0.3453689167974882,
380
+ "grad_norm": 7.09375,
381
+ "learning_rate": 8.78e-05,
382
+ "loss": 0.6982,
383
+ "step": 440
384
+ },
385
+ {
386
+ "epoch": 0.3532182103610675,
387
+ "grad_norm": 5.5,
388
+ "learning_rate": 8.98e-05,
389
+ "loss": 0.6671,
390
+ "step": 450
391
+ },
392
+ {
393
+ "epoch": 0.36106750392464676,
394
+ "grad_norm": 6.1875,
395
+ "learning_rate": 9.180000000000001e-05,
396
+ "loss": 0.6487,
397
+ "step": 460
398
+ },
399
+ {
400
+ "epoch": 0.36891679748822603,
401
+ "grad_norm": 6.875,
402
+ "learning_rate": 9.38e-05,
403
+ "loss": 0.7471,
404
+ "step": 470
405
+ },
406
+ {
407
+ "epoch": 0.37676609105180536,
408
+ "grad_norm": 5.71875,
409
+ "learning_rate": 9.58e-05,
410
+ "loss": 0.6445,
411
+ "step": 480
412
+ },
413
+ {
414
+ "epoch": 0.38461538461538464,
415
+ "grad_norm": 5.125,
416
+ "learning_rate": 9.78e-05,
417
+ "loss": 0.659,
418
+ "step": 490
419
+ },
420
+ {
421
+ "epoch": 0.3924646781789639,
422
+ "grad_norm": 7.96875,
423
+ "learning_rate": 9.98e-05,
424
+ "loss": 0.6469,
425
+ "step": 500
426
+ },
427
+ {
428
+ "epoch": 0.4003139717425432,
429
+ "grad_norm": 6.5625,
430
+ "learning_rate": 9.999901304280685e-05,
431
+ "loss": 0.6404,
432
+ "step": 510
433
+ },
434
+ {
435
+ "epoch": 0.40816326530612246,
436
+ "grad_norm": 7.5625,
437
+ "learning_rate": 9.999560138895238e-05,
438
+ "loss": 0.7289,
439
+ "step": 520
440
+ },
441
+ {
442
+ "epoch": 0.41601255886970173,
443
+ "grad_norm": 5.5625,
444
+ "learning_rate": 9.99897530200195e-05,
445
+ "loss": 0.6332,
446
+ "step": 530
447
+ },
448
+ {
449
+ "epoch": 0.423861852433281,
450
+ "grad_norm": 5.09375,
451
+ "learning_rate": 9.998146822104943e-05,
452
+ "loss": 0.7495,
453
+ "step": 540
454
+ },
455
+ {
456
+ "epoch": 0.4317111459968603,
457
+ "grad_norm": 4.625,
458
+ "learning_rate": 9.997074739583162e-05,
459
+ "loss": 0.6474,
460
+ "step": 550
461
+ },
462
+ {
463
+ "epoch": 0.43956043956043955,
464
+ "grad_norm": 5.4375,
465
+ "learning_rate": 9.995759106688393e-05,
466
+ "loss": 0.6788,
467
+ "step": 560
468
+ },
469
+ {
470
+ "epoch": 0.4474097331240188,
471
+ "grad_norm": 5.5,
472
+ "learning_rate": 9.99419998754273e-05,
473
+ "loss": 0.6584,
474
+ "step": 570
475
+ },
476
+ {
477
+ "epoch": 0.4552590266875981,
478
+ "grad_norm": 5.125,
479
+ "learning_rate": 9.992397458135438e-05,
480
+ "loss": 0.7894,
481
+ "step": 580
482
+ },
483
+ {
484
+ "epoch": 0.4631083202511774,
485
+ "grad_norm": 5.71875,
486
+ "learning_rate": 9.990351606319261e-05,
487
+ "loss": 0.7216,
488
+ "step": 590
489
+ },
490
+ {
491
+ "epoch": 0.47095761381475665,
492
+ "grad_norm": 5.40625,
493
+ "learning_rate": 9.988062531806126e-05,
494
+ "loss": 0.6335,
495
+ "step": 600
496
+ },
497
+ {
498
+ "epoch": 0.478806907378336,
499
+ "grad_norm": 7.0,
500
+ "learning_rate": 9.985530346162299e-05,
501
+ "loss": 0.6338,
502
+ "step": 610
503
+ },
504
+ {
505
+ "epoch": 0.48665620094191525,
506
+ "grad_norm": 5.375,
507
+ "learning_rate": 9.982755172802933e-05,
508
+ "loss": 0.6607,
509
+ "step": 620
510
+ },
511
+ {
512
+ "epoch": 0.4945054945054945,
513
+ "grad_norm": 5.59375,
514
+ "learning_rate": 9.979737146986064e-05,
515
+ "loss": 0.6775,
516
+ "step": 630
517
+ },
518
+ {
519
+ "epoch": 0.5023547880690737,
520
+ "grad_norm": 9.0,
521
+ "learning_rate": 9.976476415806013e-05,
522
+ "loss": 0.642,
523
+ "step": 640
524
+ },
525
+ {
526
+ "epoch": 0.5102040816326531,
527
+ "grad_norm": 7.46875,
528
+ "learning_rate": 9.972973138186217e-05,
529
+ "loss": 0.7126,
530
+ "step": 650
531
+ },
532
+ {
533
+ "epoch": 0.5180533751962323,
534
+ "grad_norm": 4.15625,
535
+ "learning_rate": 9.969227484871484e-05,
536
+ "loss": 0.6095,
537
+ "step": 660
538
+ },
539
+ {
540
+ "epoch": 0.5259026687598116,
541
+ "grad_norm": 5.03125,
542
+ "learning_rate": 9.965239638419672e-05,
543
+ "loss": 0.7593,
544
+ "step": 670
545
+ },
546
+ {
547
+ "epoch": 0.533751962323391,
548
+ "grad_norm": 5.1875,
549
+ "learning_rate": 9.961009793192792e-05,
550
+ "loss": 0.7108,
551
+ "step": 680
552
+ },
553
+ {
554
+ "epoch": 0.5416012558869702,
555
+ "grad_norm": 6.0,
556
+ "learning_rate": 9.956538155347534e-05,
557
+ "loss": 0.7327,
558
+ "step": 690
559
+ },
560
+ {
561
+ "epoch": 0.5494505494505495,
562
+ "grad_norm": 4.8125,
563
+ "learning_rate": 9.951824942825215e-05,
564
+ "loss": 0.677,
565
+ "step": 700
566
+ },
567
+ {
568
+ "epoch": 0.5572998430141287,
569
+ "grad_norm": 4.6875,
570
+ "learning_rate": 9.946870385341167e-05,
571
+ "loss": 0.6671,
572
+ "step": 710
573
+ },
574
+ {
575
+ "epoch": 0.565149136577708,
576
+ "grad_norm": 4.8125,
577
+ "learning_rate": 9.941674724373531e-05,
578
+ "loss": 0.604,
579
+ "step": 720
580
+ },
581
+ {
582
+ "epoch": 0.5729984301412873,
583
+ "grad_norm": 5.84375,
584
+ "learning_rate": 9.936238213151491e-05,
585
+ "loss": 0.6487,
586
+ "step": 730
587
+ },
588
+ {
589
+ "epoch": 0.5808477237048666,
590
+ "grad_norm": 5.90625,
591
+ "learning_rate": 9.930561116642935e-05,
592
+ "loss": 0.6417,
593
+ "step": 740
594
+ },
595
+ {
596
+ "epoch": 0.5886970172684458,
597
+ "grad_norm": 6.03125,
598
+ "learning_rate": 9.924643711541539e-05,
599
+ "loss": 0.8061,
600
+ "step": 750
601
+ },
602
+ {
603
+ "epoch": 0.5965463108320251,
604
+ "grad_norm": 4.90625,
605
+ "learning_rate": 9.918486286253279e-05,
606
+ "loss": 0.7204,
607
+ "step": 760
608
+ },
609
+ {
610
+ "epoch": 0.6043956043956044,
611
+ "grad_norm": 4.5,
612
+ "learning_rate": 9.912089140882376e-05,
613
+ "loss": 0.6467,
614
+ "step": 770
615
+ },
616
+ {
617
+ "epoch": 0.6122448979591837,
618
+ "grad_norm": 6.5,
619
+ "learning_rate": 9.90545258721667e-05,
620
+ "loss": 0.6307,
621
+ "step": 780
622
+ },
623
+ {
624
+ "epoch": 0.6200941915227629,
625
+ "grad_norm": 6.71875,
626
+ "learning_rate": 9.898576948712426e-05,
627
+ "loss": 0.8016,
628
+ "step": 790
629
+ },
630
+ {
631
+ "epoch": 0.6279434850863422,
632
+ "grad_norm": 8.0,
633
+ "learning_rate": 9.891462560478562e-05,
634
+ "loss": 0.7312,
635
+ "step": 800
636
+ },
637
+ {
638
+ "epoch": 0.6357927786499215,
639
+ "grad_norm": 6.03125,
640
+ "learning_rate": 9.884109769260325e-05,
641
+ "loss": 0.7053,
642
+ "step": 810
643
+ },
644
+ {
645
+ "epoch": 0.6436420722135008,
646
+ "grad_norm": 4.21875,
647
+ "learning_rate": 9.876518933422386e-05,
648
+ "loss": 0.6855,
649
+ "step": 820
650
+ },
651
+ {
652
+ "epoch": 0.6514913657770801,
653
+ "grad_norm": 5.28125,
654
+ "learning_rate": 9.868690422931372e-05,
655
+ "loss": 0.7619,
656
+ "step": 830
657
+ },
658
+ {
659
+ "epoch": 0.6593406593406593,
660
+ "grad_norm": 8.25,
661
+ "learning_rate": 9.860624619337844e-05,
662
+ "loss": 0.7078,
663
+ "step": 840
664
+ },
665
+ {
666
+ "epoch": 0.6671899529042387,
667
+ "grad_norm": 5.3125,
668
+ "learning_rate": 9.852321915757687e-05,
669
+ "loss": 0.6612,
670
+ "step": 850
671
+ },
672
+ {
673
+ "epoch": 0.6750392464678179,
674
+ "grad_norm": 5.8125,
675
+ "learning_rate": 9.843782716852963e-05,
676
+ "loss": 0.7664,
677
+ "step": 860
678
+ },
679
+ {
680
+ "epoch": 0.6828885400313972,
681
+ "grad_norm": 4.90625,
682
+ "learning_rate": 9.835007438812177e-05,
683
+ "loss": 0.7502,
684
+ "step": 870
685
+ },
686
+ {
687
+ "epoch": 0.6907378335949764,
688
+ "grad_norm": 4.84375,
689
+ "learning_rate": 9.82599650933e-05,
690
+ "loss": 0.6931,
691
+ "step": 880
692
+ },
693
+ {
694
+ "epoch": 0.6985871271585558,
695
+ "grad_norm": 4.78125,
696
+ "learning_rate": 9.816750367586425e-05,
697
+ "loss": 0.7217,
698
+ "step": 890
699
+ },
700
+ {
701
+ "epoch": 0.706436420722135,
702
+ "grad_norm": 5.4375,
703
+ "learning_rate": 9.807269464225355e-05,
704
+ "loss": 0.6607,
705
+ "step": 900
706
+ },
707
+ {
708
+ "epoch": 0.7142857142857143,
709
+ "grad_norm": 6.40625,
710
+ "learning_rate": 9.797554261332645e-05,
711
+ "loss": 0.7778,
712
+ "step": 910
713
+ },
714
+ {
715
+ "epoch": 0.7221350078492935,
716
+ "grad_norm": 6.90625,
717
+ "learning_rate": 9.787605232413574e-05,
718
+ "loss": 0.7161,
719
+ "step": 920
720
+ },
721
+ {
722
+ "epoch": 0.7299843014128728,
723
+ "grad_norm": 6.28125,
724
+ "learning_rate": 9.777422862369783e-05,
725
+ "loss": 0.6461,
726
+ "step": 930
727
+ },
728
+ {
729
+ "epoch": 0.7378335949764521,
730
+ "grad_norm": 6.125,
731
+ "learning_rate": 9.767007647475617e-05,
732
+ "loss": 0.6601,
733
+ "step": 940
734
+ },
735
+ {
736
+ "epoch": 0.7456828885400314,
737
+ "grad_norm": 5.375,
738
+ "learning_rate": 9.756360095353956e-05,
739
+ "loss": 0.6062,
740
+ "step": 950
741
+ },
742
+ {
743
+ "epoch": 0.7535321821036107,
744
+ "grad_norm": 7.03125,
745
+ "learning_rate": 9.745480724951473e-05,
746
+ "loss": 0.7328,
747
+ "step": 960
748
+ },
749
+ {
750
+ "epoch": 0.7613814756671899,
751
+ "grad_norm": 5.34375,
752
+ "learning_rate": 9.73437006651333e-05,
753
+ "loss": 0.6816,
754
+ "step": 970
755
+ },
756
+ {
757
+ "epoch": 0.7692307692307693,
758
+ "grad_norm": 4.90625,
759
+ "learning_rate": 9.723028661557345e-05,
760
+ "loss": 0.7171,
761
+ "step": 980
762
+ },
763
+ {
764
+ "epoch": 0.7770800627943485,
765
+ "grad_norm": 3.828125,
766
+ "learning_rate": 9.711457062847595e-05,
767
+ "loss": 0.6988,
768
+ "step": 990
769
+ },
770
+ {
771
+ "epoch": 0.7849293563579278,
772
+ "grad_norm": 5.09375,
773
+ "learning_rate": 9.699655834367478e-05,
774
+ "loss": 0.6331,
775
+ "step": 1000
776
+ },
777
+ {
778
+ "epoch": 0.7849293563579278,
779
+ "eval/nano_beir.arguana": 0.4803257504428309,
780
+ "step": 1000
781
+ },
782
+ {
783
+ "epoch": 0.7849293563579278,
784
+ "eval/nano_beir.climate_fever": 0.25852105142356685,
785
+ "step": 1000
786
+ },
787
+ {
788
+ "epoch": 0.7849293563579278,
789
+ "eval/nano_beir.dbpedia_entity": 0.33382012080541146,
790
+ "step": 1000
791
+ },
792
+ {
793
+ "epoch": 0.7849293563579278,
794
+ "eval/nano_beir.fever": 0.6414665409400568,
795
+ "step": 1000
796
+ },
797
+ {
798
+ "epoch": 0.7849293563579278,
799
+ "eval/nano_beir.fiqa": 0.4530048021812019,
800
+ "step": 1000
801
+ },
802
+ {
803
+ "epoch": 0.7849293563579278,
804
+ "eval/nano_beir.hotpotqa": 0.6827182548147874,
805
+ "step": 1000
806
+ },
807
+ {
808
+ "epoch": 0.7849293563579278,
809
+ "eval/nano_beir.nfcorpus": 0.3587781171330431,
810
+ "step": 1000
811
+ },
812
+ {
813
+ "epoch": 0.7849293563579278,
814
+ "eval/nano_beir.nq": 0.426613822215205,
815
+ "step": 1000
816
+ },
817
+ {
818
+ "epoch": 0.7849293563579278,
819
+ "eval/nano_beir.quora": 0.8451303146637551,
820
+ "step": 1000
821
+ },
822
+ {
823
+ "epoch": 0.7849293563579278,
824
+ "eval/nano_beir.scidocs": 0.24604303501015984,
825
+ "step": 1000
826
+ },
827
+ {
828
+ "epoch": 0.7849293563579278,
829
+ "eval/nano_beir.scifact": 0.7460449406252556,
830
+ "step": 1000
831
+ },
832
+ {
833
+ "epoch": 0.7849293563579278,
834
+ "eval/nano_beir.webis_touche2020": 0.2860074344347365,
835
+ "step": 1000
836
+ },
837
+ {
838
+ "epoch": 0.7849293563579278,
839
+ "eval/avg": 0.47987284872416747,
840
+ "step": 1000
841
+ },
842
+ {
843
+ "epoch": 0.792778649921507,
844
+ "grad_norm": 5.71875,
845
+ "learning_rate": 9.68762555129222e-05,
846
+ "loss": 0.6397,
847
+ "step": 1010
848
+ },
849
+ {
850
+ "epoch": 0.8006279434850864,
851
+ "grad_norm": 5.5625,
852
+ "learning_rate": 9.675366799960841e-05,
853
+ "loss": 0.8047,
854
+ "step": 1020
855
+ },
856
+ {
857
+ "epoch": 0.8084772370486656,
858
+ "grad_norm": 6.0625,
859
+ "learning_rate": 9.662880177847594e-05,
860
+ "loss": 0.6725,
861
+ "step": 1030
862
+ },
863
+ {
864
+ "epoch": 0.8163265306122449,
865
+ "grad_norm": 5.0,
866
+ "learning_rate": 9.65016629353282e-05,
867
+ "loss": 0.6879,
868
+ "step": 1040
869
+ },
870
+ {
871
+ "epoch": 0.8241758241758241,
872
+ "grad_norm": 4.90625,
873
+ "learning_rate": 9.637225766673307e-05,
874
+ "loss": 0.7164,
875
+ "step": 1050
876
+ },
877
+ {
878
+ "epoch": 0.8320251177394035,
879
+ "grad_norm": 5.40625,
880
+ "learning_rate": 9.624059227972075e-05,
881
+ "loss": 0.7116,
882
+ "step": 1060
883
+ },
884
+ {
885
+ "epoch": 0.8398744113029827,
886
+ "grad_norm": 5.96875,
887
+ "learning_rate": 9.610667319147647e-05,
888
+ "loss": 0.7343,
889
+ "step": 1070
890
+ },
891
+ {
892
+ "epoch": 0.847723704866562,
893
+ "grad_norm": 5.6875,
894
+ "learning_rate": 9.597050692902765e-05,
895
+ "loss": 0.6432,
896
+ "step": 1080
897
+ },
898
+ {
899
+ "epoch": 0.8555729984301413,
900
+ "grad_norm": 4.90625,
901
+ "learning_rate": 9.583210012892582e-05,
902
+ "loss": 0.6632,
903
+ "step": 1090
904
+ },
905
+ {
906
+ "epoch": 0.8634222919937206,
907
+ "grad_norm": 6.40625,
908
+ "learning_rate": 9.569145953692315e-05,
909
+ "loss": 0.7413,
910
+ "step": 1100
911
+ },
912
+ {
913
+ "epoch": 0.8712715855572999,
914
+ "grad_norm": 4.40625,
915
+ "learning_rate": 9.55485920076437e-05,
916
+ "loss": 0.6581,
917
+ "step": 1110
918
+ },
919
+ {
920
+ "epoch": 0.8791208791208791,
921
+ "grad_norm": 5.625,
922
+ "learning_rate": 9.540350450424926e-05,
923
+ "loss": 0.7424,
924
+ "step": 1120
925
+ },
926
+ {
927
+ "epoch": 0.8869701726844584,
928
+ "grad_norm": 5.4375,
929
+ "learning_rate": 9.525620409810007e-05,
930
+ "loss": 0.5684,
931
+ "step": 1130
932
+ },
933
+ {
934
+ "epoch": 0.8948194662480377,
935
+ "grad_norm": 6.15625,
936
+ "learning_rate": 9.510669796841014e-05,
937
+ "loss": 0.6588,
938
+ "step": 1140
939
+ },
940
+ {
941
+ "epoch": 0.902668759811617,
942
+ "grad_norm": 5.34375,
943
+ "learning_rate": 9.495499340189728e-05,
944
+ "loss": 0.6428,
945
+ "step": 1150
946
+ },
947
+ {
948
+ "epoch": 0.9105180533751962,
949
+ "grad_norm": 4.84375,
950
+ "learning_rate": 9.480109779242805e-05,
951
+ "loss": 0.6236,
952
+ "step": 1160
953
+ },
954
+ {
955
+ "epoch": 0.9183673469387755,
956
+ "grad_norm": 5.5,
957
+ "learning_rate": 9.464501864065735e-05,
958
+ "loss": 0.7007,
959
+ "step": 1170
960
+ },
961
+ {
962
+ "epoch": 0.9262166405023547,
963
+ "grad_norm": 4.75,
964
+ "learning_rate": 9.448676355366282e-05,
965
+ "loss": 0.732,
966
+ "step": 1180
967
+ },
968
+ {
969
+ "epoch": 0.9340659340659341,
970
+ "grad_norm": 4.0,
971
+ "learning_rate": 9.432634024457414e-05,
972
+ "loss": 0.759,
973
+ "step": 1190
974
+ },
975
+ {
976
+ "epoch": 0.9419152276295133,
977
+ "grad_norm": 6.4375,
978
+ "learning_rate": 9.416375653219709e-05,
979
+ "loss": 0.6467,
980
+ "step": 1200
981
+ },
982
+ {
983
+ "epoch": 0.9497645211930926,
984
+ "grad_norm": 4.53125,
985
+ "learning_rate": 9.399902034063242e-05,
986
+ "loss": 0.6971,
987
+ "step": 1210
988
+ },
989
+ {
990
+ "epoch": 0.957613814756672,
991
+ "grad_norm": 4.5,
992
+ "learning_rate": 9.383213969888972e-05,
993
+ "loss": 0.6034,
994
+ "step": 1220
995
+ },
996
+ {
997
+ "epoch": 0.9654631083202512,
998
+ "grad_norm": 4.90625,
999
+ "learning_rate": 9.366312274049602e-05,
1000
+ "loss": 0.7138,
1001
+ "step": 1230
1002
+ },
1003
+ {
1004
+ "epoch": 0.9733124018838305,
1005
+ "grad_norm": 4.59375,
1006
+ "learning_rate": 9.34919777030994e-05,
1007
+ "loss": 0.7296,
1008
+ "step": 1240
1009
+ },
1010
+ {
1011
+ "epoch": 0.9811616954474097,
1012
+ "grad_norm": 4.6875,
1013
+ "learning_rate": 9.331871292806759e-05,
1014
+ "loss": 0.6068,
1015
+ "step": 1250
1016
+ },
1017
+ {
1018
+ "epoch": 0.989010989010989,
1019
+ "grad_norm": 6.09375,
1020
+ "learning_rate": 9.314333686008125e-05,
1021
+ "loss": 0.6281,
1022
+ "step": 1260
1023
+ },
1024
+ {
1025
+ "epoch": 0.9968602825745683,
1026
+ "grad_norm": 5.71875,
1027
+ "learning_rate": 9.296585804672252e-05,
1028
+ "loss": 0.5406,
1029
+ "step": 1270
1030
+ },
1031
+ {
1032
+ "epoch": 1.0047095761381475,
1033
+ "grad_norm": 5.3125,
1034
+ "learning_rate": 9.278628513805838e-05,
1035
+ "loss": 0.6783,
1036
+ "step": 1280
1037
+ },
1038
+ {
1039
+ "epoch": 1.012558869701727,
1040
+ "grad_norm": 5.15625,
1041
+ "learning_rate": 9.260462688621905e-05,
1042
+ "loss": 0.7021,
1043
+ "step": 1290
1044
+ },
1045
+ {
1046
+ "epoch": 1.0204081632653061,
1047
+ "grad_norm": 4.84375,
1048
+ "learning_rate": 9.242089214497145e-05,
1049
+ "loss": 0.6157,
1050
+ "step": 1300
1051
+ },
1052
+ {
1053
+ "epoch": 1.0282574568288854,
1054
+ "grad_norm": 4.8125,
1055
+ "learning_rate": 9.223508986928766e-05,
1056
+ "loss": 0.5672,
1057
+ "step": 1310
1058
+ },
1059
+ {
1060
+ "epoch": 1.0361067503924646,
1061
+ "grad_norm": 4.96875,
1062
+ "learning_rate": 9.204722911490846e-05,
1063
+ "loss": 0.711,
1064
+ "step": 1320
1065
+ },
1066
+ {
1067
+ "epoch": 1.043956043956044,
1068
+ "grad_norm": 5.84375,
1069
+ "learning_rate": 9.1857319037902e-05,
1070
+ "loss": 0.5718,
1071
+ "step": 1330
1072
+ },
1073
+ {
1074
+ "epoch": 1.0518053375196232,
1075
+ "grad_norm": 6.03125,
1076
+ "learning_rate": 9.16653688942175e-05,
1077
+ "loss": 0.6359,
1078
+ "step": 1340
1079
+ },
1080
+ {
1081
+ "epoch": 1.0596546310832025,
1082
+ "grad_norm": 6.125,
1083
+ "learning_rate": 9.147138803923416e-05,
1084
+ "loss": 0.6593,
1085
+ "step": 1350
1086
+ },
1087
+ {
1088
+ "epoch": 1.0675039246467817,
1089
+ "grad_norm": 7.78125,
1090
+ "learning_rate": 9.127538592730519e-05,
1091
+ "loss": 0.7177,
1092
+ "step": 1360
1093
+ },
1094
+ {
1095
+ "epoch": 1.0753532182103611,
1096
+ "grad_norm": 4.59375,
1097
+ "learning_rate": 9.107737211129701e-05,
1098
+ "loss": 0.6882,
1099
+ "step": 1370
1100
+ },
1101
+ {
1102
+ "epoch": 1.0832025117739403,
1103
+ "grad_norm": 5.59375,
1104
+ "learning_rate": 9.087735624212365e-05,
1105
+ "loss": 0.5998,
1106
+ "step": 1380
1107
+ },
1108
+ {
1109
+ "epoch": 1.0910518053375196,
1110
+ "grad_norm": 5.21875,
1111
+ "learning_rate": 9.06753480682764e-05,
1112
+ "loss": 0.7154,
1113
+ "step": 1390
1114
+ },
1115
+ {
1116
+ "epoch": 1.098901098901099,
1117
+ "grad_norm": 6.28125,
1118
+ "learning_rate": 9.047135743534868e-05,
1119
+ "loss": 0.6295,
1120
+ "step": 1400
1121
+ },
1122
+ {
1123
+ "epoch": 1.1067503924646782,
1124
+ "grad_norm": 6.25,
1125
+ "learning_rate": 9.02653942855561e-05,
1126
+ "loss": 0.6978,
1127
+ "step": 1410
1128
+ },
1129
+ {
1130
+ "epoch": 1.1145996860282574,
1131
+ "grad_norm": 4.6875,
1132
+ "learning_rate": 9.005746865725207e-05,
1133
+ "loss": 0.6089,
1134
+ "step": 1420
1135
+ },
1136
+ {
1137
+ "epoch": 1.1224489795918366,
1138
+ "grad_norm": 7.78125,
1139
+ "learning_rate": 8.984759068443831e-05,
1140
+ "loss": 0.5945,
1141
+ "step": 1430
1142
+ },
1143
+ {
1144
+ "epoch": 1.130298273155416,
1145
+ "grad_norm": 7.34375,
1146
+ "learning_rate": 8.963577059627118e-05,
1147
+ "loss": 0.6959,
1148
+ "step": 1440
1149
+ },
1150
+ {
1151
+ "epoch": 1.1381475667189953,
1152
+ "grad_norm": 5.3125,
1153
+ "learning_rate": 8.942201871656291e-05,
1154
+ "loss": 0.6495,
1155
+ "step": 1450
1156
+ },
1157
+ {
1158
+ "epoch": 1.1459968602825745,
1159
+ "grad_norm": 5.84375,
1160
+ "learning_rate": 8.920634546327858e-05,
1161
+ "loss": 0.6512,
1162
+ "step": 1460
1163
+ },
1164
+ {
1165
+ "epoch": 1.1538461538461537,
1166
+ "grad_norm": 4.6875,
1167
+ "learning_rate": 8.898876134802826e-05,
1168
+ "loss": 0.6301,
1169
+ "step": 1470
1170
+ },
1171
+ {
1172
+ "epoch": 1.1616954474097332,
1173
+ "grad_norm": 7.0625,
1174
+ "learning_rate": 8.87692769755548e-05,
1175
+ "loss": 0.5665,
1176
+ "step": 1480
1177
+ },
1178
+ {
1179
+ "epoch": 1.1695447409733124,
1180
+ "grad_norm": 5.71875,
1181
+ "learning_rate": 8.854790304321681e-05,
1182
+ "loss": 0.6361,
1183
+ "step": 1490
1184
+ },
1185
+ {
1186
+ "epoch": 1.1773940345368916,
1187
+ "grad_norm": 5.75,
1188
+ "learning_rate": 8.832465034046749e-05,
1189
+ "loss": 0.6687,
1190
+ "step": 1500
1191
+ },
1192
+ {
1193
+ "epoch": 1.185243328100471,
1194
+ "grad_norm": 6.3125,
1195
+ "learning_rate": 8.80995297483286e-05,
1196
+ "loss": 0.6106,
1197
+ "step": 1510
1198
+ },
1199
+ {
1200
+ "epoch": 1.1930926216640503,
1201
+ "grad_norm": 6.53125,
1202
+ "learning_rate": 8.787255223886019e-05,
1203
+ "loss": 0.6526,
1204
+ "step": 1520
1205
+ },
1206
+ {
1207
+ "epoch": 1.2009419152276295,
1208
+ "grad_norm": 5.6875,
1209
+ "learning_rate": 8.764372887462586e-05,
1210
+ "loss": 0.6845,
1211
+ "step": 1530
1212
+ },
1213
+ {
1214
+ "epoch": 1.2087912087912087,
1215
+ "grad_norm": 5.78125,
1216
+ "learning_rate": 8.741307080815357e-05,
1217
+ "loss": 0.6695,
1218
+ "step": 1540
1219
+ },
1220
+ {
1221
+ "epoch": 1.2166405023547882,
1222
+ "grad_norm": 4.15625,
1223
+ "learning_rate": 8.718058928139205e-05,
1224
+ "loss": 0.6021,
1225
+ "step": 1550
1226
+ },
1227
+ {
1228
+ "epoch": 1.2244897959183674,
1229
+ "grad_norm": 5.5625,
1230
+ "learning_rate": 8.694629562516294e-05,
1231
+ "loss": 0.6173,
1232
+ "step": 1560
1233
+ },
1234
+ {
1235
+ "epoch": 1.2323390894819466,
1236
+ "grad_norm": 4.125,
1237
+ "learning_rate": 8.671020125860851e-05,
1238
+ "loss": 0.5407,
1239
+ "step": 1570
1240
+ },
1241
+ {
1242
+ "epoch": 1.2401883830455258,
1243
+ "grad_norm": 6.28125,
1244
+ "learning_rate": 8.647231768863512e-05,
1245
+ "loss": 0.6177,
1246
+ "step": 1580
1247
+ },
1248
+ {
1249
+ "epoch": 1.2480376766091053,
1250
+ "grad_norm": 5.03125,
1251
+ "learning_rate": 8.623265650935234e-05,
1252
+ "loss": 0.6132,
1253
+ "step": 1590
1254
+ },
1255
+ {
1256
+ "epoch": 1.2558869701726845,
1257
+ "grad_norm": 5.9375,
1258
+ "learning_rate": 8.599122940150795e-05,
1259
+ "loss": 0.5913,
1260
+ "step": 1600
1261
+ },
1262
+ {
1263
+ "epoch": 1.2637362637362637,
1264
+ "grad_norm": 6.28125,
1265
+ "learning_rate": 8.574804813191858e-05,
1266
+ "loss": 0.7089,
1267
+ "step": 1610
1268
+ },
1269
+ {
1270
+ "epoch": 1.2715855572998431,
1271
+ "grad_norm": 7.3125,
1272
+ "learning_rate": 8.550312455289625e-05,
1273
+ "loss": 0.5426,
1274
+ "step": 1620
1275
+ },
1276
+ {
1277
+ "epoch": 1.2794348508634223,
1278
+ "grad_norm": 4.59375,
1279
+ "learning_rate": 8.525647060167063e-05,
1280
+ "loss": 0.5678,
1281
+ "step": 1630
1282
+ },
1283
+ {
1284
+ "epoch": 1.2872841444270016,
1285
+ "grad_norm": 6.34375,
1286
+ "learning_rate": 8.500809829980734e-05,
1287
+ "loss": 0.6343,
1288
+ "step": 1640
1289
+ },
1290
+ {
1291
+ "epoch": 1.2951334379905808,
1292
+ "grad_norm": 6.5625,
1293
+ "learning_rate": 8.4758019752622e-05,
1294
+ "loss": 0.7523,
1295
+ "step": 1650
1296
+ },
1297
+ {
1298
+ "epoch": 1.30298273155416,
1299
+ "grad_norm": 4.5,
1300
+ "learning_rate": 8.450624714859016e-05,
1301
+ "loss": 0.699,
1302
+ "step": 1660
1303
+ },
1304
+ {
1305
+ "epoch": 1.3108320251177394,
1306
+ "grad_norm": 5.09375,
1307
+ "learning_rate": 8.425279275875336e-05,
1308
+ "loss": 0.6517,
1309
+ "step": 1670
1310
+ },
1311
+ {
1312
+ "epoch": 1.3186813186813187,
1313
+ "grad_norm": 4.21875,
1314
+ "learning_rate": 8.399766893612096e-05,
1315
+ "loss": 0.563,
1316
+ "step": 1680
1317
+ },
1318
+ {
1319
+ "epoch": 1.3265306122448979,
1320
+ "grad_norm": 5.5625,
1321
+ "learning_rate": 8.374088811506819e-05,
1322
+ "loss": 0.6381,
1323
+ "step": 1690
1324
+ },
1325
+ {
1326
+ "epoch": 1.3343799058084773,
1327
+ "grad_norm": 4.6875,
1328
+ "learning_rate": 8.348246281072998e-05,
1329
+ "loss": 0.6794,
1330
+ "step": 1700
1331
+ },
1332
+ {
1333
+ "epoch": 1.3422291993720565,
1334
+ "grad_norm": 5.1875,
1335
+ "learning_rate": 8.322240561839109e-05,
1336
+ "loss": 0.6494,
1337
+ "step": 1710
1338
+ },
1339
+ {
1340
+ "epoch": 1.3500784929356358,
1341
+ "grad_norm": 5.09375,
1342
+ "learning_rate": 8.296072921287217e-05,
1343
+ "loss": 0.587,
1344
+ "step": 1720
1345
+ },
1346
+ {
1347
+ "epoch": 1.3579277864992152,
1348
+ "grad_norm": 6.34375,
1349
+ "learning_rate": 8.269744634791208e-05,
1350
+ "loss": 0.6888,
1351
+ "step": 1730
1352
+ },
1353
+ {
1354
+ "epoch": 1.3657770800627944,
1355
+ "grad_norm": 4.40625,
1356
+ "learning_rate": 8.243256985554621e-05,
1357
+ "loss": 0.6629,
1358
+ "step": 1740
1359
+ },
1360
+ {
1361
+ "epoch": 1.3736263736263736,
1362
+ "grad_norm": 5.4375,
1363
+ "learning_rate": 8.21661126454811e-05,
1364
+ "loss": 0.6633,
1365
+ "step": 1750
1366
+ },
1367
+ {
1368
+ "epoch": 1.3814756671899528,
1369
+ "grad_norm": 5.03125,
1370
+ "learning_rate": 8.189808770446528e-05,
1371
+ "loss": 0.633,
1372
+ "step": 1760
1373
+ },
1374
+ {
1375
+ "epoch": 1.389324960753532,
1376
+ "grad_norm": 4.59375,
1377
+ "learning_rate": 8.162850809565623e-05,
1378
+ "loss": 0.6725,
1379
+ "step": 1770
1380
+ },
1381
+ {
1382
+ "epoch": 1.3971742543171115,
1383
+ "grad_norm": 5.71875,
1384
+ "learning_rate": 8.135738695798376e-05,
1385
+ "loss": 0.6739,
1386
+ "step": 1780
1387
+ },
1388
+ {
1389
+ "epoch": 1.4050235478806907,
1390
+ "grad_norm": 5.34375,
1391
+ "learning_rate": 8.108473750550965e-05,
1392
+ "loss": 0.6547,
1393
+ "step": 1790
1394
+ },
1395
+ {
1396
+ "epoch": 1.41287284144427,
1397
+ "grad_norm": 6.0,
1398
+ "learning_rate": 8.081057302678352e-05,
1399
+ "loss": 0.5921,
1400
+ "step": 1800
1401
+ },
1402
+ {
1403
+ "epoch": 1.4207221350078494,
1404
+ "grad_norm": 6.5625,
1405
+ "learning_rate": 8.053490688419533e-05,
1406
+ "loss": 0.6639,
1407
+ "step": 1810
1408
+ },
1409
+ {
1410
+ "epoch": 1.4285714285714286,
1411
+ "grad_norm": 6.90625,
1412
+ "learning_rate": 8.02577525133239e-05,
1413
+ "loss": 0.6445,
1414
+ "step": 1820
1415
+ },
1416
+ {
1417
+ "epoch": 1.4364207221350078,
1418
+ "grad_norm": 6.375,
1419
+ "learning_rate": 7.997912342228232e-05,
1420
+ "loss": 0.6453,
1421
+ "step": 1830
1422
+ },
1423
+ {
1424
+ "epoch": 1.4442700156985873,
1425
+ "grad_norm": 6.34375,
1426
+ "learning_rate": 7.969903319105935e-05,
1427
+ "loss": 0.6649,
1428
+ "step": 1840
1429
+ },
1430
+ {
1431
+ "epoch": 1.4521193092621665,
1432
+ "grad_norm": 5.875,
1433
+ "learning_rate": 7.941749547085777e-05,
1434
+ "loss": 0.7093,
1435
+ "step": 1850
1436
+ },
1437
+ {
1438
+ "epoch": 1.4599686028257457,
1439
+ "grad_norm": 5.71875,
1440
+ "learning_rate": 7.913452398342881e-05,
1441
+ "loss": 0.5864,
1442
+ "step": 1860
1443
+ },
1444
+ {
1445
+ "epoch": 1.467817896389325,
1446
+ "grad_norm": 4.875,
1447
+ "learning_rate": 7.885013252040359e-05,
1448
+ "loss": 0.6377,
1449
+ "step": 1870
1450
+ },
1451
+ {
1452
+ "epoch": 1.4756671899529041,
1453
+ "grad_norm": 3.625,
1454
+ "learning_rate": 7.856433494262077e-05,
1455
+ "loss": 0.6205,
1456
+ "step": 1880
1457
+ },
1458
+ {
1459
+ "epoch": 1.4835164835164836,
1460
+ "grad_norm": 5.15625,
1461
+ "learning_rate": 7.827714517945115e-05,
1462
+ "loss": 0.6009,
1463
+ "step": 1890
1464
+ },
1465
+ {
1466
+ "epoch": 1.4913657770800628,
1467
+ "grad_norm": 3.890625,
1468
+ "learning_rate": 7.798857722811856e-05,
1469
+ "loss": 0.5702,
1470
+ "step": 1900
1471
+ },
1472
+ {
1473
+ "epoch": 1.499215070643642,
1474
+ "grad_norm": 5.03125,
1475
+ "learning_rate": 7.769864515301787e-05,
1476
+ "loss": 0.5561,
1477
+ "step": 1910
1478
+ },
1479
+ {
1480
+ "epoch": 1.5070643642072215,
1481
+ "grad_norm": 6.0625,
1482
+ "learning_rate": 7.740736308502938e-05,
1483
+ "loss": 0.6446,
1484
+ "step": 1920
1485
+ },
1486
+ {
1487
+ "epoch": 1.5149136577708007,
1488
+ "grad_norm": 6.34375,
1489
+ "learning_rate": 7.711474522083015e-05,
1490
+ "loss": 0.5983,
1491
+ "step": 1930
1492
+ },
1493
+ {
1494
+ "epoch": 1.5227629513343799,
1495
+ "grad_norm": 6.1875,
1496
+ "learning_rate": 7.682080582220206e-05,
1497
+ "loss": 0.5358,
1498
+ "step": 1940
1499
+ },
1500
+ {
1501
+ "epoch": 1.5306122448979593,
1502
+ "grad_norm": 6.5,
1503
+ "learning_rate": 7.65255592153367e-05,
1504
+ "loss": 0.7394,
1505
+ "step": 1950
1506
+ },
1507
+ {
1508
+ "epoch": 1.5384615384615383,
1509
+ "grad_norm": 4.8125,
1510
+ "learning_rate": 7.622901979013717e-05,
1511
+ "loss": 0.5742,
1512
+ "step": 1960
1513
+ },
1514
+ {
1515
+ "epoch": 1.5463108320251178,
1516
+ "grad_norm": 4.5,
1517
+ "learning_rate": 7.593120199951671e-05,
1518
+ "loss": 0.6425,
1519
+ "step": 1970
1520
+ },
1521
+ {
1522
+ "epoch": 1.554160125588697,
1523
+ "grad_norm": 6.03125,
1524
+ "learning_rate": 7.563212035869425e-05,
1525
+ "loss": 0.6593,
1526
+ "step": 1980
1527
+ },
1528
+ {
1529
+ "epoch": 1.5620094191522762,
1530
+ "grad_norm": 7.03125,
1531
+ "learning_rate": 7.533178944448705e-05,
1532
+ "loss": 0.5948,
1533
+ "step": 1990
1534
+ },
1535
+ {
1536
+ "epoch": 1.5698587127158556,
1537
+ "grad_norm": 6.0,
1538
+ "learning_rate": 7.503022389460014e-05,
1539
+ "loss": 0.625,
1540
+ "step": 2000
1541
+ },
1542
+ {
1543
+ "epoch": 1.5698587127158556,
1544
+ "eval/nano_beir.arguana": 0.47093852833484745,
1545
+ "step": 2000
1546
+ },
1547
+ {
1548
+ "epoch": 1.5698587127158556,
1549
+ "eval/nano_beir.climate_fever": 0.25927619589572015,
1550
+ "step": 2000
1551
+ },
1552
+ {
1553
+ "epoch": 1.5698587127158556,
1554
+ "eval/nano_beir.dbpedia_entity": 0.3266259450992667,
1555
+ "step": 2000
1556
+ },
1557
+ {
1558
+ "epoch": 1.5698587127158556,
1559
+ "eval/nano_beir.fever": 0.6291710052755145,
1560
+ "step": 2000
1561
+ },
1562
+ {
1563
+ "epoch": 1.5698587127158556,
1564
+ "eval/nano_beir.fiqa": 0.4343483963906797,
1565
+ "step": 2000
1566
+ },
1567
+ {
1568
+ "epoch": 1.5698587127158556,
1569
+ "eval/nano_beir.hotpotqa": 0.6854629804476208,
1570
+ "step": 2000
1571
+ },
1572
+ {
1573
+ "epoch": 1.5698587127158556,
1574
+ "eval/nano_beir.nfcorpus": 0.34320335172563987,
1575
+ "step": 2000
1576
+ },
1577
+ {
1578
+ "epoch": 1.5698587127158556,
1579
+ "eval/nano_beir.nq": 0.40466132495464113,
1580
+ "step": 2000
1581
+ },
1582
+ {
1583
+ "epoch": 1.5698587127158556,
1584
+ "eval/nano_beir.quora": 0.8346175392008862,
1585
+ "step": 2000
1586
+ },
1587
+ {
1588
+ "epoch": 1.5698587127158556,
1589
+ "eval/nano_beir.scidocs": 0.2536638155561289,
1590
+ "step": 2000
1591
+ },
1592
+ {
1593
+ "epoch": 1.5698587127158556,
1594
+ "eval/nano_beir.scifact": 0.756036525808922,
1595
+ "step": 2000
1596
+ },
1597
+ {
1598
+ "epoch": 1.5698587127158556,
1599
+ "eval/nano_beir.webis_touche2020": 0.282061170550218,
1600
+ "step": 2000
1601
+ },
1602
+ {
1603
+ "epoch": 1.5698587127158556,
1604
+ "eval/avg": 0.4733388982700071,
1605
+ "step": 2000
1606
+ },
1607
+ {
1608
+ "epoch": 1.5777080062794349,
1609
+ "grad_norm": 5.1875,
1610
+ "learning_rate": 7.472743840691299e-05,
1611
+ "loss": 0.6136,
1612
+ "step": 2010
1613
+ },
1614
+ {
1615
+ "epoch": 1.585557299843014,
1616
+ "grad_norm": 4.53125,
1617
+ "learning_rate": 7.44234477387631e-05,
1618
+ "loss": 0.5169,
1619
+ "step": 2020
1620
+ },
1621
+ {
1622
+ "epoch": 1.5934065934065935,
1623
+ "grad_norm": 4.0625,
1624
+ "learning_rate": 7.411826670622676e-05,
1625
+ "loss": 0.6241,
1626
+ "step": 2030
1627
+ },
1628
+ {
1629
+ "epoch": 1.6012558869701727,
1630
+ "grad_norm": 5.5,
1631
+ "learning_rate": 7.381191018339696e-05,
1632
+ "loss": 0.6384,
1633
+ "step": 2040
1634
+ },
1635
+ {
1636
+ "epoch": 1.609105180533752,
1637
+ "grad_norm": 4.78125,
1638
+ "learning_rate": 7.350439310165842e-05,
1639
+ "loss": 0.687,
1640
+ "step": 2050
1641
+ },
1642
+ {
1643
+ "epoch": 1.6169544740973314,
1644
+ "grad_norm": 5.84375,
1645
+ "learning_rate": 7.319573044895986e-05,
1646
+ "loss": 0.6424,
1647
+ "step": 2060
1648
+ },
1649
+ {
1650
+ "epoch": 1.6248037676609104,
1651
+ "grad_norm": 6.34375,
1652
+ "learning_rate": 7.28859372690835e-05,
1653
+ "loss": 0.6997,
1654
+ "step": 2070
1655
+ },
1656
+ {
1657
+ "epoch": 1.6326530612244898,
1658
+ "grad_norm": 5.59375,
1659
+ "learning_rate": 7.257502866091192e-05,
1660
+ "loss": 0.5591,
1661
+ "step": 2080
1662
+ },
1663
+ {
1664
+ "epoch": 1.640502354788069,
1665
+ "grad_norm": 5.75,
1666
+ "learning_rate": 7.226301977769199e-05,
1667
+ "loss": 0.5663,
1668
+ "step": 2090
1669
+ },
1670
+ {
1671
+ "epoch": 1.6483516483516483,
1672
+ "grad_norm": 6.15625,
1673
+ "learning_rate": 7.194992582629654e-05,
1674
+ "loss": 0.6399,
1675
+ "step": 2100
1676
+ },
1677
+ {
1678
+ "epoch": 1.6562009419152277,
1679
+ "grad_norm": 6.5625,
1680
+ "learning_rate": 7.163576206648304e-05,
1681
+ "loss": 0.6225,
1682
+ "step": 2110
1683
+ },
1684
+ {
1685
+ "epoch": 1.664050235478807,
1686
+ "grad_norm": 6.28125,
1687
+ "learning_rate": 7.132054381014995e-05,
1688
+ "loss": 0.6461,
1689
+ "step": 2120
1690
+ },
1691
+ {
1692
+ "epoch": 1.6718995290423861,
1693
+ "grad_norm": 5.5625,
1694
+ "learning_rate": 7.100428642059033e-05,
1695
+ "loss": 0.6079,
1696
+ "step": 2130
1697
+ },
1698
+ {
1699
+ "epoch": 1.6797488226059656,
1700
+ "grad_norm": 5.8125,
1701
+ "learning_rate": 7.068700531174319e-05,
1702
+ "loss": 0.6539,
1703
+ "step": 2140
1704
+ },
1705
+ {
1706
+ "epoch": 1.6875981161695446,
1707
+ "grad_norm": 6.0,
1708
+ "learning_rate": 7.036871594744217e-05,
1709
+ "loss": 0.6949,
1710
+ "step": 2150
1711
+ },
1712
+ {
1713
+ "epoch": 1.695447409733124,
1714
+ "grad_norm": 5.34375,
1715
+ "learning_rate": 7.004943384066187e-05,
1716
+ "loss": 0.5566,
1717
+ "step": 2160
1718
+ },
1719
+ {
1720
+ "epoch": 1.7032967032967035,
1721
+ "grad_norm": 5.71875,
1722
+ "learning_rate": 6.97291745527617e-05,
1723
+ "loss": 0.691,
1724
+ "step": 2170
1725
+ },
1726
+ {
1727
+ "epoch": 1.7111459968602825,
1728
+ "grad_norm": 5.21875,
1729
+ "learning_rate": 6.940795369272754e-05,
1730
+ "loss": 0.7199,
1731
+ "step": 2180
1732
+ },
1733
+ {
1734
+ "epoch": 1.718995290423862,
1735
+ "grad_norm": 5.1875,
1736
+ "learning_rate": 6.908578691641092e-05,
1737
+ "loss": 0.6165,
1738
+ "step": 2190
1739
+ },
1740
+ {
1741
+ "epoch": 1.7268445839874411,
1742
+ "grad_norm": 5.0625,
1743
+ "learning_rate": 6.876268992576604e-05,
1744
+ "loss": 0.598,
1745
+ "step": 2200
1746
+ },
1747
+ {
1748
+ "epoch": 1.7346938775510203,
1749
+ "grad_norm": 5.125,
1750
+ "learning_rate": 6.843867846808438e-05,
1751
+ "loss": 0.5891,
1752
+ "step": 2210
1753
+ },
1754
+ {
1755
+ "epoch": 1.7425431711145998,
1756
+ "grad_norm": 5.90625,
1757
+ "learning_rate": 6.811376833522729e-05,
1758
+ "loss": 0.6333,
1759
+ "step": 2220
1760
+ },
1761
+ {
1762
+ "epoch": 1.750392464678179,
1763
+ "grad_norm": 5.59375,
1764
+ "learning_rate": 6.778797536285624e-05,
1765
+ "loss": 0.5931,
1766
+ "step": 2230
1767
+ },
1768
+ {
1769
+ "epoch": 1.7582417582417582,
1770
+ "grad_norm": 5.15625,
1771
+ "learning_rate": 6.746131542966112e-05,
1772
+ "loss": 0.6326,
1773
+ "step": 2240
1774
+ },
1775
+ {
1776
+ "epoch": 1.7660910518053377,
1777
+ "grad_norm": 5.625,
1778
+ "learning_rate": 6.713380445658618e-05,
1779
+ "loss": 0.6492,
1780
+ "step": 2250
1781
+ },
1782
+ {
1783
+ "epoch": 1.7739403453689166,
1784
+ "grad_norm": 7.0625,
1785
+ "learning_rate": 6.680545840605423e-05,
1786
+ "loss": 0.5642,
1787
+ "step": 2260
1788
+ },
1789
+ {
1790
+ "epoch": 1.781789638932496,
1791
+ "grad_norm": 5.6875,
1792
+ "learning_rate": 6.647629328118851e-05,
1793
+ "loss": 0.5615,
1794
+ "step": 2270
1795
+ },
1796
+ {
1797
+ "epoch": 1.7896389324960753,
1798
+ "grad_norm": 4.625,
1799
+ "learning_rate": 6.614632512503288e-05,
1800
+ "loss": 0.5933,
1801
+ "step": 2280
1802
+ },
1803
+ {
1804
+ "epoch": 1.7974882260596545,
1805
+ "grad_norm": 6.125,
1806
+ "learning_rate": 6.58155700197697e-05,
1807
+ "loss": 0.6289,
1808
+ "step": 2290
1809
+ },
1810
+ {
1811
+ "epoch": 1.805337519623234,
1812
+ "grad_norm": 5.21875,
1813
+ "learning_rate": 6.548404408593621e-05,
1814
+ "loss": 0.6227,
1815
+ "step": 2300
1816
+ },
1817
+ {
1818
+ "epoch": 1.8131868131868132,
1819
+ "grad_norm": 5.75,
1820
+ "learning_rate": 6.515176348163871e-05,
1821
+ "loss": 0.5941,
1822
+ "step": 2310
1823
+ },
1824
+ {
1825
+ "epoch": 1.8210361067503924,
1826
+ "grad_norm": 5.59375,
1827
+ "learning_rate": 6.481874440176506e-05,
1828
+ "loss": 0.5882,
1829
+ "step": 2320
1830
+ },
1831
+ {
1832
+ "epoch": 1.8288854003139718,
1833
+ "grad_norm": 5.6875,
1834
+ "learning_rate": 6.448500307719538e-05,
1835
+ "loss": 0.6208,
1836
+ "step": 2330
1837
+ },
1838
+ {
1839
+ "epoch": 1.836734693877551,
1840
+ "grad_norm": 4.5,
1841
+ "learning_rate": 6.415055577401102e-05,
1842
+ "loss": 0.5798,
1843
+ "step": 2340
1844
+ },
1845
+ {
1846
+ "epoch": 1.8445839874411303,
1847
+ "grad_norm": 3.9375,
1848
+ "learning_rate": 6.381541879270169e-05,
1849
+ "loss": 0.6352,
1850
+ "step": 2350
1851
+ },
1852
+ {
1853
+ "epoch": 1.8524332810047097,
1854
+ "grad_norm": 5.5,
1855
+ "learning_rate": 6.347960846737106e-05,
1856
+ "loss": 0.7087,
1857
+ "step": 2360
1858
+ },
1859
+ {
1860
+ "epoch": 1.8602825745682887,
1861
+ "grad_norm": 4.625,
1862
+ "learning_rate": 6.31431411649406e-05,
1863
+ "loss": 0.5722,
1864
+ "step": 2370
1865
+ },
1866
+ {
1867
+ "epoch": 1.8681318681318682,
1868
+ "grad_norm": 5.625,
1869
+ "learning_rate": 6.280603328435198e-05,
1870
+ "loss": 0.6634,
1871
+ "step": 2380
1872
+ },
1873
+ {
1874
+ "epoch": 1.8759811616954474,
1875
+ "grad_norm": 5.78125,
1876
+ "learning_rate": 6.24683012557677e-05,
1877
+ "loss": 0.6851,
1878
+ "step": 2390
1879
+ },
1880
+ {
1881
+ "epoch": 1.8838304552590266,
1882
+ "grad_norm": 5.875,
1883
+ "learning_rate": 6.212996153977037e-05,
1884
+ "loss": 0.6613,
1885
+ "step": 2400
1886
+ },
1887
+ {
1888
+ "epoch": 1.891679748822606,
1889
+ "grad_norm": 4.3125,
1890
+ "learning_rate": 6.179103062656042e-05,
1891
+ "loss": 0.6541,
1892
+ "step": 2410
1893
+ },
1894
+ {
1895
+ "epoch": 1.8995290423861853,
1896
+ "grad_norm": 7.34375,
1897
+ "learning_rate": 6.145152503515239e-05,
1898
+ "loss": 0.714,
1899
+ "step": 2420
1900
+ },
1901
+ {
1902
+ "epoch": 1.9073783359497645,
1903
+ "grad_norm": 5.25,
1904
+ "learning_rate": 6.111146131256983e-05,
1905
+ "loss": 0.614,
1906
+ "step": 2430
1907
+ },
1908
+ {
1909
+ "epoch": 1.915227629513344,
1910
+ "grad_norm": 5.0625,
1911
+ "learning_rate": 6.077085603303882e-05,
1912
+ "loss": 0.5535,
1913
+ "step": 2440
1914
+ },
1915
+ {
1916
+ "epoch": 1.9230769230769231,
1917
+ "grad_norm": 6.40625,
1918
+ "learning_rate": 6.042972579718019e-05,
1919
+ "loss": 0.7302,
1920
+ "step": 2450
1921
+ },
1922
+ {
1923
+ "epoch": 1.9309262166405023,
1924
+ "grad_norm": 4.5,
1925
+ "learning_rate": 6.008808723120035e-05,
1926
+ "loss": 0.6042,
1927
+ "step": 2460
1928
+ },
1929
+ {
1930
+ "epoch": 1.9387755102040818,
1931
+ "grad_norm": 5.8125,
1932
+ "learning_rate": 5.974595698608103e-05,
1933
+ "loss": 0.603,
1934
+ "step": 2470
1935
+ },
1936
+ {
1937
+ "epoch": 1.9466248037676608,
1938
+ "grad_norm": 5.1875,
1939
+ "learning_rate": 5.9403351736767696e-05,
1940
+ "loss": 0.668,
1941
+ "step": 2480
1942
+ },
1943
+ {
1944
+ "epoch": 1.9544740973312402,
1945
+ "grad_norm": 5.46875,
1946
+ "learning_rate": 5.906028818135687e-05,
1947
+ "loss": 0.6311,
1948
+ "step": 2490
1949
+ },
1950
+ {
1951
+ "epoch": 1.9623233908948194,
1952
+ "grad_norm": 5.40625,
1953
+ "learning_rate": 5.8716783040282244e-05,
1954
+ "loss": 0.5565,
1955
+ "step": 2500
1956
+ },
1957
+ {
1958
+ "epoch": 1.9701726844583987,
1959
+ "grad_norm": 4.875,
1960
+ "learning_rate": 5.837285305549978e-05,
1961
+ "loss": 0.6793,
1962
+ "step": 2510
1963
+ },
1964
+ {
1965
+ "epoch": 1.978021978021978,
1966
+ "grad_norm": 6.6875,
1967
+ "learning_rate": 5.8028514989671724e-05,
1968
+ "loss": 0.6645,
1969
+ "step": 2520
1970
+ },
1971
+ {
1972
+ "epoch": 1.9858712715855573,
1973
+ "grad_norm": 5.65625,
1974
+ "learning_rate": 5.7683785625349616e-05,
1975
+ "loss": 0.5526,
1976
+ "step": 2530
1977
+ },
1978
+ {
1979
+ "epoch": 1.9937205651491365,
1980
+ "grad_norm": 6.59375,
1981
+ "learning_rate": 5.733868176415633e-05,
1982
+ "loss": 0.6974,
1983
+ "step": 2540
1984
+ },
1985
+ {
1986
+ "epoch": 2.001569858712716,
1987
+ "grad_norm": 4.875,
1988
+ "learning_rate": 5.699322022596722e-05,
1989
+ "loss": 0.7787,
1990
+ "step": 2550
1991
+ },
1992
+ {
1993
+ "epoch": 2.009419152276295,
1994
+ "grad_norm": 5.03125,
1995
+ "learning_rate": 5.664741784809022e-05,
1996
+ "loss": 0.6121,
1997
+ "step": 2560
1998
+ },
1999
+ {
2000
+ "epoch": 2.0172684458398744,
2001
+ "grad_norm": 4.4375,
2002
+ "learning_rate": 5.630129148444543e-05,
2003
+ "loss": 0.5956,
2004
+ "step": 2570
2005
+ },
2006
+ {
2007
+ "epoch": 2.025117739403454,
2008
+ "grad_norm": 4.4375,
2009
+ "learning_rate": 5.595485800474349e-05,
2010
+ "loss": 0.6243,
2011
+ "step": 2580
2012
+ },
2013
+ {
2014
+ "epoch": 2.032967032967033,
2015
+ "grad_norm": 6.625,
2016
+ "learning_rate": 5.560813429366345e-05,
2017
+ "loss": 0.5992,
2018
+ "step": 2590
2019
+ },
2020
+ {
2021
+ "epoch": 2.0408163265306123,
2022
+ "grad_norm": 7.03125,
2023
+ "learning_rate": 5.5261137250029835e-05,
2024
+ "loss": 0.5197,
2025
+ "step": 2600
2026
+ },
2027
+ {
2028
+ "epoch": 2.0486656200941917,
2029
+ "grad_norm": 4.78125,
2030
+ "learning_rate": 5.4913883785988993e-05,
2031
+ "loss": 0.5492,
2032
+ "step": 2610
2033
+ },
2034
+ {
2035
+ "epoch": 2.0565149136577707,
2036
+ "grad_norm": 5.8125,
2037
+ "learning_rate": 5.456639082618489e-05,
2038
+ "loss": 0.5624,
2039
+ "step": 2620
2040
+ },
2041
+ {
2042
+ "epoch": 2.06436420722135,
2043
+ "grad_norm": 5.90625,
2044
+ "learning_rate": 5.421867530693414e-05,
2045
+ "loss": 0.5266,
2046
+ "step": 2630
2047
+ },
2048
+ {
2049
+ "epoch": 2.072213500784929,
2050
+ "grad_norm": 5.0,
2051
+ "learning_rate": 5.3870754175400595e-05,
2052
+ "loss": 0.5205,
2053
+ "step": 2640
2054
+ },
2055
+ {
2056
+ "epoch": 2.0800627943485086,
2057
+ "grad_norm": 5.59375,
2058
+ "learning_rate": 5.352264438876935e-05,
2059
+ "loss": 0.6017,
2060
+ "step": 2650
2061
+ },
2062
+ {
2063
+ "epoch": 2.087912087912088,
2064
+ "grad_norm": 5.125,
2065
+ "learning_rate": 5.3174362913420306e-05,
2066
+ "loss": 0.565,
2067
+ "step": 2660
2068
+ },
2069
+ {
2070
+ "epoch": 2.095761381475667,
2071
+ "grad_norm": 4.875,
2072
+ "learning_rate": 5.2825926724101236e-05,
2073
+ "loss": 0.6248,
2074
+ "step": 2670
2075
+ },
2076
+ {
2077
+ "epoch": 2.1036106750392465,
2078
+ "grad_norm": 6.09375,
2079
+ "learning_rate": 5.247735280310041e-05,
2080
+ "loss": 0.7087,
2081
+ "step": 2680
2082
+ },
2083
+ {
2084
+ "epoch": 2.111459968602826,
2085
+ "grad_norm": 5.25,
2086
+ "learning_rate": 5.212865813941899e-05,
2087
+ "loss": 0.4872,
2088
+ "step": 2690
2089
+ },
2090
+ {
2091
+ "epoch": 2.119309262166405,
2092
+ "grad_norm": 6.0,
2093
+ "learning_rate": 5.1779859727942924e-05,
2094
+ "loss": 0.6128,
2095
+ "step": 2700
2096
+ },
2097
+ {
2098
+ "epoch": 2.1271585557299844,
2099
+ "grad_norm": 4.9375,
2100
+ "learning_rate": 5.143097456861474e-05,
2101
+ "loss": 0.5494,
2102
+ "step": 2710
2103
+ },
2104
+ {
2105
+ "epoch": 2.1350078492935634,
2106
+ "grad_norm": 5.5625,
2107
+ "learning_rate": 5.1082019665604895e-05,
2108
+ "loss": 0.5941,
2109
+ "step": 2720
2110
+ },
2111
+ {
2112
+ "epoch": 2.142857142857143,
2113
+ "grad_norm": 6.0625,
2114
+ "learning_rate": 5.073301202648304e-05,
2115
+ "loss": 0.5385,
2116
+ "step": 2730
2117
+ },
2118
+ {
2119
+ "epoch": 2.1507064364207222,
2120
+ "grad_norm": 7.25,
2121
+ "learning_rate": 5.0383968661389146e-05,
2122
+ "loss": 0.5394,
2123
+ "step": 2740
2124
+ },
2125
+ {
2126
+ "epoch": 2.1585557299843012,
2127
+ "grad_norm": 7.59375,
2128
+ "learning_rate": 5.003490658220438e-05,
2129
+ "loss": 0.5874,
2130
+ "step": 2750
2131
+ },
2132
+ {
2133
+ "epoch": 2.1664050235478807,
2134
+ "grad_norm": 5.15625,
2135
+ "learning_rate": 4.968584280172206e-05,
2136
+ "loss": 0.5373,
2137
+ "step": 2760
2138
+ },
2139
+ {
2140
+ "epoch": 2.17425431711146,
2141
+ "grad_norm": 6.65625,
2142
+ "learning_rate": 4.933679433281836e-05,
2143
+ "loss": 0.5698,
2144
+ "step": 2770
2145
+ },
2146
+ {
2147
+ "epoch": 2.182103610675039,
2148
+ "grad_norm": 6.25,
2149
+ "learning_rate": 4.8987778187623245e-05,
2150
+ "loss": 0.6354,
2151
+ "step": 2780
2152
+ },
2153
+ {
2154
+ "epoch": 2.1899529042386185,
2155
+ "grad_norm": 6.6875,
2156
+ "learning_rate": 4.863881137669123e-05,
2157
+ "loss": 0.5519,
2158
+ "step": 2790
2159
+ },
2160
+ {
2161
+ "epoch": 2.197802197802198,
2162
+ "grad_norm": 5.59375,
2163
+ "learning_rate": 4.8289910908172376e-05,
2164
+ "loss": 0.6345,
2165
+ "step": 2800
2166
+ },
2167
+ {
2168
+ "epoch": 2.205651491365777,
2169
+ "grad_norm": 5.5,
2170
+ "learning_rate": 4.794109378698327e-05,
2171
+ "loss": 0.5927,
2172
+ "step": 2810
2173
+ },
2174
+ {
2175
+ "epoch": 2.2135007849293564,
2176
+ "grad_norm": 5.71875,
2177
+ "learning_rate": 4.7592377013978306e-05,
2178
+ "loss": 0.5696,
2179
+ "step": 2820
2180
+ },
2181
+ {
2182
+ "epoch": 2.221350078492936,
2183
+ "grad_norm": 5.21875,
2184
+ "learning_rate": 4.724377758512103e-05,
2185
+ "loss": 0.5136,
2186
+ "step": 2830
2187
+ },
2188
+ {
2189
+ "epoch": 2.229199372056515,
2190
+ "grad_norm": 5.75,
2191
+ "learning_rate": 4.68953124906558e-05,
2192
+ "loss": 0.6424,
2193
+ "step": 2840
2194
+ },
2195
+ {
2196
+ "epoch": 2.2370486656200943,
2197
+ "grad_norm": 5.75,
2198
+ "learning_rate": 4.654699871427971e-05,
2199
+ "loss": 0.6576,
2200
+ "step": 2850
2201
+ },
2202
+ {
2203
+ "epoch": 2.2448979591836733,
2204
+ "grad_norm": 7.34375,
2205
+ "learning_rate": 4.619885323231484e-05,
2206
+ "loss": 0.5978,
2207
+ "step": 2860
2208
+ },
2209
+ {
2210
+ "epoch": 2.2527472527472527,
2211
+ "grad_norm": 6.25,
2212
+ "learning_rate": 4.58508930128808e-05,
2213
+ "loss": 0.5672,
2214
+ "step": 2870
2215
+ },
2216
+ {
2217
+ "epoch": 2.260596546310832,
2218
+ "grad_norm": 5.90625,
2219
+ "learning_rate": 4.550313501506781e-05,
2220
+ "loss": 0.6515,
2221
+ "step": 2880
2222
+ },
2223
+ {
2224
+ "epoch": 2.268445839874411,
2225
+ "grad_norm": 5.34375,
2226
+ "learning_rate": 4.515559618811005e-05,
2227
+ "loss": 0.4906,
2228
+ "step": 2890
2229
+ },
2230
+ {
2231
+ "epoch": 2.2762951334379906,
2232
+ "grad_norm": 6.59375,
2233
+ "learning_rate": 4.4808293470559643e-05,
2234
+ "loss": 0.5254,
2235
+ "step": 2900
2236
+ },
2237
+ {
2238
+ "epoch": 2.2841444270015696,
2239
+ "grad_norm": 5.75,
2240
+ "learning_rate": 4.446124378946107e-05,
2241
+ "loss": 0.6452,
2242
+ "step": 2910
2243
+ },
2244
+ {
2245
+ "epoch": 2.291993720565149,
2246
+ "grad_norm": 5.84375,
2247
+ "learning_rate": 4.411446405952618e-05,
2248
+ "loss": 0.5403,
2249
+ "step": 2920
2250
+ },
2251
+ {
2252
+ "epoch": 2.2998430141287285,
2253
+ "grad_norm": 4.75,
2254
+ "learning_rate": 4.376797118230979e-05,
2255
+ "loss": 0.5628,
2256
+ "step": 2930
2257
+ },
2258
+ {
2259
+ "epoch": 2.3076923076923075,
2260
+ "grad_norm": 6.1875,
2261
+ "learning_rate": 4.342178204538588e-05,
2262
+ "loss": 0.5914,
2263
+ "step": 2940
2264
+ },
2265
+ {
2266
+ "epoch": 2.315541601255887,
2267
+ "grad_norm": 5.8125,
2268
+ "learning_rate": 4.307591352152459e-05,
2269
+ "loss": 0.556,
2270
+ "step": 2950
2271
+ },
2272
+ {
2273
+ "epoch": 2.3233908948194664,
2274
+ "grad_norm": 7.125,
2275
+ "learning_rate": 4.273038246786986e-05,
2276
+ "loss": 0.6112,
2277
+ "step": 2960
2278
+ },
2279
+ {
2280
+ "epoch": 2.3312401883830454,
2281
+ "grad_norm": 5.9375,
2282
+ "learning_rate": 4.238520572511773e-05,
2283
+ "loss": 0.5364,
2284
+ "step": 2970
2285
+ },
2286
+ {
2287
+ "epoch": 2.339089481946625,
2288
+ "grad_norm": 5.1875,
2289
+ "learning_rate": 4.2040400116695666e-05,
2290
+ "loss": 0.5563,
2291
+ "step": 2980
2292
+ },
2293
+ {
2294
+ "epoch": 2.3469387755102042,
2295
+ "grad_norm": 5.15625,
2296
+ "learning_rate": 4.169598244794261e-05,
2297
+ "loss": 0.5306,
2298
+ "step": 2990
2299
+ },
2300
+ {
2301
+ "epoch": 2.3547880690737832,
2302
+ "grad_norm": 5.8125,
2303
+ "learning_rate": 4.135196950528982e-05,
2304
+ "loss": 0.5319,
2305
+ "step": 3000
2306
+ }
2307
+ ],
2308
+ "logging_steps": 10,
2309
+ "max_steps": 5000,
2310
+ "num_input_tokens_seen": 0,
2311
+ "num_train_epochs": 4,
2312
+ "save_steps": 1000,
2313
+ "stateful_callbacks": {
2314
+ "TrainerControl": {
2315
+ "args": {
2316
+ "should_epoch_stop": false,
2317
+ "should_evaluate": false,
2318
+ "should_log": false,
2319
+ "should_save": true,
2320
+ "should_training_stop": false
2321
+ },
2322
+ "attributes": {}
2323
+ }
2324
+ },
2325
+ "total_flos": 0.0,
2326
+ "train_batch_size": 16,
2327
+ "trial_name": null,
2328
+ "trial_params": null
2329
+ }
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-3000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2ed821ed7ef0074e4b13618254be9ea2c38d68708aace5456b6baa256883072
3
+ size 6097
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "gradient_checkpointing": false,
20
+ "hidden_activation": "gelu",
21
+ "hidden_size": 768,
22
+ "initializer_cutoff_factor": 2.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 1152,
25
+ "layer_norm_eps": 1e-05,
26
+ "local_attention": 128,
27
+ "local_rope_theta": 10000.0,
28
+ "max_position_embeddings": 8192,
29
+ "mlp_bias": false,
30
+ "mlp_dropout": 0.0,
31
+ "model_type": "modernbert",
32
+ "norm_bias": false,
33
+ "norm_eps": 1e-05,
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 22,
36
+ "pad_token_id": 50283,
37
+ "position_embedding_type": "absolute",
38
+ "repad_logits_with_grad": false,
39
+ "sep_token_id": 50282,
40
+ "sparse_pred_ignore_index": -100,
41
+ "sparse_prediction": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.53.0",
44
+ "vocab_size": 50368
45
+ }
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb8c18aefb03ba8afaf3ddafca1cd6ea74416c497632a03907346835d107adad
3
+ size 298041696
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7ca94df487520ae2c411613298582fccb5e50edc91edd1e3311d560a4ff7427
3
+ size 596170443
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24c12cceaffe8acec04189ba71dbb81f100dd5cd03b86390253117b3fab2911d
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c24e46cf0599262f5ee2ba9ec6bddad7d3e8feb965b424cc9c4dc4563771b1e2
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5af6ed530327c4af91aa04cc19a7523ed8295478ca34bb5f3124ff88cd33f77
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74c82a4770595a28e603ecf5353ab454c0fb9b0e184187b20ebdc4685217f3a5
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9218a8f161ba45a13a24f24cd020f32b0abec4bcc53a2764e48b480568dd2d13
3
+ size 1465
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/trainer_state.json ADDED
@@ -0,0 +1,3094 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 3.1397174254317113,
6
+ "eval_steps": 1000.0,
7
+ "global_step": 4000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0,
14
+ "eval/nano_beir.arguana": 0.4400549254974839,
15
+ "step": 0
16
+ },
17
+ {
18
+ "epoch": 0,
19
+ "eval/nano_beir.climate_fever": 0.22880053057901228,
20
+ "step": 0
21
+ },
22
+ {
23
+ "epoch": 0,
24
+ "eval/nano_beir.dbpedia_entity": 0.30026716538531234,
25
+ "step": 0
26
+ },
27
+ {
28
+ "epoch": 0,
29
+ "eval/nano_beir.fever": 0.6419178788751609,
30
+ "step": 0
31
+ },
32
+ {
33
+ "epoch": 0,
34
+ "eval/nano_beir.fiqa": 0.4370593595651809,
35
+ "step": 0
36
+ },
37
+ {
38
+ "epoch": 0,
39
+ "eval/nano_beir.hotpotqa": 0.6723047477224562,
40
+ "step": 0
41
+ },
42
+ {
43
+ "epoch": 0,
44
+ "eval/nano_beir.nfcorpus": 0.3437446614516137,
45
+ "step": 0
46
+ },
47
+ {
48
+ "epoch": 0,
49
+ "eval/nano_beir.nq": 0.436711543163819,
50
+ "step": 0
51
+ },
52
+ {
53
+ "epoch": 0,
54
+ "eval/nano_beir.quora": 0.8439997970269911,
55
+ "step": 0
56
+ },
57
+ {
58
+ "epoch": 0,
59
+ "eval/nano_beir.scidocs": 0.26304331393922176,
60
+ "step": 0
61
+ },
62
+ {
63
+ "epoch": 0,
64
+ "eval/nano_beir.scifact": 0.7599341408912704,
65
+ "step": 0
66
+ },
67
+ {
68
+ "epoch": 0,
69
+ "eval/nano_beir.webis_touche2020": 0.29696490432942557,
70
+ "step": 0
71
+ },
72
+ {
73
+ "epoch": 0,
74
+ "eval/avg": 0.47206691403557904,
75
+ "step": 0
76
+ },
77
+ {
78
+ "epoch": 0.007849293563579277,
79
+ "grad_norm": 9.3125,
80
+ "learning_rate": 1.8e-06,
81
+ "loss": 0.8676,
82
+ "step": 10
83
+ },
84
+ {
85
+ "epoch": 0.015698587127158554,
86
+ "grad_norm": 7.46875,
87
+ "learning_rate": 3.8e-06,
88
+ "loss": 0.9307,
89
+ "step": 20
90
+ },
91
+ {
92
+ "epoch": 0.023547880690737835,
93
+ "grad_norm": 9.0,
94
+ "learning_rate": 5.8e-06,
95
+ "loss": 0.8335,
96
+ "step": 30
97
+ },
98
+ {
99
+ "epoch": 0.03139717425431711,
100
+ "grad_norm": 7.15625,
101
+ "learning_rate": 7.8e-06,
102
+ "loss": 0.8277,
103
+ "step": 40
104
+ },
105
+ {
106
+ "epoch": 0.03924646781789639,
107
+ "grad_norm": 7.65625,
108
+ "learning_rate": 9.800000000000001e-06,
109
+ "loss": 0.8458,
110
+ "step": 50
111
+ },
112
+ {
113
+ "epoch": 0.04709576138147567,
114
+ "grad_norm": 9.125,
115
+ "learning_rate": 1.18e-05,
116
+ "loss": 0.8074,
117
+ "step": 60
118
+ },
119
+ {
120
+ "epoch": 0.054945054945054944,
121
+ "grad_norm": 8.1875,
122
+ "learning_rate": 1.3800000000000002e-05,
123
+ "loss": 0.7318,
124
+ "step": 70
125
+ },
126
+ {
127
+ "epoch": 0.06279434850863422,
128
+ "grad_norm": 7.65625,
129
+ "learning_rate": 1.58e-05,
130
+ "loss": 0.8056,
131
+ "step": 80
132
+ },
133
+ {
134
+ "epoch": 0.0706436420722135,
135
+ "grad_norm": 8.625,
136
+ "learning_rate": 1.78e-05,
137
+ "loss": 0.8091,
138
+ "step": 90
139
+ },
140
+ {
141
+ "epoch": 0.07849293563579278,
142
+ "grad_norm": 7.78125,
143
+ "learning_rate": 1.9800000000000004e-05,
144
+ "loss": 0.8001,
145
+ "step": 100
146
+ },
147
+ {
148
+ "epoch": 0.08634222919937205,
149
+ "grad_norm": 8.0625,
150
+ "learning_rate": 2.18e-05,
151
+ "loss": 0.9097,
152
+ "step": 110
153
+ },
154
+ {
155
+ "epoch": 0.09419152276295134,
156
+ "grad_norm": 5.78125,
157
+ "learning_rate": 2.38e-05,
158
+ "loss": 0.7487,
159
+ "step": 120
160
+ },
161
+ {
162
+ "epoch": 0.10204081632653061,
163
+ "grad_norm": 6.5625,
164
+ "learning_rate": 2.58e-05,
165
+ "loss": 0.7602,
166
+ "step": 130
167
+ },
168
+ {
169
+ "epoch": 0.10989010989010989,
170
+ "grad_norm": 6.40625,
171
+ "learning_rate": 2.7800000000000005e-05,
172
+ "loss": 0.7309,
173
+ "step": 140
174
+ },
175
+ {
176
+ "epoch": 0.11773940345368916,
177
+ "grad_norm": 6.75,
178
+ "learning_rate": 2.98e-05,
179
+ "loss": 0.7629,
180
+ "step": 150
181
+ },
182
+ {
183
+ "epoch": 0.12558869701726844,
184
+ "grad_norm": 6.46875,
185
+ "learning_rate": 3.18e-05,
186
+ "loss": 0.8294,
187
+ "step": 160
188
+ },
189
+ {
190
+ "epoch": 0.13343799058084774,
191
+ "grad_norm": 6.53125,
192
+ "learning_rate": 3.38e-05,
193
+ "loss": 0.6725,
194
+ "step": 170
195
+ },
196
+ {
197
+ "epoch": 0.141287284144427,
198
+ "grad_norm": 6.59375,
199
+ "learning_rate": 3.58e-05,
200
+ "loss": 0.6788,
201
+ "step": 180
202
+ },
203
+ {
204
+ "epoch": 0.14913657770800628,
205
+ "grad_norm": 6.125,
206
+ "learning_rate": 3.7800000000000004e-05,
207
+ "loss": 0.7473,
208
+ "step": 190
209
+ },
210
+ {
211
+ "epoch": 0.15698587127158556,
212
+ "grad_norm": 6.5625,
213
+ "learning_rate": 3.9800000000000005e-05,
214
+ "loss": 0.7462,
215
+ "step": 200
216
+ },
217
+ {
218
+ "epoch": 0.16483516483516483,
219
+ "grad_norm": 6.53125,
220
+ "learning_rate": 4.18e-05,
221
+ "loss": 0.6637,
222
+ "step": 210
223
+ },
224
+ {
225
+ "epoch": 0.1726844583987441,
226
+ "grad_norm": 6.96875,
227
+ "learning_rate": 4.38e-05,
228
+ "loss": 0.7876,
229
+ "step": 220
230
+ },
231
+ {
232
+ "epoch": 0.18053375196232338,
233
+ "grad_norm": 8.0,
234
+ "learning_rate": 4.58e-05,
235
+ "loss": 0.7846,
236
+ "step": 230
237
+ },
238
+ {
239
+ "epoch": 0.18838304552590268,
240
+ "grad_norm": 9.875,
241
+ "learning_rate": 4.78e-05,
242
+ "loss": 0.7945,
243
+ "step": 240
244
+ },
245
+ {
246
+ "epoch": 0.19623233908948196,
247
+ "grad_norm": 7.625,
248
+ "learning_rate": 4.9800000000000004e-05,
249
+ "loss": 0.7045,
250
+ "step": 250
251
+ },
252
+ {
253
+ "epoch": 0.20408163265306123,
254
+ "grad_norm": 5.1875,
255
+ "learning_rate": 5.1800000000000005e-05,
256
+ "loss": 0.7334,
257
+ "step": 260
258
+ },
259
+ {
260
+ "epoch": 0.2119309262166405,
261
+ "grad_norm": 5.625,
262
+ "learning_rate": 5.380000000000001e-05,
263
+ "loss": 0.7552,
264
+ "step": 270
265
+ },
266
+ {
267
+ "epoch": 0.21978021978021978,
268
+ "grad_norm": 4.125,
269
+ "learning_rate": 5.580000000000001e-05,
270
+ "loss": 0.7312,
271
+ "step": 280
272
+ },
273
+ {
274
+ "epoch": 0.22762951334379905,
275
+ "grad_norm": 6.1875,
276
+ "learning_rate": 5.7799999999999995e-05,
277
+ "loss": 0.6843,
278
+ "step": 290
279
+ },
280
+ {
281
+ "epoch": 0.23547880690737832,
282
+ "grad_norm": 5.9375,
283
+ "learning_rate": 5.9800000000000003e-05,
284
+ "loss": 0.7139,
285
+ "step": 300
286
+ },
287
+ {
288
+ "epoch": 0.24332810047095763,
289
+ "grad_norm": 4.90625,
290
+ "learning_rate": 6.18e-05,
291
+ "loss": 0.6777,
292
+ "step": 310
293
+ },
294
+ {
295
+ "epoch": 0.25117739403453687,
296
+ "grad_norm": 5.25,
297
+ "learning_rate": 6.38e-05,
298
+ "loss": 0.7767,
299
+ "step": 320
300
+ },
301
+ {
302
+ "epoch": 0.25902668759811615,
303
+ "grad_norm": 5.71875,
304
+ "learning_rate": 6.58e-05,
305
+ "loss": 0.7373,
306
+ "step": 330
307
+ },
308
+ {
309
+ "epoch": 0.2668759811616955,
310
+ "grad_norm": 6.0,
311
+ "learning_rate": 6.780000000000001e-05,
312
+ "loss": 0.8043,
313
+ "step": 340
314
+ },
315
+ {
316
+ "epoch": 0.27472527472527475,
317
+ "grad_norm": 5.6875,
318
+ "learning_rate": 6.98e-05,
319
+ "loss": 0.748,
320
+ "step": 350
321
+ },
322
+ {
323
+ "epoch": 0.282574568288854,
324
+ "grad_norm": 5.21875,
325
+ "learning_rate": 7.18e-05,
326
+ "loss": 0.6968,
327
+ "step": 360
328
+ },
329
+ {
330
+ "epoch": 0.2904238618524333,
331
+ "grad_norm": 5.65625,
332
+ "learning_rate": 7.38e-05,
333
+ "loss": 0.6794,
334
+ "step": 370
335
+ },
336
+ {
337
+ "epoch": 0.29827315541601257,
338
+ "grad_norm": 5.71875,
339
+ "learning_rate": 7.58e-05,
340
+ "loss": 0.6401,
341
+ "step": 380
342
+ },
343
+ {
344
+ "epoch": 0.30612244897959184,
345
+ "grad_norm": 6.40625,
346
+ "learning_rate": 7.780000000000001e-05,
347
+ "loss": 0.8027,
348
+ "step": 390
349
+ },
350
+ {
351
+ "epoch": 0.3139717425431711,
352
+ "grad_norm": 6.9375,
353
+ "learning_rate": 7.98e-05,
354
+ "loss": 0.6472,
355
+ "step": 400
356
+ },
357
+ {
358
+ "epoch": 0.3218210361067504,
359
+ "grad_norm": 5.5625,
360
+ "learning_rate": 8.18e-05,
361
+ "loss": 0.7221,
362
+ "step": 410
363
+ },
364
+ {
365
+ "epoch": 0.32967032967032966,
366
+ "grad_norm": 5.5,
367
+ "learning_rate": 8.38e-05,
368
+ "loss": 0.668,
369
+ "step": 420
370
+ },
371
+ {
372
+ "epoch": 0.33751962323390894,
373
+ "grad_norm": 6.09375,
374
+ "learning_rate": 8.58e-05,
375
+ "loss": 0.6923,
376
+ "step": 430
377
+ },
378
+ {
379
+ "epoch": 0.3453689167974882,
380
+ "grad_norm": 7.09375,
381
+ "learning_rate": 8.78e-05,
382
+ "loss": 0.6982,
383
+ "step": 440
384
+ },
385
+ {
386
+ "epoch": 0.3532182103610675,
387
+ "grad_norm": 5.5,
388
+ "learning_rate": 8.98e-05,
389
+ "loss": 0.6671,
390
+ "step": 450
391
+ },
392
+ {
393
+ "epoch": 0.36106750392464676,
394
+ "grad_norm": 6.1875,
395
+ "learning_rate": 9.180000000000001e-05,
396
+ "loss": 0.6487,
397
+ "step": 460
398
+ },
399
+ {
400
+ "epoch": 0.36891679748822603,
401
+ "grad_norm": 6.875,
402
+ "learning_rate": 9.38e-05,
403
+ "loss": 0.7471,
404
+ "step": 470
405
+ },
406
+ {
407
+ "epoch": 0.37676609105180536,
408
+ "grad_norm": 5.71875,
409
+ "learning_rate": 9.58e-05,
410
+ "loss": 0.6445,
411
+ "step": 480
412
+ },
413
+ {
414
+ "epoch": 0.38461538461538464,
415
+ "grad_norm": 5.125,
416
+ "learning_rate": 9.78e-05,
417
+ "loss": 0.659,
418
+ "step": 490
419
+ },
420
+ {
421
+ "epoch": 0.3924646781789639,
422
+ "grad_norm": 7.96875,
423
+ "learning_rate": 9.98e-05,
424
+ "loss": 0.6469,
425
+ "step": 500
426
+ },
427
+ {
428
+ "epoch": 0.4003139717425432,
429
+ "grad_norm": 6.5625,
430
+ "learning_rate": 9.999901304280685e-05,
431
+ "loss": 0.6404,
432
+ "step": 510
433
+ },
434
+ {
435
+ "epoch": 0.40816326530612246,
436
+ "grad_norm": 7.5625,
437
+ "learning_rate": 9.999560138895238e-05,
438
+ "loss": 0.7289,
439
+ "step": 520
440
+ },
441
+ {
442
+ "epoch": 0.41601255886970173,
443
+ "grad_norm": 5.5625,
444
+ "learning_rate": 9.99897530200195e-05,
445
+ "loss": 0.6332,
446
+ "step": 530
447
+ },
448
+ {
449
+ "epoch": 0.423861852433281,
450
+ "grad_norm": 5.09375,
451
+ "learning_rate": 9.998146822104943e-05,
452
+ "loss": 0.7495,
453
+ "step": 540
454
+ },
455
+ {
456
+ "epoch": 0.4317111459968603,
457
+ "grad_norm": 4.625,
458
+ "learning_rate": 9.997074739583162e-05,
459
+ "loss": 0.6474,
460
+ "step": 550
461
+ },
462
+ {
463
+ "epoch": 0.43956043956043955,
464
+ "grad_norm": 5.4375,
465
+ "learning_rate": 9.995759106688393e-05,
466
+ "loss": 0.6788,
467
+ "step": 560
468
+ },
469
+ {
470
+ "epoch": 0.4474097331240188,
471
+ "grad_norm": 5.5,
472
+ "learning_rate": 9.99419998754273e-05,
473
+ "loss": 0.6584,
474
+ "step": 570
475
+ },
476
+ {
477
+ "epoch": 0.4552590266875981,
478
+ "grad_norm": 5.125,
479
+ "learning_rate": 9.992397458135438e-05,
480
+ "loss": 0.7894,
481
+ "step": 580
482
+ },
483
+ {
484
+ "epoch": 0.4631083202511774,
485
+ "grad_norm": 5.71875,
486
+ "learning_rate": 9.990351606319261e-05,
487
+ "loss": 0.7216,
488
+ "step": 590
489
+ },
490
+ {
491
+ "epoch": 0.47095761381475665,
492
+ "grad_norm": 5.40625,
493
+ "learning_rate": 9.988062531806126e-05,
494
+ "loss": 0.6335,
495
+ "step": 600
496
+ },
497
+ {
498
+ "epoch": 0.478806907378336,
499
+ "grad_norm": 7.0,
500
+ "learning_rate": 9.985530346162299e-05,
501
+ "loss": 0.6338,
502
+ "step": 610
503
+ },
504
+ {
505
+ "epoch": 0.48665620094191525,
506
+ "grad_norm": 5.375,
507
+ "learning_rate": 9.982755172802933e-05,
508
+ "loss": 0.6607,
509
+ "step": 620
510
+ },
511
+ {
512
+ "epoch": 0.4945054945054945,
513
+ "grad_norm": 5.59375,
514
+ "learning_rate": 9.979737146986064e-05,
515
+ "loss": 0.6775,
516
+ "step": 630
517
+ },
518
+ {
519
+ "epoch": 0.5023547880690737,
520
+ "grad_norm": 9.0,
521
+ "learning_rate": 9.976476415806013e-05,
522
+ "loss": 0.642,
523
+ "step": 640
524
+ },
525
+ {
526
+ "epoch": 0.5102040816326531,
527
+ "grad_norm": 7.46875,
528
+ "learning_rate": 9.972973138186217e-05,
529
+ "loss": 0.7126,
530
+ "step": 650
531
+ },
532
+ {
533
+ "epoch": 0.5180533751962323,
534
+ "grad_norm": 4.15625,
535
+ "learning_rate": 9.969227484871484e-05,
536
+ "loss": 0.6095,
537
+ "step": 660
538
+ },
539
+ {
540
+ "epoch": 0.5259026687598116,
541
+ "grad_norm": 5.03125,
542
+ "learning_rate": 9.965239638419672e-05,
543
+ "loss": 0.7593,
544
+ "step": 670
545
+ },
546
+ {
547
+ "epoch": 0.533751962323391,
548
+ "grad_norm": 5.1875,
549
+ "learning_rate": 9.961009793192792e-05,
550
+ "loss": 0.7108,
551
+ "step": 680
552
+ },
553
+ {
554
+ "epoch": 0.5416012558869702,
555
+ "grad_norm": 6.0,
556
+ "learning_rate": 9.956538155347534e-05,
557
+ "loss": 0.7327,
558
+ "step": 690
559
+ },
560
+ {
561
+ "epoch": 0.5494505494505495,
562
+ "grad_norm": 4.8125,
563
+ "learning_rate": 9.951824942825215e-05,
564
+ "loss": 0.677,
565
+ "step": 700
566
+ },
567
+ {
568
+ "epoch": 0.5572998430141287,
569
+ "grad_norm": 4.6875,
570
+ "learning_rate": 9.946870385341167e-05,
571
+ "loss": 0.6671,
572
+ "step": 710
573
+ },
574
+ {
575
+ "epoch": 0.565149136577708,
576
+ "grad_norm": 4.8125,
577
+ "learning_rate": 9.941674724373531e-05,
578
+ "loss": 0.604,
579
+ "step": 720
580
+ },
581
+ {
582
+ "epoch": 0.5729984301412873,
583
+ "grad_norm": 5.84375,
584
+ "learning_rate": 9.936238213151491e-05,
585
+ "loss": 0.6487,
586
+ "step": 730
587
+ },
588
+ {
589
+ "epoch": 0.5808477237048666,
590
+ "grad_norm": 5.90625,
591
+ "learning_rate": 9.930561116642935e-05,
592
+ "loss": 0.6417,
593
+ "step": 740
594
+ },
595
+ {
596
+ "epoch": 0.5886970172684458,
597
+ "grad_norm": 6.03125,
598
+ "learning_rate": 9.924643711541539e-05,
599
+ "loss": 0.8061,
600
+ "step": 750
601
+ },
602
+ {
603
+ "epoch": 0.5965463108320251,
604
+ "grad_norm": 4.90625,
605
+ "learning_rate": 9.918486286253279e-05,
606
+ "loss": 0.7204,
607
+ "step": 760
608
+ },
609
+ {
610
+ "epoch": 0.6043956043956044,
611
+ "grad_norm": 4.5,
612
+ "learning_rate": 9.912089140882376e-05,
613
+ "loss": 0.6467,
614
+ "step": 770
615
+ },
616
+ {
617
+ "epoch": 0.6122448979591837,
618
+ "grad_norm": 6.5,
619
+ "learning_rate": 9.90545258721667e-05,
620
+ "loss": 0.6307,
621
+ "step": 780
622
+ },
623
+ {
624
+ "epoch": 0.6200941915227629,
625
+ "grad_norm": 6.71875,
626
+ "learning_rate": 9.898576948712426e-05,
627
+ "loss": 0.8016,
628
+ "step": 790
629
+ },
630
+ {
631
+ "epoch": 0.6279434850863422,
632
+ "grad_norm": 8.0,
633
+ "learning_rate": 9.891462560478562e-05,
634
+ "loss": 0.7312,
635
+ "step": 800
636
+ },
637
+ {
638
+ "epoch": 0.6357927786499215,
639
+ "grad_norm": 6.03125,
640
+ "learning_rate": 9.884109769260325e-05,
641
+ "loss": 0.7053,
642
+ "step": 810
643
+ },
644
+ {
645
+ "epoch": 0.6436420722135008,
646
+ "grad_norm": 4.21875,
647
+ "learning_rate": 9.876518933422386e-05,
648
+ "loss": 0.6855,
649
+ "step": 820
650
+ },
651
+ {
652
+ "epoch": 0.6514913657770801,
653
+ "grad_norm": 5.28125,
654
+ "learning_rate": 9.868690422931372e-05,
655
+ "loss": 0.7619,
656
+ "step": 830
657
+ },
658
+ {
659
+ "epoch": 0.6593406593406593,
660
+ "grad_norm": 8.25,
661
+ "learning_rate": 9.860624619337844e-05,
662
+ "loss": 0.7078,
663
+ "step": 840
664
+ },
665
+ {
666
+ "epoch": 0.6671899529042387,
667
+ "grad_norm": 5.3125,
668
+ "learning_rate": 9.852321915757687e-05,
669
+ "loss": 0.6612,
670
+ "step": 850
671
+ },
672
+ {
673
+ "epoch": 0.6750392464678179,
674
+ "grad_norm": 5.8125,
675
+ "learning_rate": 9.843782716852963e-05,
676
+ "loss": 0.7664,
677
+ "step": 860
678
+ },
679
+ {
680
+ "epoch": 0.6828885400313972,
681
+ "grad_norm": 4.90625,
682
+ "learning_rate": 9.835007438812177e-05,
683
+ "loss": 0.7502,
684
+ "step": 870
685
+ },
686
+ {
687
+ "epoch": 0.6907378335949764,
688
+ "grad_norm": 4.84375,
689
+ "learning_rate": 9.82599650933e-05,
690
+ "loss": 0.6931,
691
+ "step": 880
692
+ },
693
+ {
694
+ "epoch": 0.6985871271585558,
695
+ "grad_norm": 4.78125,
696
+ "learning_rate": 9.816750367586425e-05,
697
+ "loss": 0.7217,
698
+ "step": 890
699
+ },
700
+ {
701
+ "epoch": 0.706436420722135,
702
+ "grad_norm": 5.4375,
703
+ "learning_rate": 9.807269464225355e-05,
704
+ "loss": 0.6607,
705
+ "step": 900
706
+ },
707
+ {
708
+ "epoch": 0.7142857142857143,
709
+ "grad_norm": 6.40625,
710
+ "learning_rate": 9.797554261332645e-05,
711
+ "loss": 0.7778,
712
+ "step": 910
713
+ },
714
+ {
715
+ "epoch": 0.7221350078492935,
716
+ "grad_norm": 6.90625,
717
+ "learning_rate": 9.787605232413574e-05,
718
+ "loss": 0.7161,
719
+ "step": 920
720
+ },
721
+ {
722
+ "epoch": 0.7299843014128728,
723
+ "grad_norm": 6.28125,
724
+ "learning_rate": 9.777422862369783e-05,
725
+ "loss": 0.6461,
726
+ "step": 930
727
+ },
728
+ {
729
+ "epoch": 0.7378335949764521,
730
+ "grad_norm": 6.125,
731
+ "learning_rate": 9.767007647475617e-05,
732
+ "loss": 0.6601,
733
+ "step": 940
734
+ },
735
+ {
736
+ "epoch": 0.7456828885400314,
737
+ "grad_norm": 5.375,
738
+ "learning_rate": 9.756360095353956e-05,
739
+ "loss": 0.6062,
740
+ "step": 950
741
+ },
742
+ {
743
+ "epoch": 0.7535321821036107,
744
+ "grad_norm": 7.03125,
745
+ "learning_rate": 9.745480724951473e-05,
746
+ "loss": 0.7328,
747
+ "step": 960
748
+ },
749
+ {
750
+ "epoch": 0.7613814756671899,
751
+ "grad_norm": 5.34375,
752
+ "learning_rate": 9.73437006651333e-05,
753
+ "loss": 0.6816,
754
+ "step": 970
755
+ },
756
+ {
757
+ "epoch": 0.7692307692307693,
758
+ "grad_norm": 4.90625,
759
+ "learning_rate": 9.723028661557345e-05,
760
+ "loss": 0.7171,
761
+ "step": 980
762
+ },
763
+ {
764
+ "epoch": 0.7770800627943485,
765
+ "grad_norm": 3.828125,
766
+ "learning_rate": 9.711457062847595e-05,
767
+ "loss": 0.6988,
768
+ "step": 990
769
+ },
770
+ {
771
+ "epoch": 0.7849293563579278,
772
+ "grad_norm": 5.09375,
773
+ "learning_rate": 9.699655834367478e-05,
774
+ "loss": 0.6331,
775
+ "step": 1000
776
+ },
777
+ {
778
+ "epoch": 0.7849293563579278,
779
+ "eval/nano_beir.arguana": 0.4803257504428309,
780
+ "step": 1000
781
+ },
782
+ {
783
+ "epoch": 0.7849293563579278,
784
+ "eval/nano_beir.climate_fever": 0.25852105142356685,
785
+ "step": 1000
786
+ },
787
+ {
788
+ "epoch": 0.7849293563579278,
789
+ "eval/nano_beir.dbpedia_entity": 0.33382012080541146,
790
+ "step": 1000
791
+ },
792
+ {
793
+ "epoch": 0.7849293563579278,
794
+ "eval/nano_beir.fever": 0.6414665409400568,
795
+ "step": 1000
796
+ },
797
+ {
798
+ "epoch": 0.7849293563579278,
799
+ "eval/nano_beir.fiqa": 0.4530048021812019,
800
+ "step": 1000
801
+ },
802
+ {
803
+ "epoch": 0.7849293563579278,
804
+ "eval/nano_beir.hotpotqa": 0.6827182548147874,
805
+ "step": 1000
806
+ },
807
+ {
808
+ "epoch": 0.7849293563579278,
809
+ "eval/nano_beir.nfcorpus": 0.3587781171330431,
810
+ "step": 1000
811
+ },
812
+ {
813
+ "epoch": 0.7849293563579278,
814
+ "eval/nano_beir.nq": 0.426613822215205,
815
+ "step": 1000
816
+ },
817
+ {
818
+ "epoch": 0.7849293563579278,
819
+ "eval/nano_beir.quora": 0.8451303146637551,
820
+ "step": 1000
821
+ },
822
+ {
823
+ "epoch": 0.7849293563579278,
824
+ "eval/nano_beir.scidocs": 0.24604303501015984,
825
+ "step": 1000
826
+ },
827
+ {
828
+ "epoch": 0.7849293563579278,
829
+ "eval/nano_beir.scifact": 0.7460449406252556,
830
+ "step": 1000
831
+ },
832
+ {
833
+ "epoch": 0.7849293563579278,
834
+ "eval/nano_beir.webis_touche2020": 0.2860074344347365,
835
+ "step": 1000
836
+ },
837
+ {
838
+ "epoch": 0.7849293563579278,
839
+ "eval/avg": 0.47987284872416747,
840
+ "step": 1000
841
+ },
842
+ {
843
+ "epoch": 0.792778649921507,
844
+ "grad_norm": 5.71875,
845
+ "learning_rate": 9.68762555129222e-05,
846
+ "loss": 0.6397,
847
+ "step": 1010
848
+ },
849
+ {
850
+ "epoch": 0.8006279434850864,
851
+ "grad_norm": 5.5625,
852
+ "learning_rate": 9.675366799960841e-05,
853
+ "loss": 0.8047,
854
+ "step": 1020
855
+ },
856
+ {
857
+ "epoch": 0.8084772370486656,
858
+ "grad_norm": 6.0625,
859
+ "learning_rate": 9.662880177847594e-05,
860
+ "loss": 0.6725,
861
+ "step": 1030
862
+ },
863
+ {
864
+ "epoch": 0.8163265306122449,
865
+ "grad_norm": 5.0,
866
+ "learning_rate": 9.65016629353282e-05,
867
+ "loss": 0.6879,
868
+ "step": 1040
869
+ },
870
+ {
871
+ "epoch": 0.8241758241758241,
872
+ "grad_norm": 4.90625,
873
+ "learning_rate": 9.637225766673307e-05,
874
+ "loss": 0.7164,
875
+ "step": 1050
876
+ },
877
+ {
878
+ "epoch": 0.8320251177394035,
879
+ "grad_norm": 5.40625,
880
+ "learning_rate": 9.624059227972075e-05,
881
+ "loss": 0.7116,
882
+ "step": 1060
883
+ },
884
+ {
885
+ "epoch": 0.8398744113029827,
886
+ "grad_norm": 5.96875,
887
+ "learning_rate": 9.610667319147647e-05,
888
+ "loss": 0.7343,
889
+ "step": 1070
890
+ },
891
+ {
892
+ "epoch": 0.847723704866562,
893
+ "grad_norm": 5.6875,
894
+ "learning_rate": 9.597050692902765e-05,
895
+ "loss": 0.6432,
896
+ "step": 1080
897
+ },
898
+ {
899
+ "epoch": 0.8555729984301413,
900
+ "grad_norm": 4.90625,
901
+ "learning_rate": 9.583210012892582e-05,
902
+ "loss": 0.6632,
903
+ "step": 1090
904
+ },
905
+ {
906
+ "epoch": 0.8634222919937206,
907
+ "grad_norm": 6.40625,
908
+ "learning_rate": 9.569145953692315e-05,
909
+ "loss": 0.7413,
910
+ "step": 1100
911
+ },
912
+ {
913
+ "epoch": 0.8712715855572999,
914
+ "grad_norm": 4.40625,
915
+ "learning_rate": 9.55485920076437e-05,
916
+ "loss": 0.6581,
917
+ "step": 1110
918
+ },
919
+ {
920
+ "epoch": 0.8791208791208791,
921
+ "grad_norm": 5.625,
922
+ "learning_rate": 9.540350450424926e-05,
923
+ "loss": 0.7424,
924
+ "step": 1120
925
+ },
926
+ {
927
+ "epoch": 0.8869701726844584,
928
+ "grad_norm": 5.4375,
929
+ "learning_rate": 9.525620409810007e-05,
930
+ "loss": 0.5684,
931
+ "step": 1130
932
+ },
933
+ {
934
+ "epoch": 0.8948194662480377,
935
+ "grad_norm": 6.15625,
936
+ "learning_rate": 9.510669796841014e-05,
937
+ "loss": 0.6588,
938
+ "step": 1140
939
+ },
940
+ {
941
+ "epoch": 0.902668759811617,
942
+ "grad_norm": 5.34375,
943
+ "learning_rate": 9.495499340189728e-05,
944
+ "loss": 0.6428,
945
+ "step": 1150
946
+ },
947
+ {
948
+ "epoch": 0.9105180533751962,
949
+ "grad_norm": 4.84375,
950
+ "learning_rate": 9.480109779242805e-05,
951
+ "loss": 0.6236,
952
+ "step": 1160
953
+ },
954
+ {
955
+ "epoch": 0.9183673469387755,
956
+ "grad_norm": 5.5,
957
+ "learning_rate": 9.464501864065735e-05,
958
+ "loss": 0.7007,
959
+ "step": 1170
960
+ },
961
+ {
962
+ "epoch": 0.9262166405023547,
963
+ "grad_norm": 4.75,
964
+ "learning_rate": 9.448676355366282e-05,
965
+ "loss": 0.732,
966
+ "step": 1180
967
+ },
968
+ {
969
+ "epoch": 0.9340659340659341,
970
+ "grad_norm": 4.0,
971
+ "learning_rate": 9.432634024457414e-05,
972
+ "loss": 0.759,
973
+ "step": 1190
974
+ },
975
+ {
976
+ "epoch": 0.9419152276295133,
977
+ "grad_norm": 6.4375,
978
+ "learning_rate": 9.416375653219709e-05,
979
+ "loss": 0.6467,
980
+ "step": 1200
981
+ },
982
+ {
983
+ "epoch": 0.9497645211930926,
984
+ "grad_norm": 4.53125,
985
+ "learning_rate": 9.399902034063242e-05,
986
+ "loss": 0.6971,
987
+ "step": 1210
988
+ },
989
+ {
990
+ "epoch": 0.957613814756672,
991
+ "grad_norm": 4.5,
992
+ "learning_rate": 9.383213969888972e-05,
993
+ "loss": 0.6034,
994
+ "step": 1220
995
+ },
996
+ {
997
+ "epoch": 0.9654631083202512,
998
+ "grad_norm": 4.90625,
999
+ "learning_rate": 9.366312274049602e-05,
1000
+ "loss": 0.7138,
1001
+ "step": 1230
1002
+ },
1003
+ {
1004
+ "epoch": 0.9733124018838305,
1005
+ "grad_norm": 4.59375,
1006
+ "learning_rate": 9.34919777030994e-05,
1007
+ "loss": 0.7296,
1008
+ "step": 1240
1009
+ },
1010
+ {
1011
+ "epoch": 0.9811616954474097,
1012
+ "grad_norm": 4.6875,
1013
+ "learning_rate": 9.331871292806759e-05,
1014
+ "loss": 0.6068,
1015
+ "step": 1250
1016
+ },
1017
+ {
1018
+ "epoch": 0.989010989010989,
1019
+ "grad_norm": 6.09375,
1020
+ "learning_rate": 9.314333686008125e-05,
1021
+ "loss": 0.6281,
1022
+ "step": 1260
1023
+ },
1024
+ {
1025
+ "epoch": 0.9968602825745683,
1026
+ "grad_norm": 5.71875,
1027
+ "learning_rate": 9.296585804672252e-05,
1028
+ "loss": 0.5406,
1029
+ "step": 1270
1030
+ },
1031
+ {
1032
+ "epoch": 1.0047095761381475,
1033
+ "grad_norm": 5.3125,
1034
+ "learning_rate": 9.278628513805838e-05,
1035
+ "loss": 0.6783,
1036
+ "step": 1280
1037
+ },
1038
+ {
1039
+ "epoch": 1.012558869701727,
1040
+ "grad_norm": 5.15625,
1041
+ "learning_rate": 9.260462688621905e-05,
1042
+ "loss": 0.7021,
1043
+ "step": 1290
1044
+ },
1045
+ {
1046
+ "epoch": 1.0204081632653061,
1047
+ "grad_norm": 4.84375,
1048
+ "learning_rate": 9.242089214497145e-05,
1049
+ "loss": 0.6157,
1050
+ "step": 1300
1051
+ },
1052
+ {
1053
+ "epoch": 1.0282574568288854,
1054
+ "grad_norm": 4.8125,
1055
+ "learning_rate": 9.223508986928766e-05,
1056
+ "loss": 0.5672,
1057
+ "step": 1310
1058
+ },
1059
+ {
1060
+ "epoch": 1.0361067503924646,
1061
+ "grad_norm": 4.96875,
1062
+ "learning_rate": 9.204722911490846e-05,
1063
+ "loss": 0.711,
1064
+ "step": 1320
1065
+ },
1066
+ {
1067
+ "epoch": 1.043956043956044,
1068
+ "grad_norm": 5.84375,
1069
+ "learning_rate": 9.1857319037902e-05,
1070
+ "loss": 0.5718,
1071
+ "step": 1330
1072
+ },
1073
+ {
1074
+ "epoch": 1.0518053375196232,
1075
+ "grad_norm": 6.03125,
1076
+ "learning_rate": 9.16653688942175e-05,
1077
+ "loss": 0.6359,
1078
+ "step": 1340
1079
+ },
1080
+ {
1081
+ "epoch": 1.0596546310832025,
1082
+ "grad_norm": 6.125,
1083
+ "learning_rate": 9.147138803923416e-05,
1084
+ "loss": 0.6593,
1085
+ "step": 1350
1086
+ },
1087
+ {
1088
+ "epoch": 1.0675039246467817,
1089
+ "grad_norm": 7.78125,
1090
+ "learning_rate": 9.127538592730519e-05,
1091
+ "loss": 0.7177,
1092
+ "step": 1360
1093
+ },
1094
+ {
1095
+ "epoch": 1.0753532182103611,
1096
+ "grad_norm": 4.59375,
1097
+ "learning_rate": 9.107737211129701e-05,
1098
+ "loss": 0.6882,
1099
+ "step": 1370
1100
+ },
1101
+ {
1102
+ "epoch": 1.0832025117739403,
1103
+ "grad_norm": 5.59375,
1104
+ "learning_rate": 9.087735624212365e-05,
1105
+ "loss": 0.5998,
1106
+ "step": 1380
1107
+ },
1108
+ {
1109
+ "epoch": 1.0910518053375196,
1110
+ "grad_norm": 5.21875,
1111
+ "learning_rate": 9.06753480682764e-05,
1112
+ "loss": 0.7154,
1113
+ "step": 1390
1114
+ },
1115
+ {
1116
+ "epoch": 1.098901098901099,
1117
+ "grad_norm": 6.28125,
1118
+ "learning_rate": 9.047135743534868e-05,
1119
+ "loss": 0.6295,
1120
+ "step": 1400
1121
+ },
1122
+ {
1123
+ "epoch": 1.1067503924646782,
1124
+ "grad_norm": 6.25,
1125
+ "learning_rate": 9.02653942855561e-05,
1126
+ "loss": 0.6978,
1127
+ "step": 1410
1128
+ },
1129
+ {
1130
+ "epoch": 1.1145996860282574,
1131
+ "grad_norm": 4.6875,
1132
+ "learning_rate": 9.005746865725207e-05,
1133
+ "loss": 0.6089,
1134
+ "step": 1420
1135
+ },
1136
+ {
1137
+ "epoch": 1.1224489795918366,
1138
+ "grad_norm": 7.78125,
1139
+ "learning_rate": 8.984759068443831e-05,
1140
+ "loss": 0.5945,
1141
+ "step": 1430
1142
+ },
1143
+ {
1144
+ "epoch": 1.130298273155416,
1145
+ "grad_norm": 7.34375,
1146
+ "learning_rate": 8.963577059627118e-05,
1147
+ "loss": 0.6959,
1148
+ "step": 1440
1149
+ },
1150
+ {
1151
+ "epoch": 1.1381475667189953,
1152
+ "grad_norm": 5.3125,
1153
+ "learning_rate": 8.942201871656291e-05,
1154
+ "loss": 0.6495,
1155
+ "step": 1450
1156
+ },
1157
+ {
1158
+ "epoch": 1.1459968602825745,
1159
+ "grad_norm": 5.84375,
1160
+ "learning_rate": 8.920634546327858e-05,
1161
+ "loss": 0.6512,
1162
+ "step": 1460
1163
+ },
1164
+ {
1165
+ "epoch": 1.1538461538461537,
1166
+ "grad_norm": 4.6875,
1167
+ "learning_rate": 8.898876134802826e-05,
1168
+ "loss": 0.6301,
1169
+ "step": 1470
1170
+ },
1171
+ {
1172
+ "epoch": 1.1616954474097332,
1173
+ "grad_norm": 7.0625,
1174
+ "learning_rate": 8.87692769755548e-05,
1175
+ "loss": 0.5665,
1176
+ "step": 1480
1177
+ },
1178
+ {
1179
+ "epoch": 1.1695447409733124,
1180
+ "grad_norm": 5.71875,
1181
+ "learning_rate": 8.854790304321681e-05,
1182
+ "loss": 0.6361,
1183
+ "step": 1490
1184
+ },
1185
+ {
1186
+ "epoch": 1.1773940345368916,
1187
+ "grad_norm": 5.75,
1188
+ "learning_rate": 8.832465034046749e-05,
1189
+ "loss": 0.6687,
1190
+ "step": 1500
1191
+ },
1192
+ {
1193
+ "epoch": 1.185243328100471,
1194
+ "grad_norm": 6.3125,
1195
+ "learning_rate": 8.80995297483286e-05,
1196
+ "loss": 0.6106,
1197
+ "step": 1510
1198
+ },
1199
+ {
1200
+ "epoch": 1.1930926216640503,
1201
+ "grad_norm": 6.53125,
1202
+ "learning_rate": 8.787255223886019e-05,
1203
+ "loss": 0.6526,
1204
+ "step": 1520
1205
+ },
1206
+ {
1207
+ "epoch": 1.2009419152276295,
1208
+ "grad_norm": 5.6875,
1209
+ "learning_rate": 8.764372887462586e-05,
1210
+ "loss": 0.6845,
1211
+ "step": 1530
1212
+ },
1213
+ {
1214
+ "epoch": 1.2087912087912087,
1215
+ "grad_norm": 5.78125,
1216
+ "learning_rate": 8.741307080815357e-05,
1217
+ "loss": 0.6695,
1218
+ "step": 1540
1219
+ },
1220
+ {
1221
+ "epoch": 1.2166405023547882,
1222
+ "grad_norm": 4.15625,
1223
+ "learning_rate": 8.718058928139205e-05,
1224
+ "loss": 0.6021,
1225
+ "step": 1550
1226
+ },
1227
+ {
1228
+ "epoch": 1.2244897959183674,
1229
+ "grad_norm": 5.5625,
1230
+ "learning_rate": 8.694629562516294e-05,
1231
+ "loss": 0.6173,
1232
+ "step": 1560
1233
+ },
1234
+ {
1235
+ "epoch": 1.2323390894819466,
1236
+ "grad_norm": 4.125,
1237
+ "learning_rate": 8.671020125860851e-05,
1238
+ "loss": 0.5407,
1239
+ "step": 1570
1240
+ },
1241
+ {
1242
+ "epoch": 1.2401883830455258,
1243
+ "grad_norm": 6.28125,
1244
+ "learning_rate": 8.647231768863512e-05,
1245
+ "loss": 0.6177,
1246
+ "step": 1580
1247
+ },
1248
+ {
1249
+ "epoch": 1.2480376766091053,
1250
+ "grad_norm": 5.03125,
1251
+ "learning_rate": 8.623265650935234e-05,
1252
+ "loss": 0.6132,
1253
+ "step": 1590
1254
+ },
1255
+ {
1256
+ "epoch": 1.2558869701726845,
1257
+ "grad_norm": 5.9375,
1258
+ "learning_rate": 8.599122940150795e-05,
1259
+ "loss": 0.5913,
1260
+ "step": 1600
1261
+ },
1262
+ {
1263
+ "epoch": 1.2637362637362637,
1264
+ "grad_norm": 6.28125,
1265
+ "learning_rate": 8.574804813191858e-05,
1266
+ "loss": 0.7089,
1267
+ "step": 1610
1268
+ },
1269
+ {
1270
+ "epoch": 1.2715855572998431,
1271
+ "grad_norm": 7.3125,
1272
+ "learning_rate": 8.550312455289625e-05,
1273
+ "loss": 0.5426,
1274
+ "step": 1620
1275
+ },
1276
+ {
1277
+ "epoch": 1.2794348508634223,
1278
+ "grad_norm": 4.59375,
1279
+ "learning_rate": 8.525647060167063e-05,
1280
+ "loss": 0.5678,
1281
+ "step": 1630
1282
+ },
1283
+ {
1284
+ "epoch": 1.2872841444270016,
1285
+ "grad_norm": 6.34375,
1286
+ "learning_rate": 8.500809829980734e-05,
1287
+ "loss": 0.6343,
1288
+ "step": 1640
1289
+ },
1290
+ {
1291
+ "epoch": 1.2951334379905808,
1292
+ "grad_norm": 6.5625,
1293
+ "learning_rate": 8.4758019752622e-05,
1294
+ "loss": 0.7523,
1295
+ "step": 1650
1296
+ },
1297
+ {
1298
+ "epoch": 1.30298273155416,
1299
+ "grad_norm": 4.5,
1300
+ "learning_rate": 8.450624714859016e-05,
1301
+ "loss": 0.699,
1302
+ "step": 1660
1303
+ },
1304
+ {
1305
+ "epoch": 1.3108320251177394,
1306
+ "grad_norm": 5.09375,
1307
+ "learning_rate": 8.425279275875336e-05,
1308
+ "loss": 0.6517,
1309
+ "step": 1670
1310
+ },
1311
+ {
1312
+ "epoch": 1.3186813186813187,
1313
+ "grad_norm": 4.21875,
1314
+ "learning_rate": 8.399766893612096e-05,
1315
+ "loss": 0.563,
1316
+ "step": 1680
1317
+ },
1318
+ {
1319
+ "epoch": 1.3265306122448979,
1320
+ "grad_norm": 5.5625,
1321
+ "learning_rate": 8.374088811506819e-05,
1322
+ "loss": 0.6381,
1323
+ "step": 1690
1324
+ },
1325
+ {
1326
+ "epoch": 1.3343799058084773,
1327
+ "grad_norm": 4.6875,
1328
+ "learning_rate": 8.348246281072998e-05,
1329
+ "loss": 0.6794,
1330
+ "step": 1700
1331
+ },
1332
+ {
1333
+ "epoch": 1.3422291993720565,
1334
+ "grad_norm": 5.1875,
1335
+ "learning_rate": 8.322240561839109e-05,
1336
+ "loss": 0.6494,
1337
+ "step": 1710
1338
+ },
1339
+ {
1340
+ "epoch": 1.3500784929356358,
1341
+ "grad_norm": 5.09375,
1342
+ "learning_rate": 8.296072921287217e-05,
1343
+ "loss": 0.587,
1344
+ "step": 1720
1345
+ },
1346
+ {
1347
+ "epoch": 1.3579277864992152,
1348
+ "grad_norm": 6.34375,
1349
+ "learning_rate": 8.269744634791208e-05,
1350
+ "loss": 0.6888,
1351
+ "step": 1730
1352
+ },
1353
+ {
1354
+ "epoch": 1.3657770800627944,
1355
+ "grad_norm": 4.40625,
1356
+ "learning_rate": 8.243256985554621e-05,
1357
+ "loss": 0.6629,
1358
+ "step": 1740
1359
+ },
1360
+ {
1361
+ "epoch": 1.3736263736263736,
1362
+ "grad_norm": 5.4375,
1363
+ "learning_rate": 8.21661126454811e-05,
1364
+ "loss": 0.6633,
1365
+ "step": 1750
1366
+ },
1367
+ {
1368
+ "epoch": 1.3814756671899528,
1369
+ "grad_norm": 5.03125,
1370
+ "learning_rate": 8.189808770446528e-05,
1371
+ "loss": 0.633,
1372
+ "step": 1760
1373
+ },
1374
+ {
1375
+ "epoch": 1.389324960753532,
1376
+ "grad_norm": 4.59375,
1377
+ "learning_rate": 8.162850809565623e-05,
1378
+ "loss": 0.6725,
1379
+ "step": 1770
1380
+ },
1381
+ {
1382
+ "epoch": 1.3971742543171115,
1383
+ "grad_norm": 5.71875,
1384
+ "learning_rate": 8.135738695798376e-05,
1385
+ "loss": 0.6739,
1386
+ "step": 1780
1387
+ },
1388
+ {
1389
+ "epoch": 1.4050235478806907,
1390
+ "grad_norm": 5.34375,
1391
+ "learning_rate": 8.108473750550965e-05,
1392
+ "loss": 0.6547,
1393
+ "step": 1790
1394
+ },
1395
+ {
1396
+ "epoch": 1.41287284144427,
1397
+ "grad_norm": 6.0,
1398
+ "learning_rate": 8.081057302678352e-05,
1399
+ "loss": 0.5921,
1400
+ "step": 1800
1401
+ },
1402
+ {
1403
+ "epoch": 1.4207221350078494,
1404
+ "grad_norm": 6.5625,
1405
+ "learning_rate": 8.053490688419533e-05,
1406
+ "loss": 0.6639,
1407
+ "step": 1810
1408
+ },
1409
+ {
1410
+ "epoch": 1.4285714285714286,
1411
+ "grad_norm": 6.90625,
1412
+ "learning_rate": 8.02577525133239e-05,
1413
+ "loss": 0.6445,
1414
+ "step": 1820
1415
+ },
1416
+ {
1417
+ "epoch": 1.4364207221350078,
1418
+ "grad_norm": 6.375,
1419
+ "learning_rate": 7.997912342228232e-05,
1420
+ "loss": 0.6453,
1421
+ "step": 1830
1422
+ },
1423
+ {
1424
+ "epoch": 1.4442700156985873,
1425
+ "grad_norm": 6.34375,
1426
+ "learning_rate": 7.969903319105935e-05,
1427
+ "loss": 0.6649,
1428
+ "step": 1840
1429
+ },
1430
+ {
1431
+ "epoch": 1.4521193092621665,
1432
+ "grad_norm": 5.875,
1433
+ "learning_rate": 7.941749547085777e-05,
1434
+ "loss": 0.7093,
1435
+ "step": 1850
1436
+ },
1437
+ {
1438
+ "epoch": 1.4599686028257457,
1439
+ "grad_norm": 5.71875,
1440
+ "learning_rate": 7.913452398342881e-05,
1441
+ "loss": 0.5864,
1442
+ "step": 1860
1443
+ },
1444
+ {
1445
+ "epoch": 1.467817896389325,
1446
+ "grad_norm": 4.875,
1447
+ "learning_rate": 7.885013252040359e-05,
1448
+ "loss": 0.6377,
1449
+ "step": 1870
1450
+ },
1451
+ {
1452
+ "epoch": 1.4756671899529041,
1453
+ "grad_norm": 3.625,
1454
+ "learning_rate": 7.856433494262077e-05,
1455
+ "loss": 0.6205,
1456
+ "step": 1880
1457
+ },
1458
+ {
1459
+ "epoch": 1.4835164835164836,
1460
+ "grad_norm": 5.15625,
1461
+ "learning_rate": 7.827714517945115e-05,
1462
+ "loss": 0.6009,
1463
+ "step": 1890
1464
+ },
1465
+ {
1466
+ "epoch": 1.4913657770800628,
1467
+ "grad_norm": 3.890625,
1468
+ "learning_rate": 7.798857722811856e-05,
1469
+ "loss": 0.5702,
1470
+ "step": 1900
1471
+ },
1472
+ {
1473
+ "epoch": 1.499215070643642,
1474
+ "grad_norm": 5.03125,
1475
+ "learning_rate": 7.769864515301787e-05,
1476
+ "loss": 0.5561,
1477
+ "step": 1910
1478
+ },
1479
+ {
1480
+ "epoch": 1.5070643642072215,
1481
+ "grad_norm": 6.0625,
1482
+ "learning_rate": 7.740736308502938e-05,
1483
+ "loss": 0.6446,
1484
+ "step": 1920
1485
+ },
1486
+ {
1487
+ "epoch": 1.5149136577708007,
1488
+ "grad_norm": 6.34375,
1489
+ "learning_rate": 7.711474522083015e-05,
1490
+ "loss": 0.5983,
1491
+ "step": 1930
1492
+ },
1493
+ {
1494
+ "epoch": 1.5227629513343799,
1495
+ "grad_norm": 6.1875,
1496
+ "learning_rate": 7.682080582220206e-05,
1497
+ "loss": 0.5358,
1498
+ "step": 1940
1499
+ },
1500
+ {
1501
+ "epoch": 1.5306122448979593,
1502
+ "grad_norm": 6.5,
1503
+ "learning_rate": 7.65255592153367e-05,
1504
+ "loss": 0.7394,
1505
+ "step": 1950
1506
+ },
1507
+ {
1508
+ "epoch": 1.5384615384615383,
1509
+ "grad_norm": 4.8125,
1510
+ "learning_rate": 7.622901979013717e-05,
1511
+ "loss": 0.5742,
1512
+ "step": 1960
1513
+ },
1514
+ {
1515
+ "epoch": 1.5463108320251178,
1516
+ "grad_norm": 4.5,
1517
+ "learning_rate": 7.593120199951671e-05,
1518
+ "loss": 0.6425,
1519
+ "step": 1970
1520
+ },
1521
+ {
1522
+ "epoch": 1.554160125588697,
1523
+ "grad_norm": 6.03125,
1524
+ "learning_rate": 7.563212035869425e-05,
1525
+ "loss": 0.6593,
1526
+ "step": 1980
1527
+ },
1528
+ {
1529
+ "epoch": 1.5620094191522762,
1530
+ "grad_norm": 7.03125,
1531
+ "learning_rate": 7.533178944448705e-05,
1532
+ "loss": 0.5948,
1533
+ "step": 1990
1534
+ },
1535
+ {
1536
+ "epoch": 1.5698587127158556,
1537
+ "grad_norm": 6.0,
1538
+ "learning_rate": 7.503022389460014e-05,
1539
+ "loss": 0.625,
1540
+ "step": 2000
1541
+ },
1542
+ {
1543
+ "epoch": 1.5698587127158556,
1544
+ "eval/nano_beir.arguana": 0.47093852833484745,
1545
+ "step": 2000
1546
+ },
1547
+ {
1548
+ "epoch": 1.5698587127158556,
1549
+ "eval/nano_beir.climate_fever": 0.25927619589572015,
1550
+ "step": 2000
1551
+ },
1552
+ {
1553
+ "epoch": 1.5698587127158556,
1554
+ "eval/nano_beir.dbpedia_entity": 0.3266259450992667,
1555
+ "step": 2000
1556
+ },
1557
+ {
1558
+ "epoch": 1.5698587127158556,
1559
+ "eval/nano_beir.fever": 0.6291710052755145,
1560
+ "step": 2000
1561
+ },
1562
+ {
1563
+ "epoch": 1.5698587127158556,
1564
+ "eval/nano_beir.fiqa": 0.4343483963906797,
1565
+ "step": 2000
1566
+ },
1567
+ {
1568
+ "epoch": 1.5698587127158556,
1569
+ "eval/nano_beir.hotpotqa": 0.6854629804476208,
1570
+ "step": 2000
1571
+ },
1572
+ {
1573
+ "epoch": 1.5698587127158556,
1574
+ "eval/nano_beir.nfcorpus": 0.34320335172563987,
1575
+ "step": 2000
1576
+ },
1577
+ {
1578
+ "epoch": 1.5698587127158556,
1579
+ "eval/nano_beir.nq": 0.40466132495464113,
1580
+ "step": 2000
1581
+ },
1582
+ {
1583
+ "epoch": 1.5698587127158556,
1584
+ "eval/nano_beir.quora": 0.8346175392008862,
1585
+ "step": 2000
1586
+ },
1587
+ {
1588
+ "epoch": 1.5698587127158556,
1589
+ "eval/nano_beir.scidocs": 0.2536638155561289,
1590
+ "step": 2000
1591
+ },
1592
+ {
1593
+ "epoch": 1.5698587127158556,
1594
+ "eval/nano_beir.scifact": 0.756036525808922,
1595
+ "step": 2000
1596
+ },
1597
+ {
1598
+ "epoch": 1.5698587127158556,
1599
+ "eval/nano_beir.webis_touche2020": 0.282061170550218,
1600
+ "step": 2000
1601
+ },
1602
+ {
1603
+ "epoch": 1.5698587127158556,
1604
+ "eval/avg": 0.4733388982700071,
1605
+ "step": 2000
1606
+ },
1607
+ {
1608
+ "epoch": 1.5777080062794349,
1609
+ "grad_norm": 5.1875,
1610
+ "learning_rate": 7.472743840691299e-05,
1611
+ "loss": 0.6136,
1612
+ "step": 2010
1613
+ },
1614
+ {
1615
+ "epoch": 1.585557299843014,
1616
+ "grad_norm": 4.53125,
1617
+ "learning_rate": 7.44234477387631e-05,
1618
+ "loss": 0.5169,
1619
+ "step": 2020
1620
+ },
1621
+ {
1622
+ "epoch": 1.5934065934065935,
1623
+ "grad_norm": 4.0625,
1624
+ "learning_rate": 7.411826670622676e-05,
1625
+ "loss": 0.6241,
1626
+ "step": 2030
1627
+ },
1628
+ {
1629
+ "epoch": 1.6012558869701727,
1630
+ "grad_norm": 5.5,
1631
+ "learning_rate": 7.381191018339696e-05,
1632
+ "loss": 0.6384,
1633
+ "step": 2040
1634
+ },
1635
+ {
1636
+ "epoch": 1.609105180533752,
1637
+ "grad_norm": 4.78125,
1638
+ "learning_rate": 7.350439310165842e-05,
1639
+ "loss": 0.687,
1640
+ "step": 2050
1641
+ },
1642
+ {
1643
+ "epoch": 1.6169544740973314,
1644
+ "grad_norm": 5.84375,
1645
+ "learning_rate": 7.319573044895986e-05,
1646
+ "loss": 0.6424,
1647
+ "step": 2060
1648
+ },
1649
+ {
1650
+ "epoch": 1.6248037676609104,
1651
+ "grad_norm": 6.34375,
1652
+ "learning_rate": 7.28859372690835e-05,
1653
+ "loss": 0.6997,
1654
+ "step": 2070
1655
+ },
1656
+ {
1657
+ "epoch": 1.6326530612244898,
1658
+ "grad_norm": 5.59375,
1659
+ "learning_rate": 7.257502866091192e-05,
1660
+ "loss": 0.5591,
1661
+ "step": 2080
1662
+ },
1663
+ {
1664
+ "epoch": 1.640502354788069,
1665
+ "grad_norm": 5.75,
1666
+ "learning_rate": 7.226301977769199e-05,
1667
+ "loss": 0.5663,
1668
+ "step": 2090
1669
+ },
1670
+ {
1671
+ "epoch": 1.6483516483516483,
1672
+ "grad_norm": 6.15625,
1673
+ "learning_rate": 7.194992582629654e-05,
1674
+ "loss": 0.6399,
1675
+ "step": 2100
1676
+ },
1677
+ {
1678
+ "epoch": 1.6562009419152277,
1679
+ "grad_norm": 6.5625,
1680
+ "learning_rate": 7.163576206648304e-05,
1681
+ "loss": 0.6225,
1682
+ "step": 2110
1683
+ },
1684
+ {
1685
+ "epoch": 1.664050235478807,
1686
+ "grad_norm": 6.28125,
1687
+ "learning_rate": 7.132054381014995e-05,
1688
+ "loss": 0.6461,
1689
+ "step": 2120
1690
+ },
1691
+ {
1692
+ "epoch": 1.6718995290423861,
1693
+ "grad_norm": 5.5625,
1694
+ "learning_rate": 7.100428642059033e-05,
1695
+ "loss": 0.6079,
1696
+ "step": 2130
1697
+ },
1698
+ {
1699
+ "epoch": 1.6797488226059656,
1700
+ "grad_norm": 5.8125,
1701
+ "learning_rate": 7.068700531174319e-05,
1702
+ "loss": 0.6539,
1703
+ "step": 2140
1704
+ },
1705
+ {
1706
+ "epoch": 1.6875981161695446,
1707
+ "grad_norm": 6.0,
1708
+ "learning_rate": 7.036871594744217e-05,
1709
+ "loss": 0.6949,
1710
+ "step": 2150
1711
+ },
1712
+ {
1713
+ "epoch": 1.695447409733124,
1714
+ "grad_norm": 5.34375,
1715
+ "learning_rate": 7.004943384066187e-05,
1716
+ "loss": 0.5566,
1717
+ "step": 2160
1718
+ },
1719
+ {
1720
+ "epoch": 1.7032967032967035,
1721
+ "grad_norm": 5.71875,
1722
+ "learning_rate": 6.97291745527617e-05,
1723
+ "loss": 0.691,
1724
+ "step": 2170
1725
+ },
1726
+ {
1727
+ "epoch": 1.7111459968602825,
1728
+ "grad_norm": 5.21875,
1729
+ "learning_rate": 6.940795369272754e-05,
1730
+ "loss": 0.7199,
1731
+ "step": 2180
1732
+ },
1733
+ {
1734
+ "epoch": 1.718995290423862,
1735
+ "grad_norm": 5.1875,
1736
+ "learning_rate": 6.908578691641092e-05,
1737
+ "loss": 0.6165,
1738
+ "step": 2190
1739
+ },
1740
+ {
1741
+ "epoch": 1.7268445839874411,
1742
+ "grad_norm": 5.0625,
1743
+ "learning_rate": 6.876268992576604e-05,
1744
+ "loss": 0.598,
1745
+ "step": 2200
1746
+ },
1747
+ {
1748
+ "epoch": 1.7346938775510203,
1749
+ "grad_norm": 5.125,
1750
+ "learning_rate": 6.843867846808438e-05,
1751
+ "loss": 0.5891,
1752
+ "step": 2210
1753
+ },
1754
+ {
1755
+ "epoch": 1.7425431711145998,
1756
+ "grad_norm": 5.90625,
1757
+ "learning_rate": 6.811376833522729e-05,
1758
+ "loss": 0.6333,
1759
+ "step": 2220
1760
+ },
1761
+ {
1762
+ "epoch": 1.750392464678179,
1763
+ "grad_norm": 5.59375,
1764
+ "learning_rate": 6.778797536285624e-05,
1765
+ "loss": 0.5931,
1766
+ "step": 2230
1767
+ },
1768
+ {
1769
+ "epoch": 1.7582417582417582,
1770
+ "grad_norm": 5.15625,
1771
+ "learning_rate": 6.746131542966112e-05,
1772
+ "loss": 0.6326,
1773
+ "step": 2240
1774
+ },
1775
+ {
1776
+ "epoch": 1.7660910518053377,
1777
+ "grad_norm": 5.625,
1778
+ "learning_rate": 6.713380445658618e-05,
1779
+ "loss": 0.6492,
1780
+ "step": 2250
1781
+ },
1782
+ {
1783
+ "epoch": 1.7739403453689166,
1784
+ "grad_norm": 7.0625,
1785
+ "learning_rate": 6.680545840605423e-05,
1786
+ "loss": 0.5642,
1787
+ "step": 2260
1788
+ },
1789
+ {
1790
+ "epoch": 1.781789638932496,
1791
+ "grad_norm": 5.6875,
1792
+ "learning_rate": 6.647629328118851e-05,
1793
+ "loss": 0.5615,
1794
+ "step": 2270
1795
+ },
1796
+ {
1797
+ "epoch": 1.7896389324960753,
1798
+ "grad_norm": 4.625,
1799
+ "learning_rate": 6.614632512503288e-05,
1800
+ "loss": 0.5933,
1801
+ "step": 2280
1802
+ },
1803
+ {
1804
+ "epoch": 1.7974882260596545,
1805
+ "grad_norm": 6.125,
1806
+ "learning_rate": 6.58155700197697e-05,
1807
+ "loss": 0.6289,
1808
+ "step": 2290
1809
+ },
1810
+ {
1811
+ "epoch": 1.805337519623234,
1812
+ "grad_norm": 5.21875,
1813
+ "learning_rate": 6.548404408593621e-05,
1814
+ "loss": 0.6227,
1815
+ "step": 2300
1816
+ },
1817
+ {
1818
+ "epoch": 1.8131868131868132,
1819
+ "grad_norm": 5.75,
1820
+ "learning_rate": 6.515176348163871e-05,
1821
+ "loss": 0.5941,
1822
+ "step": 2310
1823
+ },
1824
+ {
1825
+ "epoch": 1.8210361067503924,
1826
+ "grad_norm": 5.59375,
1827
+ "learning_rate": 6.481874440176506e-05,
1828
+ "loss": 0.5882,
1829
+ "step": 2320
1830
+ },
1831
+ {
1832
+ "epoch": 1.8288854003139718,
1833
+ "grad_norm": 5.6875,
1834
+ "learning_rate": 6.448500307719538e-05,
1835
+ "loss": 0.6208,
1836
+ "step": 2330
1837
+ },
1838
+ {
1839
+ "epoch": 1.836734693877551,
1840
+ "grad_norm": 4.5,
1841
+ "learning_rate": 6.415055577401102e-05,
1842
+ "loss": 0.5798,
1843
+ "step": 2340
1844
+ },
1845
+ {
1846
+ "epoch": 1.8445839874411303,
1847
+ "grad_norm": 3.9375,
1848
+ "learning_rate": 6.381541879270169e-05,
1849
+ "loss": 0.6352,
1850
+ "step": 2350
1851
+ },
1852
+ {
1853
+ "epoch": 1.8524332810047097,
1854
+ "grad_norm": 5.5,
1855
+ "learning_rate": 6.347960846737106e-05,
1856
+ "loss": 0.7087,
1857
+ "step": 2360
1858
+ },
1859
+ {
1860
+ "epoch": 1.8602825745682887,
1861
+ "grad_norm": 4.625,
1862
+ "learning_rate": 6.31431411649406e-05,
1863
+ "loss": 0.5722,
1864
+ "step": 2370
1865
+ },
1866
+ {
1867
+ "epoch": 1.8681318681318682,
1868
+ "grad_norm": 5.625,
1869
+ "learning_rate": 6.280603328435198e-05,
1870
+ "loss": 0.6634,
1871
+ "step": 2380
1872
+ },
1873
+ {
1874
+ "epoch": 1.8759811616954474,
1875
+ "grad_norm": 5.78125,
1876
+ "learning_rate": 6.24683012557677e-05,
1877
+ "loss": 0.6851,
1878
+ "step": 2390
1879
+ },
1880
+ {
1881
+ "epoch": 1.8838304552590266,
1882
+ "grad_norm": 5.875,
1883
+ "learning_rate": 6.212996153977037e-05,
1884
+ "loss": 0.6613,
1885
+ "step": 2400
1886
+ },
1887
+ {
1888
+ "epoch": 1.891679748822606,
1889
+ "grad_norm": 4.3125,
1890
+ "learning_rate": 6.179103062656042e-05,
1891
+ "loss": 0.6541,
1892
+ "step": 2410
1893
+ },
1894
+ {
1895
+ "epoch": 1.8995290423861853,
1896
+ "grad_norm": 7.34375,
1897
+ "learning_rate": 6.145152503515239e-05,
1898
+ "loss": 0.714,
1899
+ "step": 2420
1900
+ },
1901
+ {
1902
+ "epoch": 1.9073783359497645,
1903
+ "grad_norm": 5.25,
1904
+ "learning_rate": 6.111146131256983e-05,
1905
+ "loss": 0.614,
1906
+ "step": 2430
1907
+ },
1908
+ {
1909
+ "epoch": 1.915227629513344,
1910
+ "grad_norm": 5.0625,
1911
+ "learning_rate": 6.077085603303882e-05,
1912
+ "loss": 0.5535,
1913
+ "step": 2440
1914
+ },
1915
+ {
1916
+ "epoch": 1.9230769230769231,
1917
+ "grad_norm": 6.40625,
1918
+ "learning_rate": 6.042972579718019e-05,
1919
+ "loss": 0.7302,
1920
+ "step": 2450
1921
+ },
1922
+ {
1923
+ "epoch": 1.9309262166405023,
1924
+ "grad_norm": 4.5,
1925
+ "learning_rate": 6.008808723120035e-05,
1926
+ "loss": 0.6042,
1927
+ "step": 2460
1928
+ },
1929
+ {
1930
+ "epoch": 1.9387755102040818,
1931
+ "grad_norm": 5.8125,
1932
+ "learning_rate": 5.974595698608103e-05,
1933
+ "loss": 0.603,
1934
+ "step": 2470
1935
+ },
1936
+ {
1937
+ "epoch": 1.9466248037676608,
1938
+ "grad_norm": 5.1875,
1939
+ "learning_rate": 5.9403351736767696e-05,
1940
+ "loss": 0.668,
1941
+ "step": 2480
1942
+ },
1943
+ {
1944
+ "epoch": 1.9544740973312402,
1945
+ "grad_norm": 5.46875,
1946
+ "learning_rate": 5.906028818135687e-05,
1947
+ "loss": 0.6311,
1948
+ "step": 2490
1949
+ },
1950
+ {
1951
+ "epoch": 1.9623233908948194,
1952
+ "grad_norm": 5.40625,
1953
+ "learning_rate": 5.8716783040282244e-05,
1954
+ "loss": 0.5565,
1955
+ "step": 2500
1956
+ },
1957
+ {
1958
+ "epoch": 1.9701726844583987,
1959
+ "grad_norm": 4.875,
1960
+ "learning_rate": 5.837285305549978e-05,
1961
+ "loss": 0.6793,
1962
+ "step": 2510
1963
+ },
1964
+ {
1965
+ "epoch": 1.978021978021978,
1966
+ "grad_norm": 6.6875,
1967
+ "learning_rate": 5.8028514989671724e-05,
1968
+ "loss": 0.6645,
1969
+ "step": 2520
1970
+ },
1971
+ {
1972
+ "epoch": 1.9858712715855573,
1973
+ "grad_norm": 5.65625,
1974
+ "learning_rate": 5.7683785625349616e-05,
1975
+ "loss": 0.5526,
1976
+ "step": 2530
1977
+ },
1978
+ {
1979
+ "epoch": 1.9937205651491365,
1980
+ "grad_norm": 6.59375,
1981
+ "learning_rate": 5.733868176415633e-05,
1982
+ "loss": 0.6974,
1983
+ "step": 2540
1984
+ },
1985
+ {
1986
+ "epoch": 2.001569858712716,
1987
+ "grad_norm": 4.875,
1988
+ "learning_rate": 5.699322022596722e-05,
1989
+ "loss": 0.7787,
1990
+ "step": 2550
1991
+ },
1992
+ {
1993
+ "epoch": 2.009419152276295,
1994
+ "grad_norm": 5.03125,
1995
+ "learning_rate": 5.664741784809022e-05,
1996
+ "loss": 0.6121,
1997
+ "step": 2560
1998
+ },
1999
+ {
2000
+ "epoch": 2.0172684458398744,
2001
+ "grad_norm": 4.4375,
2002
+ "learning_rate": 5.630129148444543e-05,
2003
+ "loss": 0.5956,
2004
+ "step": 2570
2005
+ },
2006
+ {
2007
+ "epoch": 2.025117739403454,
2008
+ "grad_norm": 4.4375,
2009
+ "learning_rate": 5.595485800474349e-05,
2010
+ "loss": 0.6243,
2011
+ "step": 2580
2012
+ },
2013
+ {
2014
+ "epoch": 2.032967032967033,
2015
+ "grad_norm": 6.625,
2016
+ "learning_rate": 5.560813429366345e-05,
2017
+ "loss": 0.5992,
2018
+ "step": 2590
2019
+ },
2020
+ {
2021
+ "epoch": 2.0408163265306123,
2022
+ "grad_norm": 7.03125,
2023
+ "learning_rate": 5.5261137250029835e-05,
2024
+ "loss": 0.5197,
2025
+ "step": 2600
2026
+ },
2027
+ {
2028
+ "epoch": 2.0486656200941917,
2029
+ "grad_norm": 4.78125,
2030
+ "learning_rate": 5.4913883785988993e-05,
2031
+ "loss": 0.5492,
2032
+ "step": 2610
2033
+ },
2034
+ {
2035
+ "epoch": 2.0565149136577707,
2036
+ "grad_norm": 5.8125,
2037
+ "learning_rate": 5.456639082618489e-05,
2038
+ "loss": 0.5624,
2039
+ "step": 2620
2040
+ },
2041
+ {
2042
+ "epoch": 2.06436420722135,
2043
+ "grad_norm": 5.90625,
2044
+ "learning_rate": 5.421867530693414e-05,
2045
+ "loss": 0.5266,
2046
+ "step": 2630
2047
+ },
2048
+ {
2049
+ "epoch": 2.072213500784929,
2050
+ "grad_norm": 5.0,
2051
+ "learning_rate": 5.3870754175400595e-05,
2052
+ "loss": 0.5205,
2053
+ "step": 2640
2054
+ },
2055
+ {
2056
+ "epoch": 2.0800627943485086,
2057
+ "grad_norm": 5.59375,
2058
+ "learning_rate": 5.352264438876935e-05,
2059
+ "loss": 0.6017,
2060
+ "step": 2650
2061
+ },
2062
+ {
2063
+ "epoch": 2.087912087912088,
2064
+ "grad_norm": 5.125,
2065
+ "learning_rate": 5.3174362913420306e-05,
2066
+ "loss": 0.565,
2067
+ "step": 2660
2068
+ },
2069
+ {
2070
+ "epoch": 2.095761381475667,
2071
+ "grad_norm": 4.875,
2072
+ "learning_rate": 5.2825926724101236e-05,
2073
+ "loss": 0.6248,
2074
+ "step": 2670
2075
+ },
2076
+ {
2077
+ "epoch": 2.1036106750392465,
2078
+ "grad_norm": 6.09375,
2079
+ "learning_rate": 5.247735280310041e-05,
2080
+ "loss": 0.7087,
2081
+ "step": 2680
2082
+ },
2083
+ {
2084
+ "epoch": 2.111459968602826,
2085
+ "grad_norm": 5.25,
2086
+ "learning_rate": 5.212865813941899e-05,
2087
+ "loss": 0.4872,
2088
+ "step": 2690
2089
+ },
2090
+ {
2091
+ "epoch": 2.119309262166405,
2092
+ "grad_norm": 6.0,
2093
+ "learning_rate": 5.1779859727942924e-05,
2094
+ "loss": 0.6128,
2095
+ "step": 2700
2096
+ },
2097
+ {
2098
+ "epoch": 2.1271585557299844,
2099
+ "grad_norm": 4.9375,
2100
+ "learning_rate": 5.143097456861474e-05,
2101
+ "loss": 0.5494,
2102
+ "step": 2710
2103
+ },
2104
+ {
2105
+ "epoch": 2.1350078492935634,
2106
+ "grad_norm": 5.5625,
2107
+ "learning_rate": 5.1082019665604895e-05,
2108
+ "loss": 0.5941,
2109
+ "step": 2720
2110
+ },
2111
+ {
2112
+ "epoch": 2.142857142857143,
2113
+ "grad_norm": 6.0625,
2114
+ "learning_rate": 5.073301202648304e-05,
2115
+ "loss": 0.5385,
2116
+ "step": 2730
2117
+ },
2118
+ {
2119
+ "epoch": 2.1507064364207222,
2120
+ "grad_norm": 7.25,
2121
+ "learning_rate": 5.0383968661389146e-05,
2122
+ "loss": 0.5394,
2123
+ "step": 2740
2124
+ },
2125
+ {
2126
+ "epoch": 2.1585557299843012,
2127
+ "grad_norm": 7.59375,
2128
+ "learning_rate": 5.003490658220438e-05,
2129
+ "loss": 0.5874,
2130
+ "step": 2750
2131
+ },
2132
+ {
2133
+ "epoch": 2.1664050235478807,
2134
+ "grad_norm": 5.15625,
2135
+ "learning_rate": 4.968584280172206e-05,
2136
+ "loss": 0.5373,
2137
+ "step": 2760
2138
+ },
2139
+ {
2140
+ "epoch": 2.17425431711146,
2141
+ "grad_norm": 6.65625,
2142
+ "learning_rate": 4.933679433281836e-05,
2143
+ "loss": 0.5698,
2144
+ "step": 2770
2145
+ },
2146
+ {
2147
+ "epoch": 2.182103610675039,
2148
+ "grad_norm": 6.25,
2149
+ "learning_rate": 4.8987778187623245e-05,
2150
+ "loss": 0.6354,
2151
+ "step": 2780
2152
+ },
2153
+ {
2154
+ "epoch": 2.1899529042386185,
2155
+ "grad_norm": 6.6875,
2156
+ "learning_rate": 4.863881137669123e-05,
2157
+ "loss": 0.5519,
2158
+ "step": 2790
2159
+ },
2160
+ {
2161
+ "epoch": 2.197802197802198,
2162
+ "grad_norm": 5.59375,
2163
+ "learning_rate": 4.8289910908172376e-05,
2164
+ "loss": 0.6345,
2165
+ "step": 2800
2166
+ },
2167
+ {
2168
+ "epoch": 2.205651491365777,
2169
+ "grad_norm": 5.5,
2170
+ "learning_rate": 4.794109378698327e-05,
2171
+ "loss": 0.5927,
2172
+ "step": 2810
2173
+ },
2174
+ {
2175
+ "epoch": 2.2135007849293564,
2176
+ "grad_norm": 5.71875,
2177
+ "learning_rate": 4.7592377013978306e-05,
2178
+ "loss": 0.5696,
2179
+ "step": 2820
2180
+ },
2181
+ {
2182
+ "epoch": 2.221350078492936,
2183
+ "grad_norm": 5.21875,
2184
+ "learning_rate": 4.724377758512103e-05,
2185
+ "loss": 0.5136,
2186
+ "step": 2830
2187
+ },
2188
+ {
2189
+ "epoch": 2.229199372056515,
2190
+ "grad_norm": 5.75,
2191
+ "learning_rate": 4.68953124906558e-05,
2192
+ "loss": 0.6424,
2193
+ "step": 2840
2194
+ },
2195
+ {
2196
+ "epoch": 2.2370486656200943,
2197
+ "grad_norm": 5.75,
2198
+ "learning_rate": 4.654699871427971e-05,
2199
+ "loss": 0.6576,
2200
+ "step": 2850
2201
+ },
2202
+ {
2203
+ "epoch": 2.2448979591836733,
2204
+ "grad_norm": 7.34375,
2205
+ "learning_rate": 4.619885323231484e-05,
2206
+ "loss": 0.5978,
2207
+ "step": 2860
2208
+ },
2209
+ {
2210
+ "epoch": 2.2527472527472527,
2211
+ "grad_norm": 6.25,
2212
+ "learning_rate": 4.58508930128808e-05,
2213
+ "loss": 0.5672,
2214
+ "step": 2870
2215
+ },
2216
+ {
2217
+ "epoch": 2.260596546310832,
2218
+ "grad_norm": 5.90625,
2219
+ "learning_rate": 4.550313501506781e-05,
2220
+ "loss": 0.6515,
2221
+ "step": 2880
2222
+ },
2223
+ {
2224
+ "epoch": 2.268445839874411,
2225
+ "grad_norm": 5.34375,
2226
+ "learning_rate": 4.515559618811005e-05,
2227
+ "loss": 0.4906,
2228
+ "step": 2890
2229
+ },
2230
+ {
2231
+ "epoch": 2.2762951334379906,
2232
+ "grad_norm": 6.59375,
2233
+ "learning_rate": 4.4808293470559643e-05,
2234
+ "loss": 0.5254,
2235
+ "step": 2900
2236
+ },
2237
+ {
2238
+ "epoch": 2.2841444270015696,
2239
+ "grad_norm": 5.75,
2240
+ "learning_rate": 4.446124378946107e-05,
2241
+ "loss": 0.6452,
2242
+ "step": 2910
2243
+ },
2244
+ {
2245
+ "epoch": 2.291993720565149,
2246
+ "grad_norm": 5.84375,
2247
+ "learning_rate": 4.411446405952618e-05,
2248
+ "loss": 0.5403,
2249
+ "step": 2920
2250
+ },
2251
+ {
2252
+ "epoch": 2.2998430141287285,
2253
+ "grad_norm": 4.75,
2254
+ "learning_rate": 4.376797118230979e-05,
2255
+ "loss": 0.5628,
2256
+ "step": 2930
2257
+ },
2258
+ {
2259
+ "epoch": 2.3076923076923075,
2260
+ "grad_norm": 6.1875,
2261
+ "learning_rate": 4.342178204538588e-05,
2262
+ "loss": 0.5914,
2263
+ "step": 2940
2264
+ },
2265
+ {
2266
+ "epoch": 2.315541601255887,
2267
+ "grad_norm": 5.8125,
2268
+ "learning_rate": 4.307591352152459e-05,
2269
+ "loss": 0.556,
2270
+ "step": 2950
2271
+ },
2272
+ {
2273
+ "epoch": 2.3233908948194664,
2274
+ "grad_norm": 7.125,
2275
+ "learning_rate": 4.273038246786986e-05,
2276
+ "loss": 0.6112,
2277
+ "step": 2960
2278
+ },
2279
+ {
2280
+ "epoch": 2.3312401883830454,
2281
+ "grad_norm": 5.9375,
2282
+ "learning_rate": 4.238520572511773e-05,
2283
+ "loss": 0.5364,
2284
+ "step": 2970
2285
+ },
2286
+ {
2287
+ "epoch": 2.339089481946625,
2288
+ "grad_norm": 5.1875,
2289
+ "learning_rate": 4.2040400116695666e-05,
2290
+ "loss": 0.5563,
2291
+ "step": 2980
2292
+ },
2293
+ {
2294
+ "epoch": 2.3469387755102042,
2295
+ "grad_norm": 5.15625,
2296
+ "learning_rate": 4.169598244794261e-05,
2297
+ "loss": 0.5306,
2298
+ "step": 2990
2299
+ },
2300
+ {
2301
+ "epoch": 2.3547880690737832,
2302
+ "grad_norm": 5.8125,
2303
+ "learning_rate": 4.135196950528982e-05,
2304
+ "loss": 0.5319,
2305
+ "step": 3000
2306
+ },
2307
+ {
2308
+ "epoch": 2.3547880690737832,
2309
+ "eval/nano_beir.arguana": 0.4878178763543074,
2310
+ "step": 3000
2311
+ },
2312
+ {
2313
+ "epoch": 2.3547880690737832,
2314
+ "eval/nano_beir.climate_fever": 0.2621530513631117,
2315
+ "step": 3000
2316
+ },
2317
+ {
2318
+ "epoch": 2.3547880690737832,
2319
+ "eval/nano_beir.dbpedia_entity": 0.33505269293989903,
2320
+ "step": 3000
2321
+ },
2322
+ {
2323
+ "epoch": 2.3547880690737832,
2324
+ "eval/nano_beir.fever": 0.6610619863654493,
2325
+ "step": 3000
2326
+ },
2327
+ {
2328
+ "epoch": 2.3547880690737832,
2329
+ "eval/nano_beir.fiqa": 0.4342255831609204,
2330
+ "step": 3000
2331
+ },
2332
+ {
2333
+ "epoch": 2.3547880690737832,
2334
+ "eval/nano_beir.hotpotqa": 0.6733450414585055,
2335
+ "step": 3000
2336
+ },
2337
+ {
2338
+ "epoch": 2.3547880690737832,
2339
+ "eval/nano_beir.nfcorpus": 0.3444222471148137,
2340
+ "step": 3000
2341
+ },
2342
+ {
2343
+ "epoch": 2.3547880690737832,
2344
+ "eval/nano_beir.nq": 0.4083515153557494,
2345
+ "step": 3000
2346
+ },
2347
+ {
2348
+ "epoch": 2.3547880690737832,
2349
+ "eval/nano_beir.quora": 0.849812776247734,
2350
+ "step": 3000
2351
+ },
2352
+ {
2353
+ "epoch": 2.3547880690737832,
2354
+ "eval/nano_beir.scidocs": 0.2512385413605234,
2355
+ "step": 3000
2356
+ },
2357
+ {
2358
+ "epoch": 2.3547880690737832,
2359
+ "eval/nano_beir.scifact": 0.7541787180524279,
2360
+ "step": 3000
2361
+ },
2362
+ {
2363
+ "epoch": 2.3547880690737832,
2364
+ "eval/nano_beir.webis_touche2020": 0.2821943340980588,
2365
+ "step": 3000
2366
+ },
2367
+ {
2368
+ "epoch": 2.3547880690737832,
2369
+ "eval/avg": 0.47865453032262506,
2370
+ "step": 3000
2371
+ },
2372
+ {
2373
+ "epoch": 2.3626373626373627,
2374
+ "grad_norm": 5.0625,
2375
+ "learning_rate": 4.100837805544279e-05,
2376
+ "loss": 0.5269,
2377
+ "step": 3010
2378
+ },
2379
+ {
2380
+ "epoch": 2.370486656200942,
2381
+ "grad_norm": 7.09375,
2382
+ "learning_rate": 4.0665224844564053e-05,
2383
+ "loss": 0.5913,
2384
+ "step": 3020
2385
+ },
2386
+ {
2387
+ "epoch": 2.378335949764521,
2388
+ "grad_norm": 5.25,
2389
+ "learning_rate": 4.032252659745699e-05,
2390
+ "loss": 0.5511,
2391
+ "step": 3030
2392
+ },
2393
+ {
2394
+ "epoch": 2.3861852433281006,
2395
+ "grad_norm": 5.75,
2396
+ "learning_rate": 3.9980300016750694e-05,
2397
+ "loss": 0.5751,
2398
+ "step": 3040
2399
+ },
2400
+ {
2401
+ "epoch": 2.39403453689168,
2402
+ "grad_norm": 5.65625,
2403
+ "learning_rate": 3.963856178208588e-05,
2404
+ "loss": 0.6321,
2405
+ "step": 3050
2406
+ },
2407
+ {
2408
+ "epoch": 2.401883830455259,
2409
+ "grad_norm": 4.90625,
2410
+ "learning_rate": 3.9297328549302e-05,
2411
+ "loss": 0.5079,
2412
+ "step": 3060
2413
+ },
2414
+ {
2415
+ "epoch": 2.4097331240188384,
2416
+ "grad_norm": 7.96875,
2417
+ "learning_rate": 3.895661694962542e-05,
2418
+ "loss": 0.6575,
2419
+ "step": 3070
2420
+ },
2421
+ {
2422
+ "epoch": 2.4175824175824174,
2423
+ "grad_norm": 5.0,
2424
+ "learning_rate": 3.8616443588858805e-05,
2425
+ "loss": 0.5986,
2426
+ "step": 3080
2427
+ },
2428
+ {
2429
+ "epoch": 2.425431711145997,
2430
+ "grad_norm": 5.375,
2431
+ "learning_rate": 3.827682504657187e-05,
2432
+ "loss": 0.5662,
2433
+ "step": 3090
2434
+ },
2435
+ {
2436
+ "epoch": 2.4332810047095763,
2437
+ "grad_norm": 5.46875,
2438
+ "learning_rate": 3.7937777875293244e-05,
2439
+ "loss": 0.4804,
2440
+ "step": 3100
2441
+ },
2442
+ {
2443
+ "epoch": 2.4411302982731553,
2444
+ "grad_norm": 5.46875,
2445
+ "learning_rate": 3.759931859970374e-05,
2446
+ "loss": 0.6067,
2447
+ "step": 3110
2448
+ },
2449
+ {
2450
+ "epoch": 2.4489795918367347,
2451
+ "grad_norm": 6.96875,
2452
+ "learning_rate": 3.72614637158309e-05,
2453
+ "loss": 0.4898,
2454
+ "step": 3120
2455
+ },
2456
+ {
2457
+ "epoch": 2.4568288854003137,
2458
+ "grad_norm": 5.5,
2459
+ "learning_rate": 3.692422969024516e-05,
2460
+ "loss": 0.5711,
2461
+ "step": 3130
2462
+ },
2463
+ {
2464
+ "epoch": 2.464678178963893,
2465
+ "grad_norm": 6.75,
2466
+ "learning_rate": 3.6587632959257165e-05,
2467
+ "loss": 0.5717,
2468
+ "step": 3140
2469
+ },
2470
+ {
2471
+ "epoch": 2.4725274725274726,
2472
+ "grad_norm": 6.4375,
2473
+ "learning_rate": 3.625168992811671e-05,
2474
+ "loss": 0.5582,
2475
+ "step": 3150
2476
+ },
2477
+ {
2478
+ "epoch": 2.4803767660910516,
2479
+ "grad_norm": 4.4375,
2480
+ "learning_rate": 3.591641697021317e-05,
2481
+ "loss": 0.515,
2482
+ "step": 3160
2483
+ },
2484
+ {
2485
+ "epoch": 2.488226059654631,
2486
+ "grad_norm": 6.09375,
2487
+ "learning_rate": 3.558183042627755e-05,
2488
+ "loss": 0.5472,
2489
+ "step": 3170
2490
+ },
2491
+ {
2492
+ "epoch": 2.4960753532182105,
2493
+ "grad_norm": 5.09375,
2494
+ "learning_rate": 3.524794660358593e-05,
2495
+ "loss": 0.4993,
2496
+ "step": 3180
2497
+ },
2498
+ {
2499
+ "epoch": 2.5039246467817895,
2500
+ "grad_norm": 5.6875,
2501
+ "learning_rate": 3.491478177516484e-05,
2502
+ "loss": 0.5712,
2503
+ "step": 3190
2504
+ },
2505
+ {
2506
+ "epoch": 2.511773940345369,
2507
+ "grad_norm": 5.3125,
2508
+ "learning_rate": 3.4582352178997935e-05,
2509
+ "loss": 0.6757,
2510
+ "step": 3200
2511
+ },
2512
+ {
2513
+ "epoch": 2.5196232339089484,
2514
+ "grad_norm": 5.03125,
2515
+ "learning_rate": 3.425067401723477e-05,
2516
+ "loss": 0.4989,
2517
+ "step": 3210
2518
+ },
2519
+ {
2520
+ "epoch": 2.5274725274725274,
2521
+ "grad_norm": 5.9375,
2522
+ "learning_rate": 3.3919763455401015e-05,
2523
+ "loss": 0.4891,
2524
+ "step": 3220
2525
+ },
2526
+ {
2527
+ "epoch": 2.535321821036107,
2528
+ "grad_norm": 4.75,
2529
+ "learning_rate": 3.358963662161062e-05,
2530
+ "loss": 0.4767,
2531
+ "step": 3230
2532
+ },
2533
+ {
2534
+ "epoch": 2.5431711145996863,
2535
+ "grad_norm": 6.75,
2536
+ "learning_rate": 3.326030960577972e-05,
2537
+ "loss": 0.5388,
2538
+ "step": 3240
2539
+ },
2540
+ {
2541
+ "epoch": 2.5510204081632653,
2542
+ "grad_norm": 5.375,
2543
+ "learning_rate": 3.293179845884245e-05,
2544
+ "loss": 0.5458,
2545
+ "step": 3250
2546
+ },
2547
+ {
2548
+ "epoch": 2.5588697017268447,
2549
+ "grad_norm": 6.28125,
2550
+ "learning_rate": 3.2604119191968654e-05,
2551
+ "loss": 0.5843,
2552
+ "step": 3260
2553
+ },
2554
+ {
2555
+ "epoch": 2.566718995290424,
2556
+ "grad_norm": 6.8125,
2557
+ "learning_rate": 3.2277287775783525e-05,
2558
+ "loss": 0.585,
2559
+ "step": 3270
2560
+ },
2561
+ {
2562
+ "epoch": 2.574568288854003,
2563
+ "grad_norm": 6.5,
2564
+ "learning_rate": 3.195132013958918e-05,
2565
+ "loss": 0.5527,
2566
+ "step": 3280
2567
+ },
2568
+ {
2569
+ "epoch": 2.5824175824175826,
2570
+ "grad_norm": 5.75,
2571
+ "learning_rate": 3.162623217058834e-05,
2572
+ "loss": 0.5515,
2573
+ "step": 3290
2574
+ },
2575
+ {
2576
+ "epoch": 2.5902668759811616,
2577
+ "grad_norm": 5.71875,
2578
+ "learning_rate": 3.130203971310999e-05,
2579
+ "loss": 0.546,
2580
+ "step": 3300
2581
+ },
2582
+ {
2583
+ "epoch": 2.598116169544741,
2584
+ "grad_norm": 6.75,
2585
+ "learning_rate": 3.097875856783713e-05,
2586
+ "loss": 0.6335,
2587
+ "step": 3310
2588
+ },
2589
+ {
2590
+ "epoch": 2.60596546310832,
2591
+ "grad_norm": 5.34375,
2592
+ "learning_rate": 3.065640449103669e-05,
2593
+ "loss": 0.568,
2594
+ "step": 3320
2595
+ },
2596
+ {
2597
+ "epoch": 2.6138147566718994,
2598
+ "grad_norm": 5.96875,
2599
+ "learning_rate": 3.033499319379163e-05,
2600
+ "loss": 0.564,
2601
+ "step": 3330
2602
+ },
2603
+ {
2604
+ "epoch": 2.621664050235479,
2605
+ "grad_norm": 4.8125,
2606
+ "learning_rate": 3.001454034123512e-05,
2607
+ "loss": 0.5238,
2608
+ "step": 3340
2609
+ },
2610
+ {
2611
+ "epoch": 2.629513343799058,
2612
+ "grad_norm": 6.875,
2613
+ "learning_rate": 2.9695061551787108e-05,
2614
+ "loss": 0.5249,
2615
+ "step": 3350
2616
+ },
2617
+ {
2618
+ "epoch": 2.6373626373626373,
2619
+ "grad_norm": 3.375,
2620
+ "learning_rate": 2.9376572396393048e-05,
2621
+ "loss": 0.5545,
2622
+ "step": 3360
2623
+ },
2624
+ {
2625
+ "epoch": 2.6452119309262168,
2626
+ "grad_norm": 5.625,
2627
+ "learning_rate": 2.9059088397765088e-05,
2628
+ "loss": 0.616,
2629
+ "step": 3370
2630
+ },
2631
+ {
2632
+ "epoch": 2.6530612244897958,
2633
+ "grad_norm": 5.75,
2634
+ "learning_rate": 2.874262502962537e-05,
2635
+ "loss": 0.5634,
2636
+ "step": 3380
2637
+ },
2638
+ {
2639
+ "epoch": 2.660910518053375,
2640
+ "grad_norm": 6.46875,
2641
+ "learning_rate": 2.8427197715952047e-05,
2642
+ "loss": 0.597,
2643
+ "step": 3390
2644
+ },
2645
+ {
2646
+ "epoch": 2.6687598116169546,
2647
+ "grad_norm": 5.28125,
2648
+ "learning_rate": 2.811282183022736e-05,
2649
+ "loss": 0.4931,
2650
+ "step": 3400
2651
+ },
2652
+ {
2653
+ "epoch": 2.6766091051805336,
2654
+ "grad_norm": 4.1875,
2655
+ "learning_rate": 2.7799512694688468e-05,
2656
+ "loss": 0.5645,
2657
+ "step": 3410
2658
+ },
2659
+ {
2660
+ "epoch": 2.684458398744113,
2661
+ "grad_norm": 4.125,
2662
+ "learning_rate": 2.7487285579580637e-05,
2663
+ "loss": 0.4427,
2664
+ "step": 3420
2665
+ },
2666
+ {
2667
+ "epoch": 2.6923076923076925,
2668
+ "grad_norm": 5.5,
2669
+ "learning_rate": 2.7176155702412937e-05,
2670
+ "loss": 0.5652,
2671
+ "step": 3430
2672
+ },
2673
+ {
2674
+ "epoch": 2.7001569858712715,
2675
+ "grad_norm": 5.9375,
2676
+ "learning_rate": 2.686613822721666e-05,
2677
+ "loss": 0.5637,
2678
+ "step": 3440
2679
+ },
2680
+ {
2681
+ "epoch": 2.708006279434851,
2682
+ "grad_norm": 4.21875,
2683
+ "learning_rate": 2.6557248263806174e-05,
2684
+ "loss": 0.5387,
2685
+ "step": 3450
2686
+ },
2687
+ {
2688
+ "epoch": 2.7158555729984304,
2689
+ "grad_norm": 5.375,
2690
+ "learning_rate": 2.6249500867042522e-05,
2691
+ "loss": 0.5807,
2692
+ "step": 3460
2693
+ },
2694
+ {
2695
+ "epoch": 2.7237048665620094,
2696
+ "grad_norm": 6.21875,
2697
+ "learning_rate": 2.5942911036099655e-05,
2698
+ "loss": 0.5878,
2699
+ "step": 3470
2700
+ },
2701
+ {
2702
+ "epoch": 2.731554160125589,
2703
+ "grad_norm": 4.375,
2704
+ "learning_rate": 2.5637493713733374e-05,
2705
+ "loss": 0.5696,
2706
+ "step": 3480
2707
+ },
2708
+ {
2709
+ "epoch": 2.739403453689168,
2710
+ "grad_norm": 6.21875,
2711
+ "learning_rate": 2.533326378555314e-05,
2712
+ "loss": 0.5401,
2713
+ "step": 3490
2714
+ },
2715
+ {
2716
+ "epoch": 2.7472527472527473,
2717
+ "grad_norm": 6.4375,
2718
+ "learning_rate": 2.5030236079296444e-05,
2719
+ "loss": 0.5441,
2720
+ "step": 3500
2721
+ },
2722
+ {
2723
+ "epoch": 2.7551020408163263,
2724
+ "grad_norm": 5.1875,
2725
+ "learning_rate": 2.4728425364106135e-05,
2726
+ "loss": 0.5751,
2727
+ "step": 3510
2728
+ },
2729
+ {
2730
+ "epoch": 2.7629513343799057,
2731
+ "grad_norm": 5.46875,
2732
+ "learning_rate": 2.442784634981071e-05,
2733
+ "loss": 0.5527,
2734
+ "step": 3520
2735
+ },
2736
+ {
2737
+ "epoch": 2.770800627943485,
2738
+ "grad_norm": 6.1875,
2739
+ "learning_rate": 2.412851368620726e-05,
2740
+ "loss": 0.537,
2741
+ "step": 3530
2742
+ },
2743
+ {
2744
+ "epoch": 2.778649921507064,
2745
+ "grad_norm": 4.65625,
2746
+ "learning_rate": 2.383044196234753e-05,
2747
+ "loss": 0.5304,
2748
+ "step": 3540
2749
+ },
2750
+ {
2751
+ "epoch": 2.7864992150706436,
2752
+ "grad_norm": 6.75,
2753
+ "learning_rate": 2.3533645705826806e-05,
2754
+ "loss": 0.5959,
2755
+ "step": 3550
2756
+ },
2757
+ {
2758
+ "epoch": 2.794348508634223,
2759
+ "grad_norm": 7.40625,
2760
+ "learning_rate": 2.323813938207593e-05,
2761
+ "loss": 0.525,
2762
+ "step": 3560
2763
+ },
2764
+ {
2765
+ "epoch": 2.802197802197802,
2766
+ "grad_norm": 5.03125,
2767
+ "learning_rate": 2.294393739365621e-05,
2768
+ "loss": 0.6045,
2769
+ "step": 3570
2770
+ },
2771
+ {
2772
+ "epoch": 2.8100470957613815,
2773
+ "grad_norm": 5.4375,
2774
+ "learning_rate": 2.265105407955752e-05,
2775
+ "loss": 0.5697,
2776
+ "step": 3580
2777
+ },
2778
+ {
2779
+ "epoch": 2.817896389324961,
2780
+ "grad_norm": 5.1875,
2781
+ "learning_rate": 2.235950371449938e-05,
2782
+ "loss": 0.5046,
2783
+ "step": 3590
2784
+ },
2785
+ {
2786
+ "epoch": 2.82574568288854,
2787
+ "grad_norm": 7.34375,
2788
+ "learning_rate": 2.2069300508235275e-05,
2789
+ "loss": 0.6226,
2790
+ "step": 3600
2791
+ },
2792
+ {
2793
+ "epoch": 2.8335949764521193,
2794
+ "grad_norm": 5.875,
2795
+ "learning_rate": 2.1780458604860056e-05,
2796
+ "loss": 0.5541,
2797
+ "step": 3610
2798
+ },
2799
+ {
2800
+ "epoch": 2.8414442700156988,
2801
+ "grad_norm": 7.59375,
2802
+ "learning_rate": 2.14929920821206e-05,
2803
+ "loss": 0.5778,
2804
+ "step": 3620
2805
+ },
2806
+ {
2807
+ "epoch": 2.8492935635792778,
2808
+ "grad_norm": 6.8125,
2809
+ "learning_rate": 2.1206914950729672e-05,
2810
+ "loss": 0.6705,
2811
+ "step": 3630
2812
+ },
2813
+ {
2814
+ "epoch": 2.857142857142857,
2815
+ "grad_norm": 5.65625,
2816
+ "learning_rate": 2.0922241153683066e-05,
2817
+ "loss": 0.5871,
2818
+ "step": 3640
2819
+ },
2820
+ {
2821
+ "epoch": 2.8649921507064366,
2822
+ "grad_norm": 6.78125,
2823
+ "learning_rate": 2.0638984565580023e-05,
2824
+ "loss": 0.5753,
2825
+ "step": 3650
2826
+ },
2827
+ {
2828
+ "epoch": 2.8728414442700156,
2829
+ "grad_norm": 6.46875,
2830
+ "learning_rate": 2.035715899194704e-05,
2831
+ "loss": 0.6001,
2832
+ "step": 3660
2833
+ },
2834
+ {
2835
+ "epoch": 2.880690737833595,
2836
+ "grad_norm": 5.46875,
2837
+ "learning_rate": 2.007677816856498e-05,
2838
+ "loss": 0.6271,
2839
+ "step": 3670
2840
+ },
2841
+ {
2842
+ "epoch": 2.8885400313971745,
2843
+ "grad_norm": 6.09375,
2844
+ "learning_rate": 1.9797855760799612e-05,
2845
+ "loss": 0.5911,
2846
+ "step": 3680
2847
+ },
2848
+ {
2849
+ "epoch": 2.8963893249607535,
2850
+ "grad_norm": 4.96875,
2851
+ "learning_rate": 1.9520405362935594e-05,
2852
+ "loss": 0.5447,
2853
+ "step": 3690
2854
+ },
2855
+ {
2856
+ "epoch": 2.904238618524333,
2857
+ "grad_norm": 5.9375,
2858
+ "learning_rate": 1.9244440497513893e-05,
2859
+ "loss": 0.5026,
2860
+ "step": 3700
2861
+ },
2862
+ {
2863
+ "epoch": 2.912087912087912,
2864
+ "grad_norm": 5.5,
2865
+ "learning_rate": 1.896997461467272e-05,
2866
+ "loss": 0.5871,
2867
+ "step": 3710
2868
+ },
2869
+ {
2870
+ "epoch": 2.9199372056514914,
2871
+ "grad_norm": 6.0625,
2872
+ "learning_rate": 1.869702109149199e-05,
2873
+ "loss": 0.4704,
2874
+ "step": 3720
2875
+ },
2876
+ {
2877
+ "epoch": 2.9277864992150704,
2878
+ "grad_norm": 4.90625,
2879
+ "learning_rate": 1.842559323134136e-05,
2880
+ "loss": 0.5895,
2881
+ "step": 3730
2882
+ },
2883
+ {
2884
+ "epoch": 2.93563579277865,
2885
+ "grad_norm": 8.25,
2886
+ "learning_rate": 1.8155704263231776e-05,
2887
+ "loss": 0.5151,
2888
+ "step": 3740
2889
+ },
2890
+ {
2891
+ "epoch": 2.9434850863422293,
2892
+ "grad_norm": 6.96875,
2893
+ "learning_rate": 1.788736734117078e-05,
2894
+ "loss": 0.6168,
2895
+ "step": 3750
2896
+ },
2897
+ {
2898
+ "epoch": 2.9513343799058083,
2899
+ "grad_norm": 4.40625,
2900
+ "learning_rate": 1.762059554352143e-05,
2901
+ "loss": 0.5272,
2902
+ "step": 3760
2903
+ },
2904
+ {
2905
+ "epoch": 2.9591836734693877,
2906
+ "grad_norm": 5.5625,
2907
+ "learning_rate": 1.7355401872364758e-05,
2908
+ "loss": 0.6053,
2909
+ "step": 3770
2910
+ },
2911
+ {
2912
+ "epoch": 2.967032967032967,
2913
+ "grad_norm": 4.09375,
2914
+ "learning_rate": 1.709179925286617e-05,
2915
+ "loss": 0.6135,
2916
+ "step": 3780
2917
+ },
2918
+ {
2919
+ "epoch": 2.974882260596546,
2920
+ "grad_norm": 6.78125,
2921
+ "learning_rate": 1.6829800532645447e-05,
2922
+ "loss": 0.5289,
2923
+ "step": 3790
2924
+ },
2925
+ {
2926
+ "epoch": 2.9827315541601256,
2927
+ "grad_norm": 6.34375,
2928
+ "learning_rate": 1.6569418481150595e-05,
2929
+ "loss": 0.5953,
2930
+ "step": 3800
2931
+ },
2932
+ {
2933
+ "epoch": 2.990580847723705,
2934
+ "grad_norm": 5.03125,
2935
+ "learning_rate": 1.6310665789035467e-05,
2936
+ "loss": 0.5632,
2937
+ "step": 3810
2938
+ },
2939
+ {
2940
+ "epoch": 2.998430141287284,
2941
+ "grad_norm": 4.96875,
2942
+ "learning_rate": 1.605355506754121e-05,
2943
+ "loss": 0.6042,
2944
+ "step": 3820
2945
+ },
2946
+ {
2947
+ "epoch": 3.0062794348508635,
2948
+ "grad_norm": 6.28125,
2949
+ "learning_rate": 1.5798098847881666e-05,
2950
+ "loss": 0.6241,
2951
+ "step": 3830
2952
+ },
2953
+ {
2954
+ "epoch": 3.014128728414443,
2955
+ "grad_norm": 5.71875,
2956
+ "learning_rate": 1.554430958063259e-05,
2957
+ "loss": 0.5378,
2958
+ "step": 3840
2959
+ },
2960
+ {
2961
+ "epoch": 3.021978021978022,
2962
+ "grad_norm": 5.25,
2963
+ "learning_rate": 1.529219963512481e-05,
2964
+ "loss": 0.4997,
2965
+ "step": 3850
2966
+ },
2967
+ {
2968
+ "epoch": 3.0298273155416013,
2969
+ "grad_norm": 5.0625,
2970
+ "learning_rate": 1.5041781298841423e-05,
2971
+ "loss": 0.4719,
2972
+ "step": 3860
2973
+ },
2974
+ {
2975
+ "epoch": 3.0376766091051803,
2976
+ "grad_norm": 5.6875,
2977
+ "learning_rate": 1.4793066776818842e-05,
2978
+ "loss": 0.5238,
2979
+ "step": 3870
2980
+ },
2981
+ {
2982
+ "epoch": 3.0455259026687598,
2983
+ "grad_norm": 5.4375,
2984
+ "learning_rate": 1.4546068191051987e-05,
2985
+ "loss": 0.5892,
2986
+ "step": 3880
2987
+ },
2988
+ {
2989
+ "epoch": 3.053375196232339,
2990
+ "grad_norm": 8.5625,
2991
+ "learning_rate": 1.4300797579903474e-05,
2992
+ "loss": 0.4706,
2993
+ "step": 3890
2994
+ },
2995
+ {
2996
+ "epoch": 3.061224489795918,
2997
+ "grad_norm": 3.28125,
2998
+ "learning_rate": 1.4057266897516841e-05,
2999
+ "loss": 0.5153,
3000
+ "step": 3900
3001
+ },
3002
+ {
3003
+ "epoch": 3.0690737833594977,
3004
+ "grad_norm": 4.15625,
3005
+ "learning_rate": 1.3815488013233984e-05,
3006
+ "loss": 0.5358,
3007
+ "step": 3910
3008
+ },
3009
+ {
3010
+ "epoch": 3.076923076923077,
3011
+ "grad_norm": 6.8125,
3012
+ "learning_rate": 1.3575472711016635e-05,
3013
+ "loss": 0.6259,
3014
+ "step": 3920
3015
+ },
3016
+ {
3017
+ "epoch": 3.084772370486656,
3018
+ "grad_norm": 6.09375,
3019
+ "learning_rate": 1.3337232688872009e-05,
3020
+ "loss": 0.5336,
3021
+ "step": 3930
3022
+ },
3023
+ {
3024
+ "epoch": 3.0926216640502355,
3025
+ "grad_norm": 6.09375,
3026
+ "learning_rate": 1.3100779558282672e-05,
3027
+ "loss": 0.5329,
3028
+ "step": 3940
3029
+ },
3030
+ {
3031
+ "epoch": 3.100470957613815,
3032
+ "grad_norm": 5.1875,
3033
+ "learning_rate": 1.2866124843640615e-05,
3034
+ "loss": 0.5897,
3035
+ "step": 3950
3036
+ },
3037
+ {
3038
+ "epoch": 3.108320251177394,
3039
+ "grad_norm": 7.59375,
3040
+ "learning_rate": 1.2633279981685608e-05,
3041
+ "loss": 0.5754,
3042
+ "step": 3960
3043
+ },
3044
+ {
3045
+ "epoch": 3.1161695447409734,
3046
+ "grad_norm": 6.0625,
3047
+ "learning_rate": 1.240225632094773e-05,
3048
+ "loss": 0.5477,
3049
+ "step": 3970
3050
+ },
3051
+ {
3052
+ "epoch": 3.1240188383045524,
3053
+ "grad_norm": 5.3125,
3054
+ "learning_rate": 1.217306512119425e-05,
3055
+ "loss": 0.5441,
3056
+ "step": 3980
3057
+ },
3058
+ {
3059
+ "epoch": 3.131868131868132,
3060
+ "grad_norm": 5.78125,
3061
+ "learning_rate": 1.194571755288092e-05,
3062
+ "loss": 0.5345,
3063
+ "step": 3990
3064
+ },
3065
+ {
3066
+ "epoch": 3.1397174254317113,
3067
+ "grad_norm": 6.125,
3068
+ "learning_rate": 1.1720224696607474e-05,
3069
+ "loss": 0.5756,
3070
+ "step": 4000
3071
+ }
3072
+ ],
3073
+ "logging_steps": 10,
3074
+ "max_steps": 5000,
3075
+ "num_input_tokens_seen": 0,
3076
+ "num_train_epochs": 4,
3077
+ "save_steps": 1000,
3078
+ "stateful_callbacks": {
3079
+ "TrainerControl": {
3080
+ "args": {
3081
+ "should_epoch_stop": false,
3082
+ "should_evaluate": false,
3083
+ "should_log": false,
3084
+ "should_save": true,
3085
+ "should_training_stop": false
3086
+ },
3087
+ "attributes": {}
3088
+ }
3089
+ },
3090
+ "total_flos": 0.0,
3091
+ "train_batch_size": 16,
3092
+ "trial_name": null,
3093
+ "trial_params": null
3094
+ }
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-4000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2ed821ed7ef0074e4b13618254be9ea2c38d68708aace5456b6baa256883072
3
+ size 6097
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/checkpoint-5000/removed ADDED
File without changes
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "gradient_checkpointing": false,
20
+ "hidden_activation": "gelu",
21
+ "hidden_size": 768,
22
+ "initializer_cutoff_factor": 2.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 1152,
25
+ "layer_norm_eps": 1e-05,
26
+ "local_attention": 128,
27
+ "local_rope_theta": 10000.0,
28
+ "max_position_embeddings": 8192,
29
+ "mlp_bias": false,
30
+ "mlp_dropout": 0.0,
31
+ "model_type": "modernbert",
32
+ "norm_bias": false,
33
+ "norm_eps": 1e-05,
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 22,
36
+ "pad_token_id": 50283,
37
+ "position_embedding_type": "absolute",
38
+ "repad_logits_with_grad": false,
39
+ "sep_token_id": 50282,
40
+ "sparse_pred_ignore_index": -100,
41
+ "sparse_prediction": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.53.0",
44
+ "vocab_size": 50368
45
+ }
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5a4e3e60f62a3f1aa05b880c77e0b56f807efafa62f60a5f452019530af6711
3
+ size 298041696
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:823620ff1dd0a6e380f4c8e0ec885b6b1aa2023dde15f5e41b13c2b6e43f80ab
3
+ size 596170443
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24c12cceaffe8acec04189ba71dbb81f100dd5cd03b86390253117b3fab2911d
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c24e46cf0599262f5ee2ba9ec6bddad7d3e8feb965b424cc9c4dc4563771b1e2
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5af6ed530327c4af91aa04cc19a7523ed8295478ca34bb5f3124ff88cd33f77
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74c82a4770595a28e603ecf5353ab454c0fb9b0e184187b20ebdc4685217f3a5
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy/runs/Dec10_21-48-05_nid005066/events.out.tfevents.1765396268.nid005066.34579.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09ef82f0afb6dad91602237429ec2d0623846df5ccc010205b1196cd445f0068
3
+ size 119196