DylanJHJ commited on
Commit
74371a8
·
1 Parent(s): 7fa7ec1

update the ones with request

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +40 -0
  2. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/config.json +45 -0
  3. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/model.safetensors +3 -0
  4. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/optimizer.pt +3 -0
  5. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/rng_state_0.pth +3 -0
  6. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/rng_state_1.pth +3 -0
  7. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/rng_state_2.pth +3 -0
  8. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/rng_state_3.pth +3 -0
  9. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/scheduler.pt +3 -0
  10. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/trainer_state.json +799 -0
  11. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/training_args.bin +3 -0
  12. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/config.json +45 -0
  13. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/model.safetensors +3 -0
  14. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/optimizer.pt +3 -0
  15. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/rng_state_0.pth +3 -0
  16. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/rng_state_1.pth +3 -0
  17. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/rng_state_2.pth +3 -0
  18. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/rng_state_3.pth +3 -0
  19. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/scheduler.pt +3 -0
  20. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/trainer_state.json +1564 -0
  21. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/training_args.bin +3 -0
  22. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/config.json +45 -0
  23. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/model.safetensors +3 -0
  24. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/optimizer.pt +3 -0
  25. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/rng_state_0.pth +3 -0
  26. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/rng_state_1.pth +3 -0
  27. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/rng_state_2.pth +3 -0
  28. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/rng_state_3.pth +3 -0
  29. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/scheduler.pt +3 -0
  30. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/trainer_state.json +2329 -0
  31. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/training_args.bin +3 -0
  32. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/config.json +45 -0
  33. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/model.safetensors +3 -0
  34. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/optimizer.pt +3 -0
  35. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/rng_state_0.pth +3 -0
  36. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/rng_state_1.pth +3 -0
  37. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/rng_state_2.pth +3 -0
  38. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/rng_state_3.pth +3 -0
  39. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/scheduler.pt +3 -0
  40. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/trainer_state.json +3094 -0
  41. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/training_args.bin +3 -0
  42. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-5000/moved +0 -0
  43. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/config.json +45 -0
  44. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/model.safetensors +3 -0
  45. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/optimizer.pt +3 -0
  46. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/rng_state_0.pth +3 -0
  47. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/rng_state_1.pth +3 -0
  48. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/rng_state_2.pth +3 -0
  49. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/rng_state_3.pth +3 -0
  50. modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/runs/Jan01_23-09-47_nid005111/events.out.tfevents.1767301973.nid005111.95408.0 +3 -0
.gitattributes CHANGED
@@ -53,3 +53,43 @@ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco/chec
53
  modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
54
  modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco/checkpoint-1000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
55
  modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco/checkpoint-1000/model.safetensors filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
53
  modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
54
  modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco/checkpoint-1000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
55
  modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco/checkpoint-1000/model.safetensors filter=lfs diff=lfs merge=lfs -text
56
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco.request/checkpoint-2000/model.safetensors filter=lfs diff=lfs merge=lfs -text
57
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco.request/checkpoint-2000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
58
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco.request/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
59
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco.request/checkpoint-3000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
60
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco.request/model.safetensors filter=lfs diff=lfs merge=lfs -text
61
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco.request/optimizer.pt filter=lfs diff=lfs merge=lfs -text
62
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco.request/checkpoint-1000/model.safetensors filter=lfs diff=lfs merge=lfs -text
63
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco.request/checkpoint-1000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
64
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco.request/checkpoint-4000/model.safetensors filter=lfs diff=lfs merge=lfs -text
65
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.msmarco.request/checkpoint-4000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
66
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/model.safetensors filter=lfs diff=lfs merge=lfs -text
67
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
68
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
69
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
70
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/model.safetensors filter=lfs diff=lfs merge=lfs -text
71
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/optimizer.pt filter=lfs diff=lfs merge=lfs -text
72
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/model.safetensors filter=lfs diff=lfs merge=lfs -text
73
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
74
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/model.safetensors filter=lfs diff=lfs merge=lfs -text
75
+ modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
76
+ .git/lfs/objects/0c/e8/0ce892ac499c726f9f0ffc58f21eaed98059a99787e8a5e0b47b09519dd56c03 filter=lfs diff=lfs merge=lfs -text
77
+ .git/lfs/objects/b5/a4/b5a4e3e60f62a3f1aa05b880c77e0b56f807efafa62f60a5f452019530af6711 filter=lfs diff=lfs merge=lfs -text
78
+ .git/lfs/objects/2e/d8/2ed820e6dfb3db7d393c18aa5f5eb9c5baf64f9c9c906ace62bac5ba200d4cf2 filter=lfs diff=lfs merge=lfs -text
79
+ .git/lfs/objects/0d/7d/0d7de69bdc8b2c1365cd9379c80c2e9660583d3b9d21c5da9f70bc756aef8c5c filter=lfs diff=lfs merge=lfs -text
80
+ .git/lfs/objects/c7/ca/c7ca94df487520ae2c411613298582fccb5e50edc91edd1e3311d560a4ff7427 filter=lfs diff=lfs merge=lfs -text
81
+ .git/lfs/objects/40/49/4049b989fa1e41ca8b3f1d3acf616f06129dd1998d0d24bb71a4e9077014975b filter=lfs diff=lfs merge=lfs -text
82
+ .git/lfs/objects/31/5e/315edc508eedf0940ded0c431d5b741a533b41fd7759fa8996d6f2a53fa7a1f5 filter=lfs diff=lfs merge=lfs -text
83
+ .git/lfs/objects/98/16/9816537f9d6d804d6866451e1583a3857ffd0af4e789dd38f7361227c01b1b62 filter=lfs diff=lfs merge=lfs -text
84
+ .git/lfs/objects/eb/8c/eb8c18aefb03ba8afaf3ddafca1cd6ea74416c497632a03907346835d107adad filter=lfs diff=lfs merge=lfs -text
85
+ .git/lfs/objects/44/44/4444b588f984289c2834e5eed337f71344ab7374e6ca84c4ebf27f527fdfb31e filter=lfs diff=lfs merge=lfs -text
86
+ .git/lfs/objects/f8/78/f8788fb899610dac5c81332d3ad006033814e54fdc7526f95926d4a8228a5000 filter=lfs diff=lfs merge=lfs -text
87
+ .git/lfs/objects/1e/e5/1ee565563332516e4024443fa4e532456a9f9be6bcbe4be7011aa8a9b89ed496 filter=lfs diff=lfs merge=lfs -text
88
+ .git/lfs/objects/a5/4b/a54bb859b7620f51ba6e2f7feb41f03eed4f1e229e4d5a26ba32b89f7cfda494 filter=lfs diff=lfs merge=lfs -text
89
+ .git/lfs/objects/8b/09/8b0972cf18a384a6114761e112ac68fee94239f09830af05f762267cbfd9ae4d filter=lfs diff=lfs merge=lfs -text
90
+ .git/lfs/objects/68/4d/684d731b0a9245de462cf0af539b773d34256b419759d530a7b160bea702561c filter=lfs diff=lfs merge=lfs -text
91
+ .git/lfs/objects/1b/a4/1ba44b14f4a98062eaecb16a59796513bafcff78abd007a3fba0d5e20e889ea8 filter=lfs diff=lfs merge=lfs -text
92
+ .git/lfs/objects/0b/21/0b21d57153ab71ab9dadd49944db15e74447f04e9b0cb67c61655d0106e4a0d6 filter=lfs diff=lfs merge=lfs -text
93
+ .git/lfs/objects/6d/41/6d41fde859da34b9501b484af69475f50a37de92a28b6fa47a87c5caff3703ae filter=lfs diff=lfs merge=lfs -text
94
+ .git/lfs/objects/82/36/823620ff1dd0a6e380f4c8e0ec885b6b1aa2023dde15f5e41b13c2b6e43f80ab filter=lfs diff=lfs merge=lfs -text
95
+ .git/lfs/objects/1f/8c/1f8cea23e59027c2c5b36168d940ec4d3844b1f29c814cb8d6202d62fa45f9c0 filter=lfs diff=lfs merge=lfs -text
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "gradient_checkpointing": false,
20
+ "hidden_activation": "gelu",
21
+ "hidden_size": 768,
22
+ "initializer_cutoff_factor": 2.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 1152,
25
+ "layer_norm_eps": 1e-05,
26
+ "local_attention": 128,
27
+ "local_rope_theta": 10000.0,
28
+ "max_position_embeddings": 8192,
29
+ "mlp_bias": false,
30
+ "mlp_dropout": 0.0,
31
+ "model_type": "modernbert",
32
+ "norm_bias": false,
33
+ "norm_eps": 1e-05,
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 22,
36
+ "pad_token_id": 50283,
37
+ "position_embedding_type": "absolute",
38
+ "repad_logits_with_grad": false,
39
+ "sep_token_id": 50282,
40
+ "sparse_pred_ignore_index": -100,
41
+ "sparse_prediction": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.53.0",
44
+ "vocab_size": 50368
45
+ }
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a268b8d3b223be4857defa18f90c46cf839f603e9f616465a0472f05acc3215a
3
+ size 298041696
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5831f1d98a7fb576cf915ca7b4410402babd2e88a871c415aeb759500f4add58
3
+ size 596170443
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3044a568f2c48f55afab6e63cf0c046d3586784e77e39b1af825abecfb2cdba8
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86accf27064cdd503053e90476a6bd10de333d4ff0594535ad55ea13a473c91d
3
+ size 15429
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18ca8d714ef40be035404c1957b5a4dee84e1f43980408393f8aa710552ee6f6
3
+ size 15429
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2cfdebe99e40accc9c9d8f09c63136a14abda997d9b501969ec8e16e9d183179
3
+ size 15429
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:053b181ac2db5ceae48310494587c492212456802136bc3852c62ccc7b9ff38c
3
+ size 1465
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/trainer_state.json ADDED
@@ -0,0 +1,799 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 0.7849293563579278,
6
+ "eval_steps": 1000.0,
7
+ "global_step": 1000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0,
14
+ "eval/nano_beir.arguana": 0.4400549254974839,
15
+ "step": 0
16
+ },
17
+ {
18
+ "epoch": 0,
19
+ "eval/nano_beir.climate_fever": 0.22880053057901228,
20
+ "step": 0
21
+ },
22
+ {
23
+ "epoch": 0,
24
+ "eval/nano_beir.dbpedia_entity": 0.30026716538531234,
25
+ "step": 0
26
+ },
27
+ {
28
+ "epoch": 0,
29
+ "eval/nano_beir.fever": 0.6419178788751609,
30
+ "step": 0
31
+ },
32
+ {
33
+ "epoch": 0,
34
+ "eval/nano_beir.fiqa": 0.4370593595651809,
35
+ "step": 0
36
+ },
37
+ {
38
+ "epoch": 0,
39
+ "eval/nano_beir.hotpotqa": 0.6723047477224562,
40
+ "step": 0
41
+ },
42
+ {
43
+ "epoch": 0,
44
+ "eval/nano_beir.nfcorpus": 0.3437446614516137,
45
+ "step": 0
46
+ },
47
+ {
48
+ "epoch": 0,
49
+ "eval/nano_beir.nq": 0.436711543163819,
50
+ "step": 0
51
+ },
52
+ {
53
+ "epoch": 0,
54
+ "eval/nano_beir.quora": 0.8439997970269911,
55
+ "step": 0
56
+ },
57
+ {
58
+ "epoch": 0,
59
+ "eval/nano_beir.scidocs": 0.26304331393922176,
60
+ "step": 0
61
+ },
62
+ {
63
+ "epoch": 0,
64
+ "eval/nano_beir.scifact": 0.7599341408912704,
65
+ "step": 0
66
+ },
67
+ {
68
+ "epoch": 0,
69
+ "eval/nano_beir.webis_touche2020": 0.29696490432942557,
70
+ "step": 0
71
+ },
72
+ {
73
+ "epoch": 0,
74
+ "eval/avg": 0.47206691403557904,
75
+ "step": 0
76
+ },
77
+ {
78
+ "epoch": 0.007849293563579277,
79
+ "grad_norm": 8.25,
80
+ "learning_rate": 1.8e-06,
81
+ "loss": 0.7314,
82
+ "step": 10
83
+ },
84
+ {
85
+ "epoch": 0.015698587127158554,
86
+ "grad_norm": 7.03125,
87
+ "learning_rate": 3.8e-06,
88
+ "loss": 0.8796,
89
+ "step": 20
90
+ },
91
+ {
92
+ "epoch": 0.023547880690737835,
93
+ "grad_norm": 7.3125,
94
+ "learning_rate": 5.8e-06,
95
+ "loss": 0.7311,
96
+ "step": 30
97
+ },
98
+ {
99
+ "epoch": 0.03139717425431711,
100
+ "grad_norm": 7.15625,
101
+ "learning_rate": 7.8e-06,
102
+ "loss": 0.8214,
103
+ "step": 40
104
+ },
105
+ {
106
+ "epoch": 0.03924646781789639,
107
+ "grad_norm": 9.1875,
108
+ "learning_rate": 9.800000000000001e-06,
109
+ "loss": 0.8489,
110
+ "step": 50
111
+ },
112
+ {
113
+ "epoch": 0.04709576138147567,
114
+ "grad_norm": 6.84375,
115
+ "learning_rate": 1.18e-05,
116
+ "loss": 0.7661,
117
+ "step": 60
118
+ },
119
+ {
120
+ "epoch": 0.054945054945054944,
121
+ "grad_norm": 6.6875,
122
+ "learning_rate": 1.3800000000000002e-05,
123
+ "loss": 0.7525,
124
+ "step": 70
125
+ },
126
+ {
127
+ "epoch": 0.06279434850863422,
128
+ "grad_norm": 7.5625,
129
+ "learning_rate": 1.58e-05,
130
+ "loss": 0.8653,
131
+ "step": 80
132
+ },
133
+ {
134
+ "epoch": 0.0706436420722135,
135
+ "grad_norm": 6.25,
136
+ "learning_rate": 1.78e-05,
137
+ "loss": 0.7607,
138
+ "step": 90
139
+ },
140
+ {
141
+ "epoch": 0.07849293563579278,
142
+ "grad_norm": 5.28125,
143
+ "learning_rate": 1.9800000000000004e-05,
144
+ "loss": 0.7259,
145
+ "step": 100
146
+ },
147
+ {
148
+ "epoch": 0.08634222919937205,
149
+ "grad_norm": 5.59375,
150
+ "learning_rate": 2.18e-05,
151
+ "loss": 0.7862,
152
+ "step": 110
153
+ },
154
+ {
155
+ "epoch": 0.09419152276295134,
156
+ "grad_norm": 5.8125,
157
+ "learning_rate": 2.38e-05,
158
+ "loss": 0.8212,
159
+ "step": 120
160
+ },
161
+ {
162
+ "epoch": 0.10204081632653061,
163
+ "grad_norm": 6.0,
164
+ "learning_rate": 2.58e-05,
165
+ "loss": 0.6884,
166
+ "step": 130
167
+ },
168
+ {
169
+ "epoch": 0.10989010989010989,
170
+ "grad_norm": 5.9375,
171
+ "learning_rate": 2.7800000000000005e-05,
172
+ "loss": 0.712,
173
+ "step": 140
174
+ },
175
+ {
176
+ "epoch": 0.11773940345368916,
177
+ "grad_norm": 5.53125,
178
+ "learning_rate": 2.98e-05,
179
+ "loss": 0.6231,
180
+ "step": 150
181
+ },
182
+ {
183
+ "epoch": 0.12558869701726844,
184
+ "grad_norm": 5.59375,
185
+ "learning_rate": 3.18e-05,
186
+ "loss": 0.6733,
187
+ "step": 160
188
+ },
189
+ {
190
+ "epoch": 0.13343799058084774,
191
+ "grad_norm": 6.84375,
192
+ "learning_rate": 3.38e-05,
193
+ "loss": 0.6973,
194
+ "step": 170
195
+ },
196
+ {
197
+ "epoch": 0.141287284144427,
198
+ "grad_norm": 5.4375,
199
+ "learning_rate": 3.58e-05,
200
+ "loss": 0.5928,
201
+ "step": 180
202
+ },
203
+ {
204
+ "epoch": 0.14913657770800628,
205
+ "grad_norm": 6.375,
206
+ "learning_rate": 3.7800000000000004e-05,
207
+ "loss": 0.6189,
208
+ "step": 190
209
+ },
210
+ {
211
+ "epoch": 0.15698587127158556,
212
+ "grad_norm": 5.28125,
213
+ "learning_rate": 3.9800000000000005e-05,
214
+ "loss": 0.6661,
215
+ "step": 200
216
+ },
217
+ {
218
+ "epoch": 0.16483516483516483,
219
+ "grad_norm": 4.875,
220
+ "learning_rate": 4.18e-05,
221
+ "loss": 0.634,
222
+ "step": 210
223
+ },
224
+ {
225
+ "epoch": 0.1726844583987441,
226
+ "grad_norm": 6.78125,
227
+ "learning_rate": 4.38e-05,
228
+ "loss": 0.6963,
229
+ "step": 220
230
+ },
231
+ {
232
+ "epoch": 0.18053375196232338,
233
+ "grad_norm": 5.875,
234
+ "learning_rate": 4.58e-05,
235
+ "loss": 0.6744,
236
+ "step": 230
237
+ },
238
+ {
239
+ "epoch": 0.18838304552590268,
240
+ "grad_norm": 5.90625,
241
+ "learning_rate": 4.78e-05,
242
+ "loss": 0.6635,
243
+ "step": 240
244
+ },
245
+ {
246
+ "epoch": 0.19623233908948196,
247
+ "grad_norm": 4.9375,
248
+ "learning_rate": 4.9800000000000004e-05,
249
+ "loss": 0.6874,
250
+ "step": 250
251
+ },
252
+ {
253
+ "epoch": 0.20408163265306123,
254
+ "grad_norm": 5.28125,
255
+ "learning_rate": 5.1800000000000005e-05,
256
+ "loss": 0.6516,
257
+ "step": 260
258
+ },
259
+ {
260
+ "epoch": 0.2119309262166405,
261
+ "grad_norm": 5.40625,
262
+ "learning_rate": 5.380000000000001e-05,
263
+ "loss": 0.7579,
264
+ "step": 270
265
+ },
266
+ {
267
+ "epoch": 0.21978021978021978,
268
+ "grad_norm": 4.15625,
269
+ "learning_rate": 5.580000000000001e-05,
270
+ "loss": 0.6867,
271
+ "step": 280
272
+ },
273
+ {
274
+ "epoch": 0.22762951334379905,
275
+ "grad_norm": 6.59375,
276
+ "learning_rate": 5.7799999999999995e-05,
277
+ "loss": 0.7038,
278
+ "step": 290
279
+ },
280
+ {
281
+ "epoch": 0.23547880690737832,
282
+ "grad_norm": 5.9375,
283
+ "learning_rate": 5.9800000000000003e-05,
284
+ "loss": 0.6681,
285
+ "step": 300
286
+ },
287
+ {
288
+ "epoch": 0.24332810047095763,
289
+ "grad_norm": 4.4375,
290
+ "learning_rate": 6.18e-05,
291
+ "loss": 0.6133,
292
+ "step": 310
293
+ },
294
+ {
295
+ "epoch": 0.25117739403453687,
296
+ "grad_norm": 4.4375,
297
+ "learning_rate": 6.38e-05,
298
+ "loss": 0.6203,
299
+ "step": 320
300
+ },
301
+ {
302
+ "epoch": 0.25902668759811615,
303
+ "grad_norm": 6.3125,
304
+ "learning_rate": 6.58e-05,
305
+ "loss": 0.7164,
306
+ "step": 330
307
+ },
308
+ {
309
+ "epoch": 0.2668759811616955,
310
+ "grad_norm": 4.84375,
311
+ "learning_rate": 6.780000000000001e-05,
312
+ "loss": 0.7157,
313
+ "step": 340
314
+ },
315
+ {
316
+ "epoch": 0.27472527472527475,
317
+ "grad_norm": 5.09375,
318
+ "learning_rate": 6.98e-05,
319
+ "loss": 0.7176,
320
+ "step": 350
321
+ },
322
+ {
323
+ "epoch": 0.282574568288854,
324
+ "grad_norm": 5.1875,
325
+ "learning_rate": 7.18e-05,
326
+ "loss": 0.702,
327
+ "step": 360
328
+ },
329
+ {
330
+ "epoch": 0.2904238618524333,
331
+ "grad_norm": 5.25,
332
+ "learning_rate": 7.38e-05,
333
+ "loss": 0.5834,
334
+ "step": 370
335
+ },
336
+ {
337
+ "epoch": 0.29827315541601257,
338
+ "grad_norm": 5.46875,
339
+ "learning_rate": 7.58e-05,
340
+ "loss": 0.594,
341
+ "step": 380
342
+ },
343
+ {
344
+ "epoch": 0.30612244897959184,
345
+ "grad_norm": 6.625,
346
+ "learning_rate": 7.780000000000001e-05,
347
+ "loss": 0.7667,
348
+ "step": 390
349
+ },
350
+ {
351
+ "epoch": 0.3139717425431711,
352
+ "grad_norm": 6.8125,
353
+ "learning_rate": 7.98e-05,
354
+ "loss": 0.6477,
355
+ "step": 400
356
+ },
357
+ {
358
+ "epoch": 0.3218210361067504,
359
+ "grad_norm": 5.21875,
360
+ "learning_rate": 8.18e-05,
361
+ "loss": 0.6104,
362
+ "step": 410
363
+ },
364
+ {
365
+ "epoch": 0.32967032967032966,
366
+ "grad_norm": 6.03125,
367
+ "learning_rate": 8.38e-05,
368
+ "loss": 0.6879,
369
+ "step": 420
370
+ },
371
+ {
372
+ "epoch": 0.33751962323390894,
373
+ "grad_norm": 5.90625,
374
+ "learning_rate": 8.58e-05,
375
+ "loss": 0.6514,
376
+ "step": 430
377
+ },
378
+ {
379
+ "epoch": 0.3453689167974882,
380
+ "grad_norm": 6.03125,
381
+ "learning_rate": 8.78e-05,
382
+ "loss": 0.649,
383
+ "step": 440
384
+ },
385
+ {
386
+ "epoch": 0.3532182103610675,
387
+ "grad_norm": 4.75,
388
+ "learning_rate": 8.98e-05,
389
+ "loss": 0.6204,
390
+ "step": 450
391
+ },
392
+ {
393
+ "epoch": 0.36106750392464676,
394
+ "grad_norm": 5.125,
395
+ "learning_rate": 9.180000000000001e-05,
396
+ "loss": 0.6633,
397
+ "step": 460
398
+ },
399
+ {
400
+ "epoch": 0.36891679748822603,
401
+ "grad_norm": 4.59375,
402
+ "learning_rate": 9.38e-05,
403
+ "loss": 0.7058,
404
+ "step": 470
405
+ },
406
+ {
407
+ "epoch": 0.37676609105180536,
408
+ "grad_norm": 5.9375,
409
+ "learning_rate": 9.58e-05,
410
+ "loss": 0.5653,
411
+ "step": 480
412
+ },
413
+ {
414
+ "epoch": 0.38461538461538464,
415
+ "grad_norm": 5.6875,
416
+ "learning_rate": 9.78e-05,
417
+ "loss": 0.6143,
418
+ "step": 490
419
+ },
420
+ {
421
+ "epoch": 0.3924646781789639,
422
+ "grad_norm": 5.28125,
423
+ "learning_rate": 9.98e-05,
424
+ "loss": 0.6218,
425
+ "step": 500
426
+ },
427
+ {
428
+ "epoch": 0.4003139717425432,
429
+ "grad_norm": 6.03125,
430
+ "learning_rate": 9.999901304280685e-05,
431
+ "loss": 0.6115,
432
+ "step": 510
433
+ },
434
+ {
435
+ "epoch": 0.40816326530612246,
436
+ "grad_norm": 4.53125,
437
+ "learning_rate": 9.999560138895238e-05,
438
+ "loss": 0.6206,
439
+ "step": 520
440
+ },
441
+ {
442
+ "epoch": 0.41601255886970173,
443
+ "grad_norm": 5.15625,
444
+ "learning_rate": 9.99897530200195e-05,
445
+ "loss": 0.629,
446
+ "step": 530
447
+ },
448
+ {
449
+ "epoch": 0.423861852433281,
450
+ "grad_norm": 4.6875,
451
+ "learning_rate": 9.998146822104943e-05,
452
+ "loss": 0.6492,
453
+ "step": 540
454
+ },
455
+ {
456
+ "epoch": 0.4317111459968603,
457
+ "grad_norm": 5.1875,
458
+ "learning_rate": 9.997074739583162e-05,
459
+ "loss": 0.585,
460
+ "step": 550
461
+ },
462
+ {
463
+ "epoch": 0.43956043956043955,
464
+ "grad_norm": 7.1875,
465
+ "learning_rate": 9.995759106688393e-05,
466
+ "loss": 0.558,
467
+ "step": 560
468
+ },
469
+ {
470
+ "epoch": 0.4474097331240188,
471
+ "grad_norm": 5.28125,
472
+ "learning_rate": 9.99419998754273e-05,
473
+ "loss": 0.6242,
474
+ "step": 570
475
+ },
476
+ {
477
+ "epoch": 0.4552590266875981,
478
+ "grad_norm": 4.53125,
479
+ "learning_rate": 9.992397458135438e-05,
480
+ "loss": 0.682,
481
+ "step": 580
482
+ },
483
+ {
484
+ "epoch": 0.4631083202511774,
485
+ "grad_norm": 4.3125,
486
+ "learning_rate": 9.990351606319261e-05,
487
+ "loss": 0.6194,
488
+ "step": 590
489
+ },
490
+ {
491
+ "epoch": 0.47095761381475665,
492
+ "grad_norm": 5.90625,
493
+ "learning_rate": 9.988062531806126e-05,
494
+ "loss": 0.6211,
495
+ "step": 600
496
+ },
497
+ {
498
+ "epoch": 0.478806907378336,
499
+ "grad_norm": 4.5625,
500
+ "learning_rate": 9.985530346162299e-05,
501
+ "loss": 0.617,
502
+ "step": 610
503
+ },
504
+ {
505
+ "epoch": 0.48665620094191525,
506
+ "grad_norm": 5.125,
507
+ "learning_rate": 9.982755172802933e-05,
508
+ "loss": 0.5901,
509
+ "step": 620
510
+ },
511
+ {
512
+ "epoch": 0.4945054945054945,
513
+ "grad_norm": 4.15625,
514
+ "learning_rate": 9.979737146986064e-05,
515
+ "loss": 0.6067,
516
+ "step": 630
517
+ },
518
+ {
519
+ "epoch": 0.5023547880690737,
520
+ "grad_norm": 5.4375,
521
+ "learning_rate": 9.976476415806013e-05,
522
+ "loss": 0.5976,
523
+ "step": 640
524
+ },
525
+ {
526
+ "epoch": 0.5102040816326531,
527
+ "grad_norm": 5.75,
528
+ "learning_rate": 9.972973138186217e-05,
529
+ "loss": 0.6467,
530
+ "step": 650
531
+ },
532
+ {
533
+ "epoch": 0.5180533751962323,
534
+ "grad_norm": 7.25,
535
+ "learning_rate": 9.969227484871484e-05,
536
+ "loss": 0.5904,
537
+ "step": 660
538
+ },
539
+ {
540
+ "epoch": 0.5259026687598116,
541
+ "grad_norm": 4.65625,
542
+ "learning_rate": 9.965239638419672e-05,
543
+ "loss": 0.6814,
544
+ "step": 670
545
+ },
546
+ {
547
+ "epoch": 0.533751962323391,
548
+ "grad_norm": 6.125,
549
+ "learning_rate": 9.961009793192792e-05,
550
+ "loss": 0.6684,
551
+ "step": 680
552
+ },
553
+ {
554
+ "epoch": 0.5416012558869702,
555
+ "grad_norm": 4.40625,
556
+ "learning_rate": 9.956538155347534e-05,
557
+ "loss": 0.6206,
558
+ "step": 690
559
+ },
560
+ {
561
+ "epoch": 0.5494505494505495,
562
+ "grad_norm": 5.875,
563
+ "learning_rate": 9.951824942825215e-05,
564
+ "loss": 0.6096,
565
+ "step": 700
566
+ },
567
+ {
568
+ "epoch": 0.5572998430141287,
569
+ "grad_norm": 5.71875,
570
+ "learning_rate": 9.946870385341167e-05,
571
+ "loss": 0.5468,
572
+ "step": 710
573
+ },
574
+ {
575
+ "epoch": 0.565149136577708,
576
+ "grad_norm": 4.84375,
577
+ "learning_rate": 9.941674724373531e-05,
578
+ "loss": 0.5883,
579
+ "step": 720
580
+ },
581
+ {
582
+ "epoch": 0.5729984301412873,
583
+ "grad_norm": 4.75,
584
+ "learning_rate": 9.936238213151491e-05,
585
+ "loss": 0.596,
586
+ "step": 730
587
+ },
588
+ {
589
+ "epoch": 0.5808477237048666,
590
+ "grad_norm": 4.9375,
591
+ "learning_rate": 9.930561116642935e-05,
592
+ "loss": 0.5178,
593
+ "step": 740
594
+ },
595
+ {
596
+ "epoch": 0.5886970172684458,
597
+ "grad_norm": 5.59375,
598
+ "learning_rate": 9.924643711541539e-05,
599
+ "loss": 0.6438,
600
+ "step": 750
601
+ },
602
+ {
603
+ "epoch": 0.5965463108320251,
604
+ "grad_norm": 4.96875,
605
+ "learning_rate": 9.918486286253279e-05,
606
+ "loss": 0.6975,
607
+ "step": 760
608
+ },
609
+ {
610
+ "epoch": 0.6043956043956044,
611
+ "grad_norm": 4.375,
612
+ "learning_rate": 9.912089140882376e-05,
613
+ "loss": 0.5828,
614
+ "step": 770
615
+ },
616
+ {
617
+ "epoch": 0.6122448979591837,
618
+ "grad_norm": 7.0625,
619
+ "learning_rate": 9.90545258721667e-05,
620
+ "loss": 0.5019,
621
+ "step": 780
622
+ },
623
+ {
624
+ "epoch": 0.6200941915227629,
625
+ "grad_norm": 6.75,
626
+ "learning_rate": 9.898576948712426e-05,
627
+ "loss": 0.6489,
628
+ "step": 790
629
+ },
630
+ {
631
+ "epoch": 0.6279434850863422,
632
+ "grad_norm": 5.28125,
633
+ "learning_rate": 9.891462560478562e-05,
634
+ "loss": 0.6364,
635
+ "step": 800
636
+ },
637
+ {
638
+ "epoch": 0.6357927786499215,
639
+ "grad_norm": 4.6875,
640
+ "learning_rate": 9.884109769260325e-05,
641
+ "loss": 0.5458,
642
+ "step": 810
643
+ },
644
+ {
645
+ "epoch": 0.6436420722135008,
646
+ "grad_norm": 3.703125,
647
+ "learning_rate": 9.876518933422386e-05,
648
+ "loss": 0.5859,
649
+ "step": 820
650
+ },
651
+ {
652
+ "epoch": 0.6514913657770801,
653
+ "grad_norm": 4.375,
654
+ "learning_rate": 9.868690422931372e-05,
655
+ "loss": 0.6369,
656
+ "step": 830
657
+ },
658
+ {
659
+ "epoch": 0.6593406593406593,
660
+ "grad_norm": 5.40625,
661
+ "learning_rate": 9.860624619337844e-05,
662
+ "loss": 0.7126,
663
+ "step": 840
664
+ },
665
+ {
666
+ "epoch": 0.6671899529042387,
667
+ "grad_norm": 4.9375,
668
+ "learning_rate": 9.852321915757687e-05,
669
+ "loss": 0.5857,
670
+ "step": 850
671
+ },
672
+ {
673
+ "epoch": 0.6750392464678179,
674
+ "grad_norm": 5.0,
675
+ "learning_rate": 9.843782716852963e-05,
676
+ "loss": 0.645,
677
+ "step": 860
678
+ },
679
+ {
680
+ "epoch": 0.6828885400313972,
681
+ "grad_norm": 3.78125,
682
+ "learning_rate": 9.835007438812177e-05,
683
+ "loss": 0.6538,
684
+ "step": 870
685
+ },
686
+ {
687
+ "epoch": 0.6907378335949764,
688
+ "grad_norm": 5.15625,
689
+ "learning_rate": 9.82599650933e-05,
690
+ "loss": 0.6045,
691
+ "step": 880
692
+ },
693
+ {
694
+ "epoch": 0.6985871271585558,
695
+ "grad_norm": 3.59375,
696
+ "learning_rate": 9.816750367586425e-05,
697
+ "loss": 0.6282,
698
+ "step": 890
699
+ },
700
+ {
701
+ "epoch": 0.706436420722135,
702
+ "grad_norm": 5.1875,
703
+ "learning_rate": 9.807269464225355e-05,
704
+ "loss": 0.5509,
705
+ "step": 900
706
+ },
707
+ {
708
+ "epoch": 0.7142857142857143,
709
+ "grad_norm": 7.125,
710
+ "learning_rate": 9.797554261332645e-05,
711
+ "loss": 0.6392,
712
+ "step": 910
713
+ },
714
+ {
715
+ "epoch": 0.7221350078492935,
716
+ "grad_norm": 5.34375,
717
+ "learning_rate": 9.787605232413574e-05,
718
+ "loss": 0.6152,
719
+ "step": 920
720
+ },
721
+ {
722
+ "epoch": 0.7299843014128728,
723
+ "grad_norm": 4.28125,
724
+ "learning_rate": 9.777422862369783e-05,
725
+ "loss": 0.7706,
726
+ "step": 930
727
+ },
728
+ {
729
+ "epoch": 0.7378335949764521,
730
+ "grad_norm": 4.59375,
731
+ "learning_rate": 9.767007647475617e-05,
732
+ "loss": 0.6656,
733
+ "step": 940
734
+ },
735
+ {
736
+ "epoch": 0.7456828885400314,
737
+ "grad_norm": 4.53125,
738
+ "learning_rate": 9.756360095353956e-05,
739
+ "loss": 0.6592,
740
+ "step": 950
741
+ },
742
+ {
743
+ "epoch": 0.7535321821036107,
744
+ "grad_norm": 4.40625,
745
+ "learning_rate": 9.745480724951473e-05,
746
+ "loss": 0.6141,
747
+ "step": 960
748
+ },
749
+ {
750
+ "epoch": 0.7613814756671899,
751
+ "grad_norm": 4.96875,
752
+ "learning_rate": 9.73437006651333e-05,
753
+ "loss": 0.6161,
754
+ "step": 970
755
+ },
756
+ {
757
+ "epoch": 0.7692307692307693,
758
+ "grad_norm": 4.25,
759
+ "learning_rate": 9.723028661557345e-05,
760
+ "loss": 0.7166,
761
+ "step": 980
762
+ },
763
+ {
764
+ "epoch": 0.7770800627943485,
765
+ "grad_norm": 5.34375,
766
+ "learning_rate": 9.711457062847595e-05,
767
+ "loss": 0.6028,
768
+ "step": 990
769
+ },
770
+ {
771
+ "epoch": 0.7849293563579278,
772
+ "grad_norm": 4.34375,
773
+ "learning_rate": 9.699655834367478e-05,
774
+ "loss": 0.5262,
775
+ "step": 1000
776
+ }
777
+ ],
778
+ "logging_steps": 10,
779
+ "max_steps": 5000,
780
+ "num_input_tokens_seen": 0,
781
+ "num_train_epochs": 4,
782
+ "save_steps": 1000,
783
+ "stateful_callbacks": {
784
+ "TrainerControl": {
785
+ "args": {
786
+ "should_epoch_stop": false,
787
+ "should_evaluate": false,
788
+ "should_log": false,
789
+ "should_save": true,
790
+ "should_training_stop": false
791
+ },
792
+ "attributes": {}
793
+ }
794
+ },
795
+ "total_flos": 0.0,
796
+ "train_batch_size": 16,
797
+ "trial_name": null,
798
+ "trial_params": null
799
+ }
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-1000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9365177436ad70bb9d6000d96d3ccdc5c5ca38536defdcfec6b548401edbc07f
3
+ size 6481
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "gradient_checkpointing": false,
20
+ "hidden_activation": "gelu",
21
+ "hidden_size": 768,
22
+ "initializer_cutoff_factor": 2.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 1152,
25
+ "layer_norm_eps": 1e-05,
26
+ "local_attention": 128,
27
+ "local_rope_theta": 10000.0,
28
+ "max_position_embeddings": 8192,
29
+ "mlp_bias": false,
30
+ "mlp_dropout": 0.0,
31
+ "model_type": "modernbert",
32
+ "norm_bias": false,
33
+ "norm_eps": 1e-05,
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 22,
36
+ "pad_token_id": 50283,
37
+ "position_embedding_type": "absolute",
38
+ "repad_logits_with_grad": false,
39
+ "sep_token_id": 50282,
40
+ "sparse_pred_ignore_index": -100,
41
+ "sparse_prediction": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.53.0",
44
+ "vocab_size": 50368
45
+ }
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf0a75f2dc3580598001eea2e77afba906bc5adc06f1a5696547fcb216fa27a6
3
+ size 298041696
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f2241ba51f63d9e2dc7bf192d7e5d8fffa9c1d8416702001aa9bccb56377121
3
+ size 596170443
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18bb6b6e7da6250d22c25b95c64287be4aea598d9d97ef67b0fd69b9be869ed7
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13db4dd10b22b5464f5ef0b539152b46390eae6a827b355108cb54b1af2eb830
3
+ size 15429
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06e293f13fb6e873e8081636b6b07115410b397c55f1ed08b8772b059c8eb74e
3
+ size 15429
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c595c077664f94b7f91088561951be20e39483e01865774a83bc46f3971a3152
3
+ size 15429
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff65ca8b6e225d1d82cc63887ee464e8c896624115a4dc693eae1fff20815699
3
+ size 1465
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/trainer_state.json ADDED
@@ -0,0 +1,1564 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 1.5698587127158556,
6
+ "eval_steps": 1000.0,
7
+ "global_step": 2000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0,
14
+ "eval/nano_beir.arguana": 0.4400549254974839,
15
+ "step": 0
16
+ },
17
+ {
18
+ "epoch": 0,
19
+ "eval/nano_beir.climate_fever": 0.22880053057901228,
20
+ "step": 0
21
+ },
22
+ {
23
+ "epoch": 0,
24
+ "eval/nano_beir.dbpedia_entity": 0.30026716538531234,
25
+ "step": 0
26
+ },
27
+ {
28
+ "epoch": 0,
29
+ "eval/nano_beir.fever": 0.6419178788751609,
30
+ "step": 0
31
+ },
32
+ {
33
+ "epoch": 0,
34
+ "eval/nano_beir.fiqa": 0.4370593595651809,
35
+ "step": 0
36
+ },
37
+ {
38
+ "epoch": 0,
39
+ "eval/nano_beir.hotpotqa": 0.6723047477224562,
40
+ "step": 0
41
+ },
42
+ {
43
+ "epoch": 0,
44
+ "eval/nano_beir.nfcorpus": 0.3437446614516137,
45
+ "step": 0
46
+ },
47
+ {
48
+ "epoch": 0,
49
+ "eval/nano_beir.nq": 0.436711543163819,
50
+ "step": 0
51
+ },
52
+ {
53
+ "epoch": 0,
54
+ "eval/nano_beir.quora": 0.8439997970269911,
55
+ "step": 0
56
+ },
57
+ {
58
+ "epoch": 0,
59
+ "eval/nano_beir.scidocs": 0.26304331393922176,
60
+ "step": 0
61
+ },
62
+ {
63
+ "epoch": 0,
64
+ "eval/nano_beir.scifact": 0.7599341408912704,
65
+ "step": 0
66
+ },
67
+ {
68
+ "epoch": 0,
69
+ "eval/nano_beir.webis_touche2020": 0.29696490432942557,
70
+ "step": 0
71
+ },
72
+ {
73
+ "epoch": 0,
74
+ "eval/avg": 0.47206691403557904,
75
+ "step": 0
76
+ },
77
+ {
78
+ "epoch": 0.007849293563579277,
79
+ "grad_norm": 8.25,
80
+ "learning_rate": 1.8e-06,
81
+ "loss": 0.7314,
82
+ "step": 10
83
+ },
84
+ {
85
+ "epoch": 0.015698587127158554,
86
+ "grad_norm": 7.03125,
87
+ "learning_rate": 3.8e-06,
88
+ "loss": 0.8796,
89
+ "step": 20
90
+ },
91
+ {
92
+ "epoch": 0.023547880690737835,
93
+ "grad_norm": 7.3125,
94
+ "learning_rate": 5.8e-06,
95
+ "loss": 0.7311,
96
+ "step": 30
97
+ },
98
+ {
99
+ "epoch": 0.03139717425431711,
100
+ "grad_norm": 7.15625,
101
+ "learning_rate": 7.8e-06,
102
+ "loss": 0.8214,
103
+ "step": 40
104
+ },
105
+ {
106
+ "epoch": 0.03924646781789639,
107
+ "grad_norm": 9.1875,
108
+ "learning_rate": 9.800000000000001e-06,
109
+ "loss": 0.8489,
110
+ "step": 50
111
+ },
112
+ {
113
+ "epoch": 0.04709576138147567,
114
+ "grad_norm": 6.84375,
115
+ "learning_rate": 1.18e-05,
116
+ "loss": 0.7661,
117
+ "step": 60
118
+ },
119
+ {
120
+ "epoch": 0.054945054945054944,
121
+ "grad_norm": 6.6875,
122
+ "learning_rate": 1.3800000000000002e-05,
123
+ "loss": 0.7525,
124
+ "step": 70
125
+ },
126
+ {
127
+ "epoch": 0.06279434850863422,
128
+ "grad_norm": 7.5625,
129
+ "learning_rate": 1.58e-05,
130
+ "loss": 0.8653,
131
+ "step": 80
132
+ },
133
+ {
134
+ "epoch": 0.0706436420722135,
135
+ "grad_norm": 6.25,
136
+ "learning_rate": 1.78e-05,
137
+ "loss": 0.7607,
138
+ "step": 90
139
+ },
140
+ {
141
+ "epoch": 0.07849293563579278,
142
+ "grad_norm": 5.28125,
143
+ "learning_rate": 1.9800000000000004e-05,
144
+ "loss": 0.7259,
145
+ "step": 100
146
+ },
147
+ {
148
+ "epoch": 0.08634222919937205,
149
+ "grad_norm": 5.59375,
150
+ "learning_rate": 2.18e-05,
151
+ "loss": 0.7862,
152
+ "step": 110
153
+ },
154
+ {
155
+ "epoch": 0.09419152276295134,
156
+ "grad_norm": 5.8125,
157
+ "learning_rate": 2.38e-05,
158
+ "loss": 0.8212,
159
+ "step": 120
160
+ },
161
+ {
162
+ "epoch": 0.10204081632653061,
163
+ "grad_norm": 6.0,
164
+ "learning_rate": 2.58e-05,
165
+ "loss": 0.6884,
166
+ "step": 130
167
+ },
168
+ {
169
+ "epoch": 0.10989010989010989,
170
+ "grad_norm": 5.9375,
171
+ "learning_rate": 2.7800000000000005e-05,
172
+ "loss": 0.712,
173
+ "step": 140
174
+ },
175
+ {
176
+ "epoch": 0.11773940345368916,
177
+ "grad_norm": 5.53125,
178
+ "learning_rate": 2.98e-05,
179
+ "loss": 0.6231,
180
+ "step": 150
181
+ },
182
+ {
183
+ "epoch": 0.12558869701726844,
184
+ "grad_norm": 5.59375,
185
+ "learning_rate": 3.18e-05,
186
+ "loss": 0.6733,
187
+ "step": 160
188
+ },
189
+ {
190
+ "epoch": 0.13343799058084774,
191
+ "grad_norm": 6.84375,
192
+ "learning_rate": 3.38e-05,
193
+ "loss": 0.6973,
194
+ "step": 170
195
+ },
196
+ {
197
+ "epoch": 0.141287284144427,
198
+ "grad_norm": 5.4375,
199
+ "learning_rate": 3.58e-05,
200
+ "loss": 0.5928,
201
+ "step": 180
202
+ },
203
+ {
204
+ "epoch": 0.14913657770800628,
205
+ "grad_norm": 6.375,
206
+ "learning_rate": 3.7800000000000004e-05,
207
+ "loss": 0.6189,
208
+ "step": 190
209
+ },
210
+ {
211
+ "epoch": 0.15698587127158556,
212
+ "grad_norm": 5.28125,
213
+ "learning_rate": 3.9800000000000005e-05,
214
+ "loss": 0.6661,
215
+ "step": 200
216
+ },
217
+ {
218
+ "epoch": 0.16483516483516483,
219
+ "grad_norm": 4.875,
220
+ "learning_rate": 4.18e-05,
221
+ "loss": 0.634,
222
+ "step": 210
223
+ },
224
+ {
225
+ "epoch": 0.1726844583987441,
226
+ "grad_norm": 6.78125,
227
+ "learning_rate": 4.38e-05,
228
+ "loss": 0.6963,
229
+ "step": 220
230
+ },
231
+ {
232
+ "epoch": 0.18053375196232338,
233
+ "grad_norm": 5.875,
234
+ "learning_rate": 4.58e-05,
235
+ "loss": 0.6744,
236
+ "step": 230
237
+ },
238
+ {
239
+ "epoch": 0.18838304552590268,
240
+ "grad_norm": 5.90625,
241
+ "learning_rate": 4.78e-05,
242
+ "loss": 0.6635,
243
+ "step": 240
244
+ },
245
+ {
246
+ "epoch": 0.19623233908948196,
247
+ "grad_norm": 4.9375,
248
+ "learning_rate": 4.9800000000000004e-05,
249
+ "loss": 0.6874,
250
+ "step": 250
251
+ },
252
+ {
253
+ "epoch": 0.20408163265306123,
254
+ "grad_norm": 5.28125,
255
+ "learning_rate": 5.1800000000000005e-05,
256
+ "loss": 0.6516,
257
+ "step": 260
258
+ },
259
+ {
260
+ "epoch": 0.2119309262166405,
261
+ "grad_norm": 5.40625,
262
+ "learning_rate": 5.380000000000001e-05,
263
+ "loss": 0.7579,
264
+ "step": 270
265
+ },
266
+ {
267
+ "epoch": 0.21978021978021978,
268
+ "grad_norm": 4.15625,
269
+ "learning_rate": 5.580000000000001e-05,
270
+ "loss": 0.6867,
271
+ "step": 280
272
+ },
273
+ {
274
+ "epoch": 0.22762951334379905,
275
+ "grad_norm": 6.59375,
276
+ "learning_rate": 5.7799999999999995e-05,
277
+ "loss": 0.7038,
278
+ "step": 290
279
+ },
280
+ {
281
+ "epoch": 0.23547880690737832,
282
+ "grad_norm": 5.9375,
283
+ "learning_rate": 5.9800000000000003e-05,
284
+ "loss": 0.6681,
285
+ "step": 300
286
+ },
287
+ {
288
+ "epoch": 0.24332810047095763,
289
+ "grad_norm": 4.4375,
290
+ "learning_rate": 6.18e-05,
291
+ "loss": 0.6133,
292
+ "step": 310
293
+ },
294
+ {
295
+ "epoch": 0.25117739403453687,
296
+ "grad_norm": 4.4375,
297
+ "learning_rate": 6.38e-05,
298
+ "loss": 0.6203,
299
+ "step": 320
300
+ },
301
+ {
302
+ "epoch": 0.25902668759811615,
303
+ "grad_norm": 6.3125,
304
+ "learning_rate": 6.58e-05,
305
+ "loss": 0.7164,
306
+ "step": 330
307
+ },
308
+ {
309
+ "epoch": 0.2668759811616955,
310
+ "grad_norm": 4.84375,
311
+ "learning_rate": 6.780000000000001e-05,
312
+ "loss": 0.7157,
313
+ "step": 340
314
+ },
315
+ {
316
+ "epoch": 0.27472527472527475,
317
+ "grad_norm": 5.09375,
318
+ "learning_rate": 6.98e-05,
319
+ "loss": 0.7176,
320
+ "step": 350
321
+ },
322
+ {
323
+ "epoch": 0.282574568288854,
324
+ "grad_norm": 5.1875,
325
+ "learning_rate": 7.18e-05,
326
+ "loss": 0.702,
327
+ "step": 360
328
+ },
329
+ {
330
+ "epoch": 0.2904238618524333,
331
+ "grad_norm": 5.25,
332
+ "learning_rate": 7.38e-05,
333
+ "loss": 0.5834,
334
+ "step": 370
335
+ },
336
+ {
337
+ "epoch": 0.29827315541601257,
338
+ "grad_norm": 5.46875,
339
+ "learning_rate": 7.58e-05,
340
+ "loss": 0.594,
341
+ "step": 380
342
+ },
343
+ {
344
+ "epoch": 0.30612244897959184,
345
+ "grad_norm": 6.625,
346
+ "learning_rate": 7.780000000000001e-05,
347
+ "loss": 0.7667,
348
+ "step": 390
349
+ },
350
+ {
351
+ "epoch": 0.3139717425431711,
352
+ "grad_norm": 6.8125,
353
+ "learning_rate": 7.98e-05,
354
+ "loss": 0.6477,
355
+ "step": 400
356
+ },
357
+ {
358
+ "epoch": 0.3218210361067504,
359
+ "grad_norm": 5.21875,
360
+ "learning_rate": 8.18e-05,
361
+ "loss": 0.6104,
362
+ "step": 410
363
+ },
364
+ {
365
+ "epoch": 0.32967032967032966,
366
+ "grad_norm": 6.03125,
367
+ "learning_rate": 8.38e-05,
368
+ "loss": 0.6879,
369
+ "step": 420
370
+ },
371
+ {
372
+ "epoch": 0.33751962323390894,
373
+ "grad_norm": 5.90625,
374
+ "learning_rate": 8.58e-05,
375
+ "loss": 0.6514,
376
+ "step": 430
377
+ },
378
+ {
379
+ "epoch": 0.3453689167974882,
380
+ "grad_norm": 6.03125,
381
+ "learning_rate": 8.78e-05,
382
+ "loss": 0.649,
383
+ "step": 440
384
+ },
385
+ {
386
+ "epoch": 0.3532182103610675,
387
+ "grad_norm": 4.75,
388
+ "learning_rate": 8.98e-05,
389
+ "loss": 0.6204,
390
+ "step": 450
391
+ },
392
+ {
393
+ "epoch": 0.36106750392464676,
394
+ "grad_norm": 5.125,
395
+ "learning_rate": 9.180000000000001e-05,
396
+ "loss": 0.6633,
397
+ "step": 460
398
+ },
399
+ {
400
+ "epoch": 0.36891679748822603,
401
+ "grad_norm": 4.59375,
402
+ "learning_rate": 9.38e-05,
403
+ "loss": 0.7058,
404
+ "step": 470
405
+ },
406
+ {
407
+ "epoch": 0.37676609105180536,
408
+ "grad_norm": 5.9375,
409
+ "learning_rate": 9.58e-05,
410
+ "loss": 0.5653,
411
+ "step": 480
412
+ },
413
+ {
414
+ "epoch": 0.38461538461538464,
415
+ "grad_norm": 5.6875,
416
+ "learning_rate": 9.78e-05,
417
+ "loss": 0.6143,
418
+ "step": 490
419
+ },
420
+ {
421
+ "epoch": 0.3924646781789639,
422
+ "grad_norm": 5.28125,
423
+ "learning_rate": 9.98e-05,
424
+ "loss": 0.6218,
425
+ "step": 500
426
+ },
427
+ {
428
+ "epoch": 0.4003139717425432,
429
+ "grad_norm": 6.03125,
430
+ "learning_rate": 9.999901304280685e-05,
431
+ "loss": 0.6115,
432
+ "step": 510
433
+ },
434
+ {
435
+ "epoch": 0.40816326530612246,
436
+ "grad_norm": 4.53125,
437
+ "learning_rate": 9.999560138895238e-05,
438
+ "loss": 0.6206,
439
+ "step": 520
440
+ },
441
+ {
442
+ "epoch": 0.41601255886970173,
443
+ "grad_norm": 5.15625,
444
+ "learning_rate": 9.99897530200195e-05,
445
+ "loss": 0.629,
446
+ "step": 530
447
+ },
448
+ {
449
+ "epoch": 0.423861852433281,
450
+ "grad_norm": 4.6875,
451
+ "learning_rate": 9.998146822104943e-05,
452
+ "loss": 0.6492,
453
+ "step": 540
454
+ },
455
+ {
456
+ "epoch": 0.4317111459968603,
457
+ "grad_norm": 5.1875,
458
+ "learning_rate": 9.997074739583162e-05,
459
+ "loss": 0.585,
460
+ "step": 550
461
+ },
462
+ {
463
+ "epoch": 0.43956043956043955,
464
+ "grad_norm": 7.1875,
465
+ "learning_rate": 9.995759106688393e-05,
466
+ "loss": 0.558,
467
+ "step": 560
468
+ },
469
+ {
470
+ "epoch": 0.4474097331240188,
471
+ "grad_norm": 5.28125,
472
+ "learning_rate": 9.99419998754273e-05,
473
+ "loss": 0.6242,
474
+ "step": 570
475
+ },
476
+ {
477
+ "epoch": 0.4552590266875981,
478
+ "grad_norm": 4.53125,
479
+ "learning_rate": 9.992397458135438e-05,
480
+ "loss": 0.682,
481
+ "step": 580
482
+ },
483
+ {
484
+ "epoch": 0.4631083202511774,
485
+ "grad_norm": 4.3125,
486
+ "learning_rate": 9.990351606319261e-05,
487
+ "loss": 0.6194,
488
+ "step": 590
489
+ },
490
+ {
491
+ "epoch": 0.47095761381475665,
492
+ "grad_norm": 5.90625,
493
+ "learning_rate": 9.988062531806126e-05,
494
+ "loss": 0.6211,
495
+ "step": 600
496
+ },
497
+ {
498
+ "epoch": 0.478806907378336,
499
+ "grad_norm": 4.5625,
500
+ "learning_rate": 9.985530346162299e-05,
501
+ "loss": 0.617,
502
+ "step": 610
503
+ },
504
+ {
505
+ "epoch": 0.48665620094191525,
506
+ "grad_norm": 5.125,
507
+ "learning_rate": 9.982755172802933e-05,
508
+ "loss": 0.5901,
509
+ "step": 620
510
+ },
511
+ {
512
+ "epoch": 0.4945054945054945,
513
+ "grad_norm": 4.15625,
514
+ "learning_rate": 9.979737146986064e-05,
515
+ "loss": 0.6067,
516
+ "step": 630
517
+ },
518
+ {
519
+ "epoch": 0.5023547880690737,
520
+ "grad_norm": 5.4375,
521
+ "learning_rate": 9.976476415806013e-05,
522
+ "loss": 0.5976,
523
+ "step": 640
524
+ },
525
+ {
526
+ "epoch": 0.5102040816326531,
527
+ "grad_norm": 5.75,
528
+ "learning_rate": 9.972973138186217e-05,
529
+ "loss": 0.6467,
530
+ "step": 650
531
+ },
532
+ {
533
+ "epoch": 0.5180533751962323,
534
+ "grad_norm": 7.25,
535
+ "learning_rate": 9.969227484871484e-05,
536
+ "loss": 0.5904,
537
+ "step": 660
538
+ },
539
+ {
540
+ "epoch": 0.5259026687598116,
541
+ "grad_norm": 4.65625,
542
+ "learning_rate": 9.965239638419672e-05,
543
+ "loss": 0.6814,
544
+ "step": 670
545
+ },
546
+ {
547
+ "epoch": 0.533751962323391,
548
+ "grad_norm": 6.125,
549
+ "learning_rate": 9.961009793192792e-05,
550
+ "loss": 0.6684,
551
+ "step": 680
552
+ },
553
+ {
554
+ "epoch": 0.5416012558869702,
555
+ "grad_norm": 4.40625,
556
+ "learning_rate": 9.956538155347534e-05,
557
+ "loss": 0.6206,
558
+ "step": 690
559
+ },
560
+ {
561
+ "epoch": 0.5494505494505495,
562
+ "grad_norm": 5.875,
563
+ "learning_rate": 9.951824942825215e-05,
564
+ "loss": 0.6096,
565
+ "step": 700
566
+ },
567
+ {
568
+ "epoch": 0.5572998430141287,
569
+ "grad_norm": 5.71875,
570
+ "learning_rate": 9.946870385341167e-05,
571
+ "loss": 0.5468,
572
+ "step": 710
573
+ },
574
+ {
575
+ "epoch": 0.565149136577708,
576
+ "grad_norm": 4.84375,
577
+ "learning_rate": 9.941674724373531e-05,
578
+ "loss": 0.5883,
579
+ "step": 720
580
+ },
581
+ {
582
+ "epoch": 0.5729984301412873,
583
+ "grad_norm": 4.75,
584
+ "learning_rate": 9.936238213151491e-05,
585
+ "loss": 0.596,
586
+ "step": 730
587
+ },
588
+ {
589
+ "epoch": 0.5808477237048666,
590
+ "grad_norm": 4.9375,
591
+ "learning_rate": 9.930561116642935e-05,
592
+ "loss": 0.5178,
593
+ "step": 740
594
+ },
595
+ {
596
+ "epoch": 0.5886970172684458,
597
+ "grad_norm": 5.59375,
598
+ "learning_rate": 9.924643711541539e-05,
599
+ "loss": 0.6438,
600
+ "step": 750
601
+ },
602
+ {
603
+ "epoch": 0.5965463108320251,
604
+ "grad_norm": 4.96875,
605
+ "learning_rate": 9.918486286253279e-05,
606
+ "loss": 0.6975,
607
+ "step": 760
608
+ },
609
+ {
610
+ "epoch": 0.6043956043956044,
611
+ "grad_norm": 4.375,
612
+ "learning_rate": 9.912089140882376e-05,
613
+ "loss": 0.5828,
614
+ "step": 770
615
+ },
616
+ {
617
+ "epoch": 0.6122448979591837,
618
+ "grad_norm": 7.0625,
619
+ "learning_rate": 9.90545258721667e-05,
620
+ "loss": 0.5019,
621
+ "step": 780
622
+ },
623
+ {
624
+ "epoch": 0.6200941915227629,
625
+ "grad_norm": 6.75,
626
+ "learning_rate": 9.898576948712426e-05,
627
+ "loss": 0.6489,
628
+ "step": 790
629
+ },
630
+ {
631
+ "epoch": 0.6279434850863422,
632
+ "grad_norm": 5.28125,
633
+ "learning_rate": 9.891462560478562e-05,
634
+ "loss": 0.6364,
635
+ "step": 800
636
+ },
637
+ {
638
+ "epoch": 0.6357927786499215,
639
+ "grad_norm": 4.6875,
640
+ "learning_rate": 9.884109769260325e-05,
641
+ "loss": 0.5458,
642
+ "step": 810
643
+ },
644
+ {
645
+ "epoch": 0.6436420722135008,
646
+ "grad_norm": 3.703125,
647
+ "learning_rate": 9.876518933422386e-05,
648
+ "loss": 0.5859,
649
+ "step": 820
650
+ },
651
+ {
652
+ "epoch": 0.6514913657770801,
653
+ "grad_norm": 4.375,
654
+ "learning_rate": 9.868690422931372e-05,
655
+ "loss": 0.6369,
656
+ "step": 830
657
+ },
658
+ {
659
+ "epoch": 0.6593406593406593,
660
+ "grad_norm": 5.40625,
661
+ "learning_rate": 9.860624619337844e-05,
662
+ "loss": 0.7126,
663
+ "step": 840
664
+ },
665
+ {
666
+ "epoch": 0.6671899529042387,
667
+ "grad_norm": 4.9375,
668
+ "learning_rate": 9.852321915757687e-05,
669
+ "loss": 0.5857,
670
+ "step": 850
671
+ },
672
+ {
673
+ "epoch": 0.6750392464678179,
674
+ "grad_norm": 5.0,
675
+ "learning_rate": 9.843782716852963e-05,
676
+ "loss": 0.645,
677
+ "step": 860
678
+ },
679
+ {
680
+ "epoch": 0.6828885400313972,
681
+ "grad_norm": 3.78125,
682
+ "learning_rate": 9.835007438812177e-05,
683
+ "loss": 0.6538,
684
+ "step": 870
685
+ },
686
+ {
687
+ "epoch": 0.6907378335949764,
688
+ "grad_norm": 5.15625,
689
+ "learning_rate": 9.82599650933e-05,
690
+ "loss": 0.6045,
691
+ "step": 880
692
+ },
693
+ {
694
+ "epoch": 0.6985871271585558,
695
+ "grad_norm": 3.59375,
696
+ "learning_rate": 9.816750367586425e-05,
697
+ "loss": 0.6282,
698
+ "step": 890
699
+ },
700
+ {
701
+ "epoch": 0.706436420722135,
702
+ "grad_norm": 5.1875,
703
+ "learning_rate": 9.807269464225355e-05,
704
+ "loss": 0.5509,
705
+ "step": 900
706
+ },
707
+ {
708
+ "epoch": 0.7142857142857143,
709
+ "grad_norm": 7.125,
710
+ "learning_rate": 9.797554261332645e-05,
711
+ "loss": 0.6392,
712
+ "step": 910
713
+ },
714
+ {
715
+ "epoch": 0.7221350078492935,
716
+ "grad_norm": 5.34375,
717
+ "learning_rate": 9.787605232413574e-05,
718
+ "loss": 0.6152,
719
+ "step": 920
720
+ },
721
+ {
722
+ "epoch": 0.7299843014128728,
723
+ "grad_norm": 4.28125,
724
+ "learning_rate": 9.777422862369783e-05,
725
+ "loss": 0.7706,
726
+ "step": 930
727
+ },
728
+ {
729
+ "epoch": 0.7378335949764521,
730
+ "grad_norm": 4.59375,
731
+ "learning_rate": 9.767007647475617e-05,
732
+ "loss": 0.6656,
733
+ "step": 940
734
+ },
735
+ {
736
+ "epoch": 0.7456828885400314,
737
+ "grad_norm": 4.53125,
738
+ "learning_rate": 9.756360095353956e-05,
739
+ "loss": 0.6592,
740
+ "step": 950
741
+ },
742
+ {
743
+ "epoch": 0.7535321821036107,
744
+ "grad_norm": 4.40625,
745
+ "learning_rate": 9.745480724951473e-05,
746
+ "loss": 0.6141,
747
+ "step": 960
748
+ },
749
+ {
750
+ "epoch": 0.7613814756671899,
751
+ "grad_norm": 4.96875,
752
+ "learning_rate": 9.73437006651333e-05,
753
+ "loss": 0.6161,
754
+ "step": 970
755
+ },
756
+ {
757
+ "epoch": 0.7692307692307693,
758
+ "grad_norm": 4.25,
759
+ "learning_rate": 9.723028661557345e-05,
760
+ "loss": 0.7166,
761
+ "step": 980
762
+ },
763
+ {
764
+ "epoch": 0.7770800627943485,
765
+ "grad_norm": 5.34375,
766
+ "learning_rate": 9.711457062847595e-05,
767
+ "loss": 0.6028,
768
+ "step": 990
769
+ },
770
+ {
771
+ "epoch": 0.7849293563579278,
772
+ "grad_norm": 4.34375,
773
+ "learning_rate": 9.699655834367478e-05,
774
+ "loss": 0.5262,
775
+ "step": 1000
776
+ },
777
+ {
778
+ "epoch": 0.7849293563579278,
779
+ "eval/nano_beir.arguana": 0.4712918167441636,
780
+ "step": 1000
781
+ },
782
+ {
783
+ "epoch": 0.7849293563579278,
784
+ "eval/nano_beir.climate_fever": 0.23854189010584295,
785
+ "step": 1000
786
+ },
787
+ {
788
+ "epoch": 0.7849293563579278,
789
+ "eval/nano_beir.dbpedia_entity": 0.33317100014903445,
790
+ "step": 1000
791
+ },
792
+ {
793
+ "epoch": 0.7849293563579278,
794
+ "eval/nano_beir.fever": 0.6283057972805869,
795
+ "step": 1000
796
+ },
797
+ {
798
+ "epoch": 0.7849293563579278,
799
+ "eval/nano_beir.fiqa": 0.4448076388918781,
800
+ "step": 1000
801
+ },
802
+ {
803
+ "epoch": 0.7849293563579278,
804
+ "eval/nano_beir.hotpotqa": 0.6717658203813929,
805
+ "step": 1000
806
+ },
807
+ {
808
+ "epoch": 0.7849293563579278,
809
+ "eval/nano_beir.nfcorpus": 0.35468777200298285,
810
+ "step": 1000
811
+ },
812
+ {
813
+ "epoch": 0.7849293563579278,
814
+ "eval/nano_beir.nq": 0.41666515563087825,
815
+ "step": 1000
816
+ },
817
+ {
818
+ "epoch": 0.7849293563579278,
819
+ "eval/nano_beir.quora": 0.8413355846974869,
820
+ "step": 1000
821
+ },
822
+ {
823
+ "epoch": 0.7849293563579278,
824
+ "eval/nano_beir.scidocs": 0.25553825348355963,
825
+ "step": 1000
826
+ },
827
+ {
828
+ "epoch": 0.7849293563579278,
829
+ "eval/nano_beir.scifact": 0.7484102747744907,
830
+ "step": 1000
831
+ },
832
+ {
833
+ "epoch": 0.7849293563579278,
834
+ "eval/nano_beir.webis_touche2020": 0.2895097071551271,
835
+ "step": 1000
836
+ },
837
+ {
838
+ "epoch": 0.7849293563579278,
839
+ "eval/avg": 0.47450255927478535,
840
+ "step": 1000
841
+ },
842
+ {
843
+ "epoch": 0.792778649921507,
844
+ "grad_norm": 4.90625,
845
+ "learning_rate": 9.68762555129222e-05,
846
+ "loss": 0.5077,
847
+ "step": 1010
848
+ },
849
+ {
850
+ "epoch": 0.8006279434850864,
851
+ "grad_norm": 5.0,
852
+ "learning_rate": 9.675366799960841e-05,
853
+ "loss": 0.7276,
854
+ "step": 1020
855
+ },
856
+ {
857
+ "epoch": 0.8084772370486656,
858
+ "grad_norm": 6.59375,
859
+ "learning_rate": 9.662880177847594e-05,
860
+ "loss": 0.5999,
861
+ "step": 1030
862
+ },
863
+ {
864
+ "epoch": 0.8163265306122449,
865
+ "grad_norm": 3.875,
866
+ "learning_rate": 9.65016629353282e-05,
867
+ "loss": 0.6675,
868
+ "step": 1040
869
+ },
870
+ {
871
+ "epoch": 0.8241758241758241,
872
+ "grad_norm": 3.875,
873
+ "learning_rate": 9.637225766673307e-05,
874
+ "loss": 0.6682,
875
+ "step": 1050
876
+ },
877
+ {
878
+ "epoch": 0.8320251177394035,
879
+ "grad_norm": 5.125,
880
+ "learning_rate": 9.624059227972075e-05,
881
+ "loss": 0.6381,
882
+ "step": 1060
883
+ },
884
+ {
885
+ "epoch": 0.8398744113029827,
886
+ "grad_norm": 4.8125,
887
+ "learning_rate": 9.610667319147647e-05,
888
+ "loss": 0.607,
889
+ "step": 1070
890
+ },
891
+ {
892
+ "epoch": 0.847723704866562,
893
+ "grad_norm": 4.75,
894
+ "learning_rate": 9.597050692902765e-05,
895
+ "loss": 0.5786,
896
+ "step": 1080
897
+ },
898
+ {
899
+ "epoch": 0.8555729984301413,
900
+ "grad_norm": 6.125,
901
+ "learning_rate": 9.583210012892582e-05,
902
+ "loss": 0.5928,
903
+ "step": 1090
904
+ },
905
+ {
906
+ "epoch": 0.8634222919937206,
907
+ "grad_norm": 3.734375,
908
+ "learning_rate": 9.569145953692315e-05,
909
+ "loss": 0.595,
910
+ "step": 1100
911
+ },
912
+ {
913
+ "epoch": 0.8712715855572999,
914
+ "grad_norm": 4.8125,
915
+ "learning_rate": 9.55485920076437e-05,
916
+ "loss": 0.604,
917
+ "step": 1110
918
+ },
919
+ {
920
+ "epoch": 0.8791208791208791,
921
+ "grad_norm": 5.0625,
922
+ "learning_rate": 9.540350450424926e-05,
923
+ "loss": 0.6252,
924
+ "step": 1120
925
+ },
926
+ {
927
+ "epoch": 0.8869701726844584,
928
+ "grad_norm": 5.46875,
929
+ "learning_rate": 9.525620409810007e-05,
930
+ "loss": 0.5765,
931
+ "step": 1130
932
+ },
933
+ {
934
+ "epoch": 0.8948194662480377,
935
+ "grad_norm": 4.40625,
936
+ "learning_rate": 9.510669796841014e-05,
937
+ "loss": 0.5604,
938
+ "step": 1140
939
+ },
940
+ {
941
+ "epoch": 0.902668759811617,
942
+ "grad_norm": 6.34375,
943
+ "learning_rate": 9.495499340189728e-05,
944
+ "loss": 0.555,
945
+ "step": 1150
946
+ },
947
+ {
948
+ "epoch": 0.9105180533751962,
949
+ "grad_norm": 4.71875,
950
+ "learning_rate": 9.480109779242805e-05,
951
+ "loss": 0.6496,
952
+ "step": 1160
953
+ },
954
+ {
955
+ "epoch": 0.9183673469387755,
956
+ "grad_norm": 4.71875,
957
+ "learning_rate": 9.464501864065735e-05,
958
+ "loss": 0.584,
959
+ "step": 1170
960
+ },
961
+ {
962
+ "epoch": 0.9262166405023547,
963
+ "grad_norm": 4.1875,
964
+ "learning_rate": 9.448676355366282e-05,
965
+ "loss": 0.6644,
966
+ "step": 1180
967
+ },
968
+ {
969
+ "epoch": 0.9340659340659341,
970
+ "grad_norm": 3.890625,
971
+ "learning_rate": 9.432634024457414e-05,
972
+ "loss": 0.6077,
973
+ "step": 1190
974
+ },
975
+ {
976
+ "epoch": 0.9419152276295133,
977
+ "grad_norm": 6.1875,
978
+ "learning_rate": 9.416375653219709e-05,
979
+ "loss": 0.5531,
980
+ "step": 1200
981
+ },
982
+ {
983
+ "epoch": 0.9497645211930926,
984
+ "grad_norm": 3.921875,
985
+ "learning_rate": 9.399902034063242e-05,
986
+ "loss": 0.6593,
987
+ "step": 1210
988
+ },
989
+ {
990
+ "epoch": 0.957613814756672,
991
+ "grad_norm": 4.15625,
992
+ "learning_rate": 9.383213969888972e-05,
993
+ "loss": 0.5477,
994
+ "step": 1220
995
+ },
996
+ {
997
+ "epoch": 0.9654631083202512,
998
+ "grad_norm": 4.09375,
999
+ "learning_rate": 9.366312274049602e-05,
1000
+ "loss": 0.6071,
1001
+ "step": 1230
1002
+ },
1003
+ {
1004
+ "epoch": 0.9733124018838305,
1005
+ "grad_norm": 4.34375,
1006
+ "learning_rate": 9.34919777030994e-05,
1007
+ "loss": 0.5946,
1008
+ "step": 1240
1009
+ },
1010
+ {
1011
+ "epoch": 0.9811616954474097,
1012
+ "grad_norm": 4.8125,
1013
+ "learning_rate": 9.331871292806759e-05,
1014
+ "loss": 0.5837,
1015
+ "step": 1250
1016
+ },
1017
+ {
1018
+ "epoch": 0.989010989010989,
1019
+ "grad_norm": 5.1875,
1020
+ "learning_rate": 9.314333686008125e-05,
1021
+ "loss": 0.6056,
1022
+ "step": 1260
1023
+ },
1024
+ {
1025
+ "epoch": 0.9968602825745683,
1026
+ "grad_norm": 4.90625,
1027
+ "learning_rate": 9.296585804672252e-05,
1028
+ "loss": 0.5941,
1029
+ "step": 1270
1030
+ },
1031
+ {
1032
+ "epoch": 1.0047095761381475,
1033
+ "grad_norm": 5.03125,
1034
+ "learning_rate": 9.278628513805838e-05,
1035
+ "loss": 0.6723,
1036
+ "step": 1280
1037
+ },
1038
+ {
1039
+ "epoch": 1.012558869701727,
1040
+ "grad_norm": 3.875,
1041
+ "learning_rate": 9.260462688621905e-05,
1042
+ "loss": 0.6221,
1043
+ "step": 1290
1044
+ },
1045
+ {
1046
+ "epoch": 1.0204081632653061,
1047
+ "grad_norm": 4.3125,
1048
+ "learning_rate": 9.242089214497145e-05,
1049
+ "loss": 0.5627,
1050
+ "step": 1300
1051
+ },
1052
+ {
1053
+ "epoch": 1.0282574568288854,
1054
+ "grad_norm": 4.6875,
1055
+ "learning_rate": 9.223508986928766e-05,
1056
+ "loss": 0.6045,
1057
+ "step": 1310
1058
+ },
1059
+ {
1060
+ "epoch": 1.0361067503924646,
1061
+ "grad_norm": 3.515625,
1062
+ "learning_rate": 9.204722911490846e-05,
1063
+ "loss": 0.5723,
1064
+ "step": 1320
1065
+ },
1066
+ {
1067
+ "epoch": 1.043956043956044,
1068
+ "grad_norm": 5.625,
1069
+ "learning_rate": 9.1857319037902e-05,
1070
+ "loss": 0.5616,
1071
+ "step": 1330
1072
+ },
1073
+ {
1074
+ "epoch": 1.0518053375196232,
1075
+ "grad_norm": 5.5625,
1076
+ "learning_rate": 9.16653688942175e-05,
1077
+ "loss": 0.5994,
1078
+ "step": 1340
1079
+ },
1080
+ {
1081
+ "epoch": 1.0596546310832025,
1082
+ "grad_norm": 4.65625,
1083
+ "learning_rate": 9.147138803923416e-05,
1084
+ "loss": 0.6163,
1085
+ "step": 1350
1086
+ },
1087
+ {
1088
+ "epoch": 1.0675039246467817,
1089
+ "grad_norm": 8.0,
1090
+ "learning_rate": 9.127538592730519e-05,
1091
+ "loss": 0.6001,
1092
+ "step": 1360
1093
+ },
1094
+ {
1095
+ "epoch": 1.0753532182103611,
1096
+ "grad_norm": 4.4375,
1097
+ "learning_rate": 9.107737211129701e-05,
1098
+ "loss": 0.55,
1099
+ "step": 1370
1100
+ },
1101
+ {
1102
+ "epoch": 1.0832025117739403,
1103
+ "grad_norm": 5.21875,
1104
+ "learning_rate": 9.087735624212365e-05,
1105
+ "loss": 0.5109,
1106
+ "step": 1380
1107
+ },
1108
+ {
1109
+ "epoch": 1.0910518053375196,
1110
+ "grad_norm": 4.59375,
1111
+ "learning_rate": 9.06753480682764e-05,
1112
+ "loss": 0.6688,
1113
+ "step": 1390
1114
+ },
1115
+ {
1116
+ "epoch": 1.098901098901099,
1117
+ "grad_norm": 5.3125,
1118
+ "learning_rate": 9.047135743534868e-05,
1119
+ "loss": 0.6504,
1120
+ "step": 1400
1121
+ },
1122
+ {
1123
+ "epoch": 1.1067503924646782,
1124
+ "grad_norm": 4.625,
1125
+ "learning_rate": 9.02653942855561e-05,
1126
+ "loss": 0.6315,
1127
+ "step": 1410
1128
+ },
1129
+ {
1130
+ "epoch": 1.1145996860282574,
1131
+ "grad_norm": 6.25,
1132
+ "learning_rate": 9.005746865725207e-05,
1133
+ "loss": 0.6297,
1134
+ "step": 1420
1135
+ },
1136
+ {
1137
+ "epoch": 1.1224489795918366,
1138
+ "grad_norm": 5.5625,
1139
+ "learning_rate": 8.984759068443831e-05,
1140
+ "loss": 0.5962,
1141
+ "step": 1430
1142
+ },
1143
+ {
1144
+ "epoch": 1.130298273155416,
1145
+ "grad_norm": 5.0625,
1146
+ "learning_rate": 8.963577059627118e-05,
1147
+ "loss": 0.5681,
1148
+ "step": 1440
1149
+ },
1150
+ {
1151
+ "epoch": 1.1381475667189953,
1152
+ "grad_norm": 4.78125,
1153
+ "learning_rate": 8.942201871656291e-05,
1154
+ "loss": 0.6486,
1155
+ "step": 1450
1156
+ },
1157
+ {
1158
+ "epoch": 1.1459968602825745,
1159
+ "grad_norm": 5.40625,
1160
+ "learning_rate": 8.920634546327858e-05,
1161
+ "loss": 0.5462,
1162
+ "step": 1460
1163
+ },
1164
+ {
1165
+ "epoch": 1.1538461538461537,
1166
+ "grad_norm": 5.6875,
1167
+ "learning_rate": 8.898876134802826e-05,
1168
+ "loss": 0.5653,
1169
+ "step": 1470
1170
+ },
1171
+ {
1172
+ "epoch": 1.1616954474097332,
1173
+ "grad_norm": 4.5625,
1174
+ "learning_rate": 8.87692769755548e-05,
1175
+ "loss": 0.492,
1176
+ "step": 1480
1177
+ },
1178
+ {
1179
+ "epoch": 1.1695447409733124,
1180
+ "grad_norm": 4.34375,
1181
+ "learning_rate": 8.854790304321681e-05,
1182
+ "loss": 0.5613,
1183
+ "step": 1490
1184
+ },
1185
+ {
1186
+ "epoch": 1.1773940345368916,
1187
+ "grad_norm": 5.15625,
1188
+ "learning_rate": 8.832465034046749e-05,
1189
+ "loss": 0.5728,
1190
+ "step": 1500
1191
+ },
1192
+ {
1193
+ "epoch": 1.185243328100471,
1194
+ "grad_norm": 4.34375,
1195
+ "learning_rate": 8.80995297483286e-05,
1196
+ "loss": 0.5405,
1197
+ "step": 1510
1198
+ },
1199
+ {
1200
+ "epoch": 1.1930926216640503,
1201
+ "grad_norm": 4.53125,
1202
+ "learning_rate": 8.787255223886019e-05,
1203
+ "loss": 0.5618,
1204
+ "step": 1520
1205
+ },
1206
+ {
1207
+ "epoch": 1.2009419152276295,
1208
+ "grad_norm": 4.875,
1209
+ "learning_rate": 8.764372887462586e-05,
1210
+ "loss": 0.5731,
1211
+ "step": 1530
1212
+ },
1213
+ {
1214
+ "epoch": 1.2087912087912087,
1215
+ "grad_norm": 4.71875,
1216
+ "learning_rate": 8.741307080815357e-05,
1217
+ "loss": 0.5745,
1218
+ "step": 1540
1219
+ },
1220
+ {
1221
+ "epoch": 1.2166405023547882,
1222
+ "grad_norm": 5.15625,
1223
+ "learning_rate": 8.718058928139205e-05,
1224
+ "loss": 0.5266,
1225
+ "step": 1550
1226
+ },
1227
+ {
1228
+ "epoch": 1.2244897959183674,
1229
+ "grad_norm": 4.53125,
1230
+ "learning_rate": 8.694629562516294e-05,
1231
+ "loss": 0.6069,
1232
+ "step": 1560
1233
+ },
1234
+ {
1235
+ "epoch": 1.2323390894819466,
1236
+ "grad_norm": 3.5,
1237
+ "learning_rate": 8.671020125860851e-05,
1238
+ "loss": 0.5433,
1239
+ "step": 1570
1240
+ },
1241
+ {
1242
+ "epoch": 1.2401883830455258,
1243
+ "grad_norm": 4.0,
1244
+ "learning_rate": 8.647231768863512e-05,
1245
+ "loss": 0.4842,
1246
+ "step": 1580
1247
+ },
1248
+ {
1249
+ "epoch": 1.2480376766091053,
1250
+ "grad_norm": 4.6875,
1251
+ "learning_rate": 8.623265650935234e-05,
1252
+ "loss": 0.58,
1253
+ "step": 1590
1254
+ },
1255
+ {
1256
+ "epoch": 1.2558869701726845,
1257
+ "grad_norm": 4.125,
1258
+ "learning_rate": 8.599122940150795e-05,
1259
+ "loss": 0.5331,
1260
+ "step": 1600
1261
+ },
1262
+ {
1263
+ "epoch": 1.2637362637362637,
1264
+ "grad_norm": 6.0,
1265
+ "learning_rate": 8.574804813191858e-05,
1266
+ "loss": 0.6408,
1267
+ "step": 1610
1268
+ },
1269
+ {
1270
+ "epoch": 1.2715855572998431,
1271
+ "grad_norm": 4.90625,
1272
+ "learning_rate": 8.550312455289625e-05,
1273
+ "loss": 0.5044,
1274
+ "step": 1620
1275
+ },
1276
+ {
1277
+ "epoch": 1.2794348508634223,
1278
+ "grad_norm": 4.96875,
1279
+ "learning_rate": 8.525647060167063e-05,
1280
+ "loss": 0.6098,
1281
+ "step": 1630
1282
+ },
1283
+ {
1284
+ "epoch": 1.2872841444270016,
1285
+ "grad_norm": 4.1875,
1286
+ "learning_rate": 8.500809829980734e-05,
1287
+ "loss": 0.5911,
1288
+ "step": 1640
1289
+ },
1290
+ {
1291
+ "epoch": 1.2951334379905808,
1292
+ "grad_norm": 5.3125,
1293
+ "learning_rate": 8.4758019752622e-05,
1294
+ "loss": 0.5876,
1295
+ "step": 1650
1296
+ },
1297
+ {
1298
+ "epoch": 1.30298273155416,
1299
+ "grad_norm": 5.09375,
1300
+ "learning_rate": 8.450624714859016e-05,
1301
+ "loss": 0.5618,
1302
+ "step": 1660
1303
+ },
1304
+ {
1305
+ "epoch": 1.3108320251177394,
1306
+ "grad_norm": 3.78125,
1307
+ "learning_rate": 8.425279275875336e-05,
1308
+ "loss": 0.5629,
1309
+ "step": 1670
1310
+ },
1311
+ {
1312
+ "epoch": 1.3186813186813187,
1313
+ "grad_norm": 3.953125,
1314
+ "learning_rate": 8.399766893612096e-05,
1315
+ "loss": 0.5672,
1316
+ "step": 1680
1317
+ },
1318
+ {
1319
+ "epoch": 1.3265306122448979,
1320
+ "grad_norm": 5.5,
1321
+ "learning_rate": 8.374088811506819e-05,
1322
+ "loss": 0.6229,
1323
+ "step": 1690
1324
+ },
1325
+ {
1326
+ "epoch": 1.3343799058084773,
1327
+ "grad_norm": 7.78125,
1328
+ "learning_rate": 8.348246281072998e-05,
1329
+ "loss": 0.5945,
1330
+ "step": 1700
1331
+ },
1332
+ {
1333
+ "epoch": 1.3422291993720565,
1334
+ "grad_norm": 4.125,
1335
+ "learning_rate": 8.322240561839109e-05,
1336
+ "loss": 0.5755,
1337
+ "step": 1710
1338
+ },
1339
+ {
1340
+ "epoch": 1.3500784929356358,
1341
+ "grad_norm": 5.90625,
1342
+ "learning_rate": 8.296072921287217e-05,
1343
+ "loss": 0.5472,
1344
+ "step": 1720
1345
+ },
1346
+ {
1347
+ "epoch": 1.3579277864992152,
1348
+ "grad_norm": 6.34375,
1349
+ "learning_rate": 8.269744634791208e-05,
1350
+ "loss": 0.6287,
1351
+ "step": 1730
1352
+ },
1353
+ {
1354
+ "epoch": 1.3657770800627944,
1355
+ "grad_norm": 4.34375,
1356
+ "learning_rate": 8.243256985554621e-05,
1357
+ "loss": 0.5684,
1358
+ "step": 1740
1359
+ },
1360
+ {
1361
+ "epoch": 1.3736263736263736,
1362
+ "grad_norm": 5.875,
1363
+ "learning_rate": 8.21661126454811e-05,
1364
+ "loss": 0.5455,
1365
+ "step": 1750
1366
+ },
1367
+ {
1368
+ "epoch": 1.3814756671899528,
1369
+ "grad_norm": 4.3125,
1370
+ "learning_rate": 8.189808770446528e-05,
1371
+ "loss": 0.5403,
1372
+ "step": 1760
1373
+ },
1374
+ {
1375
+ "epoch": 1.389324960753532,
1376
+ "grad_norm": 5.03125,
1377
+ "learning_rate": 8.162850809565623e-05,
1378
+ "loss": 0.5343,
1379
+ "step": 1770
1380
+ },
1381
+ {
1382
+ "epoch": 1.3971742543171115,
1383
+ "grad_norm": 4.53125,
1384
+ "learning_rate": 8.135738695798376e-05,
1385
+ "loss": 0.5692,
1386
+ "step": 1780
1387
+ },
1388
+ {
1389
+ "epoch": 1.4050235478806907,
1390
+ "grad_norm": 5.125,
1391
+ "learning_rate": 8.108473750550965e-05,
1392
+ "loss": 0.5382,
1393
+ "step": 1790
1394
+ },
1395
+ {
1396
+ "epoch": 1.41287284144427,
1397
+ "grad_norm": 5.6875,
1398
+ "learning_rate": 8.081057302678352e-05,
1399
+ "loss": 0.5316,
1400
+ "step": 1800
1401
+ },
1402
+ {
1403
+ "epoch": 1.4207221350078494,
1404
+ "grad_norm": 5.15625,
1405
+ "learning_rate": 8.053490688419533e-05,
1406
+ "loss": 0.5221,
1407
+ "step": 1810
1408
+ },
1409
+ {
1410
+ "epoch": 1.4285714285714286,
1411
+ "grad_norm": 6.6875,
1412
+ "learning_rate": 8.02577525133239e-05,
1413
+ "loss": 0.6501,
1414
+ "step": 1820
1415
+ },
1416
+ {
1417
+ "epoch": 1.4364207221350078,
1418
+ "grad_norm": 7.34375,
1419
+ "learning_rate": 7.997912342228232e-05,
1420
+ "loss": 0.5745,
1421
+ "step": 1830
1422
+ },
1423
+ {
1424
+ "epoch": 1.4442700156985873,
1425
+ "grad_norm": 4.15625,
1426
+ "learning_rate": 7.969903319105935e-05,
1427
+ "loss": 0.5797,
1428
+ "step": 1840
1429
+ },
1430
+ {
1431
+ "epoch": 1.4521193092621665,
1432
+ "grad_norm": 6.21875,
1433
+ "learning_rate": 7.941749547085777e-05,
1434
+ "loss": 0.6899,
1435
+ "step": 1850
1436
+ },
1437
+ {
1438
+ "epoch": 1.4599686028257457,
1439
+ "grad_norm": 4.625,
1440
+ "learning_rate": 7.913452398342881e-05,
1441
+ "loss": 0.5891,
1442
+ "step": 1860
1443
+ },
1444
+ {
1445
+ "epoch": 1.467817896389325,
1446
+ "grad_norm": 7.1875,
1447
+ "learning_rate": 7.885013252040359e-05,
1448
+ "loss": 0.6118,
1449
+ "step": 1870
1450
+ },
1451
+ {
1452
+ "epoch": 1.4756671899529041,
1453
+ "grad_norm": 6.8125,
1454
+ "learning_rate": 7.856433494262077e-05,
1455
+ "loss": 0.6102,
1456
+ "step": 1880
1457
+ },
1458
+ {
1459
+ "epoch": 1.4835164835164836,
1460
+ "grad_norm": 5.0,
1461
+ "learning_rate": 7.827714517945115e-05,
1462
+ "loss": 0.5459,
1463
+ "step": 1890
1464
+ },
1465
+ {
1466
+ "epoch": 1.4913657770800628,
1467
+ "grad_norm": 4.40625,
1468
+ "learning_rate": 7.798857722811856e-05,
1469
+ "loss": 0.5717,
1470
+ "step": 1900
1471
+ },
1472
+ {
1473
+ "epoch": 1.499215070643642,
1474
+ "grad_norm": 3.734375,
1475
+ "learning_rate": 7.769864515301787e-05,
1476
+ "loss": 0.5279,
1477
+ "step": 1910
1478
+ },
1479
+ {
1480
+ "epoch": 1.5070643642072215,
1481
+ "grad_norm": 4.6875,
1482
+ "learning_rate": 7.740736308502938e-05,
1483
+ "loss": 0.66,
1484
+ "step": 1920
1485
+ },
1486
+ {
1487
+ "epoch": 1.5149136577708007,
1488
+ "grad_norm": 6.5,
1489
+ "learning_rate": 7.711474522083015e-05,
1490
+ "loss": 0.546,
1491
+ "step": 1930
1492
+ },
1493
+ {
1494
+ "epoch": 1.5227629513343799,
1495
+ "grad_norm": 4.625,
1496
+ "learning_rate": 7.682080582220206e-05,
1497
+ "loss": 0.5207,
1498
+ "step": 1940
1499
+ },
1500
+ {
1501
+ "epoch": 1.5306122448979593,
1502
+ "grad_norm": 6.4375,
1503
+ "learning_rate": 7.65255592153367e-05,
1504
+ "loss": 0.6005,
1505
+ "step": 1950
1506
+ },
1507
+ {
1508
+ "epoch": 1.5384615384615383,
1509
+ "grad_norm": 4.15625,
1510
+ "learning_rate": 7.622901979013717e-05,
1511
+ "loss": 0.5897,
1512
+ "step": 1960
1513
+ },
1514
+ {
1515
+ "epoch": 1.5463108320251178,
1516
+ "grad_norm": 4.40625,
1517
+ "learning_rate": 7.593120199951671e-05,
1518
+ "loss": 0.5639,
1519
+ "step": 1970
1520
+ },
1521
+ {
1522
+ "epoch": 1.554160125588697,
1523
+ "grad_norm": 4.59375,
1524
+ "learning_rate": 7.563212035869425e-05,
1525
+ "loss": 0.5479,
1526
+ "step": 1980
1527
+ },
1528
+ {
1529
+ "epoch": 1.5620094191522762,
1530
+ "grad_norm": 4.8125,
1531
+ "learning_rate": 7.533178944448705e-05,
1532
+ "loss": 0.5539,
1533
+ "step": 1990
1534
+ },
1535
+ {
1536
+ "epoch": 1.5698587127158556,
1537
+ "grad_norm": 6.09375,
1538
+ "learning_rate": 7.503022389460014e-05,
1539
+ "loss": 0.625,
1540
+ "step": 2000
1541
+ }
1542
+ ],
1543
+ "logging_steps": 10,
1544
+ "max_steps": 5000,
1545
+ "num_input_tokens_seen": 0,
1546
+ "num_train_epochs": 4,
1547
+ "save_steps": 1000,
1548
+ "stateful_callbacks": {
1549
+ "TrainerControl": {
1550
+ "args": {
1551
+ "should_epoch_stop": false,
1552
+ "should_evaluate": false,
1553
+ "should_log": false,
1554
+ "should_save": true,
1555
+ "should_training_stop": false
1556
+ },
1557
+ "attributes": {}
1558
+ }
1559
+ },
1560
+ "total_flos": 0.0,
1561
+ "train_batch_size": 16,
1562
+ "trial_name": null,
1563
+ "trial_params": null
1564
+ }
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-2000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9365177436ad70bb9d6000d96d3ccdc5c5ca38536defdcfec6b548401edbc07f
3
+ size 6481
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "gradient_checkpointing": false,
20
+ "hidden_activation": "gelu",
21
+ "hidden_size": 768,
22
+ "initializer_cutoff_factor": 2.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 1152,
25
+ "layer_norm_eps": 1e-05,
26
+ "local_attention": 128,
27
+ "local_rope_theta": 10000.0,
28
+ "max_position_embeddings": 8192,
29
+ "mlp_bias": false,
30
+ "mlp_dropout": 0.0,
31
+ "model_type": "modernbert",
32
+ "norm_bias": false,
33
+ "norm_eps": 1e-05,
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 22,
36
+ "pad_token_id": 50283,
37
+ "position_embedding_type": "absolute",
38
+ "repad_logits_with_grad": false,
39
+ "sep_token_id": 50282,
40
+ "sparse_pred_ignore_index": -100,
41
+ "sparse_prediction": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.53.0",
44
+ "vocab_size": 50368
45
+ }
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bb963d2d2da684d38c649275acf790b53b2b9efeb1a54ccb045770f523ccc1d
3
+ size 298041696
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:707d877637528a6ecd5001a42e1fe755b11b439f37b6d35579a2d7247f6dcfc2
3
+ size 596170443
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac0e770c61e3acef7bd144efee65a450ed487b244ed293a8dd801394ffcae775
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8fb125336725f7741cb4daa1e3d06e225bbacfde8d41c4dcabb6762c222e62c6
3
+ size 15429
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:153c865f77c7129ba565bded50f334683d51c80f20e3cfec39e62f8737b86f0d
3
+ size 15429
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d017ce00fcebac7edc058ddd138f194eb0340f2d8ad0879bdab08f922ed0846e
3
+ size 15429
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01a8f419d963b829163d14296852a07ad341e5e63e49d44de7c0a10c9369d217
3
+ size 1465
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/trainer_state.json ADDED
@@ -0,0 +1,2329 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 2.3547880690737832,
6
+ "eval_steps": 1000.0,
7
+ "global_step": 3000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0,
14
+ "eval/nano_beir.arguana": 0.4400549254974839,
15
+ "step": 0
16
+ },
17
+ {
18
+ "epoch": 0,
19
+ "eval/nano_beir.climate_fever": 0.22880053057901228,
20
+ "step": 0
21
+ },
22
+ {
23
+ "epoch": 0,
24
+ "eval/nano_beir.dbpedia_entity": 0.30026716538531234,
25
+ "step": 0
26
+ },
27
+ {
28
+ "epoch": 0,
29
+ "eval/nano_beir.fever": 0.6419178788751609,
30
+ "step": 0
31
+ },
32
+ {
33
+ "epoch": 0,
34
+ "eval/nano_beir.fiqa": 0.4370593595651809,
35
+ "step": 0
36
+ },
37
+ {
38
+ "epoch": 0,
39
+ "eval/nano_beir.hotpotqa": 0.6723047477224562,
40
+ "step": 0
41
+ },
42
+ {
43
+ "epoch": 0,
44
+ "eval/nano_beir.nfcorpus": 0.3437446614516137,
45
+ "step": 0
46
+ },
47
+ {
48
+ "epoch": 0,
49
+ "eval/nano_beir.nq": 0.436711543163819,
50
+ "step": 0
51
+ },
52
+ {
53
+ "epoch": 0,
54
+ "eval/nano_beir.quora": 0.8439997970269911,
55
+ "step": 0
56
+ },
57
+ {
58
+ "epoch": 0,
59
+ "eval/nano_beir.scidocs": 0.26304331393922176,
60
+ "step": 0
61
+ },
62
+ {
63
+ "epoch": 0,
64
+ "eval/nano_beir.scifact": 0.7599341408912704,
65
+ "step": 0
66
+ },
67
+ {
68
+ "epoch": 0,
69
+ "eval/nano_beir.webis_touche2020": 0.29696490432942557,
70
+ "step": 0
71
+ },
72
+ {
73
+ "epoch": 0,
74
+ "eval/avg": 0.47206691403557904,
75
+ "step": 0
76
+ },
77
+ {
78
+ "epoch": 0.007849293563579277,
79
+ "grad_norm": 8.25,
80
+ "learning_rate": 1.8e-06,
81
+ "loss": 0.7314,
82
+ "step": 10
83
+ },
84
+ {
85
+ "epoch": 0.015698587127158554,
86
+ "grad_norm": 7.03125,
87
+ "learning_rate": 3.8e-06,
88
+ "loss": 0.8796,
89
+ "step": 20
90
+ },
91
+ {
92
+ "epoch": 0.023547880690737835,
93
+ "grad_norm": 7.3125,
94
+ "learning_rate": 5.8e-06,
95
+ "loss": 0.7311,
96
+ "step": 30
97
+ },
98
+ {
99
+ "epoch": 0.03139717425431711,
100
+ "grad_norm": 7.15625,
101
+ "learning_rate": 7.8e-06,
102
+ "loss": 0.8214,
103
+ "step": 40
104
+ },
105
+ {
106
+ "epoch": 0.03924646781789639,
107
+ "grad_norm": 9.1875,
108
+ "learning_rate": 9.800000000000001e-06,
109
+ "loss": 0.8489,
110
+ "step": 50
111
+ },
112
+ {
113
+ "epoch": 0.04709576138147567,
114
+ "grad_norm": 6.84375,
115
+ "learning_rate": 1.18e-05,
116
+ "loss": 0.7661,
117
+ "step": 60
118
+ },
119
+ {
120
+ "epoch": 0.054945054945054944,
121
+ "grad_norm": 6.6875,
122
+ "learning_rate": 1.3800000000000002e-05,
123
+ "loss": 0.7525,
124
+ "step": 70
125
+ },
126
+ {
127
+ "epoch": 0.06279434850863422,
128
+ "grad_norm": 7.5625,
129
+ "learning_rate": 1.58e-05,
130
+ "loss": 0.8653,
131
+ "step": 80
132
+ },
133
+ {
134
+ "epoch": 0.0706436420722135,
135
+ "grad_norm": 6.25,
136
+ "learning_rate": 1.78e-05,
137
+ "loss": 0.7607,
138
+ "step": 90
139
+ },
140
+ {
141
+ "epoch": 0.07849293563579278,
142
+ "grad_norm": 5.28125,
143
+ "learning_rate": 1.9800000000000004e-05,
144
+ "loss": 0.7259,
145
+ "step": 100
146
+ },
147
+ {
148
+ "epoch": 0.08634222919937205,
149
+ "grad_norm": 5.59375,
150
+ "learning_rate": 2.18e-05,
151
+ "loss": 0.7862,
152
+ "step": 110
153
+ },
154
+ {
155
+ "epoch": 0.09419152276295134,
156
+ "grad_norm": 5.8125,
157
+ "learning_rate": 2.38e-05,
158
+ "loss": 0.8212,
159
+ "step": 120
160
+ },
161
+ {
162
+ "epoch": 0.10204081632653061,
163
+ "grad_norm": 6.0,
164
+ "learning_rate": 2.58e-05,
165
+ "loss": 0.6884,
166
+ "step": 130
167
+ },
168
+ {
169
+ "epoch": 0.10989010989010989,
170
+ "grad_norm": 5.9375,
171
+ "learning_rate": 2.7800000000000005e-05,
172
+ "loss": 0.712,
173
+ "step": 140
174
+ },
175
+ {
176
+ "epoch": 0.11773940345368916,
177
+ "grad_norm": 5.53125,
178
+ "learning_rate": 2.98e-05,
179
+ "loss": 0.6231,
180
+ "step": 150
181
+ },
182
+ {
183
+ "epoch": 0.12558869701726844,
184
+ "grad_norm": 5.59375,
185
+ "learning_rate": 3.18e-05,
186
+ "loss": 0.6733,
187
+ "step": 160
188
+ },
189
+ {
190
+ "epoch": 0.13343799058084774,
191
+ "grad_norm": 6.84375,
192
+ "learning_rate": 3.38e-05,
193
+ "loss": 0.6973,
194
+ "step": 170
195
+ },
196
+ {
197
+ "epoch": 0.141287284144427,
198
+ "grad_norm": 5.4375,
199
+ "learning_rate": 3.58e-05,
200
+ "loss": 0.5928,
201
+ "step": 180
202
+ },
203
+ {
204
+ "epoch": 0.14913657770800628,
205
+ "grad_norm": 6.375,
206
+ "learning_rate": 3.7800000000000004e-05,
207
+ "loss": 0.6189,
208
+ "step": 190
209
+ },
210
+ {
211
+ "epoch": 0.15698587127158556,
212
+ "grad_norm": 5.28125,
213
+ "learning_rate": 3.9800000000000005e-05,
214
+ "loss": 0.6661,
215
+ "step": 200
216
+ },
217
+ {
218
+ "epoch": 0.16483516483516483,
219
+ "grad_norm": 4.875,
220
+ "learning_rate": 4.18e-05,
221
+ "loss": 0.634,
222
+ "step": 210
223
+ },
224
+ {
225
+ "epoch": 0.1726844583987441,
226
+ "grad_norm": 6.78125,
227
+ "learning_rate": 4.38e-05,
228
+ "loss": 0.6963,
229
+ "step": 220
230
+ },
231
+ {
232
+ "epoch": 0.18053375196232338,
233
+ "grad_norm": 5.875,
234
+ "learning_rate": 4.58e-05,
235
+ "loss": 0.6744,
236
+ "step": 230
237
+ },
238
+ {
239
+ "epoch": 0.18838304552590268,
240
+ "grad_norm": 5.90625,
241
+ "learning_rate": 4.78e-05,
242
+ "loss": 0.6635,
243
+ "step": 240
244
+ },
245
+ {
246
+ "epoch": 0.19623233908948196,
247
+ "grad_norm": 4.9375,
248
+ "learning_rate": 4.9800000000000004e-05,
249
+ "loss": 0.6874,
250
+ "step": 250
251
+ },
252
+ {
253
+ "epoch": 0.20408163265306123,
254
+ "grad_norm": 5.28125,
255
+ "learning_rate": 5.1800000000000005e-05,
256
+ "loss": 0.6516,
257
+ "step": 260
258
+ },
259
+ {
260
+ "epoch": 0.2119309262166405,
261
+ "grad_norm": 5.40625,
262
+ "learning_rate": 5.380000000000001e-05,
263
+ "loss": 0.7579,
264
+ "step": 270
265
+ },
266
+ {
267
+ "epoch": 0.21978021978021978,
268
+ "grad_norm": 4.15625,
269
+ "learning_rate": 5.580000000000001e-05,
270
+ "loss": 0.6867,
271
+ "step": 280
272
+ },
273
+ {
274
+ "epoch": 0.22762951334379905,
275
+ "grad_norm": 6.59375,
276
+ "learning_rate": 5.7799999999999995e-05,
277
+ "loss": 0.7038,
278
+ "step": 290
279
+ },
280
+ {
281
+ "epoch": 0.23547880690737832,
282
+ "grad_norm": 5.9375,
283
+ "learning_rate": 5.9800000000000003e-05,
284
+ "loss": 0.6681,
285
+ "step": 300
286
+ },
287
+ {
288
+ "epoch": 0.24332810047095763,
289
+ "grad_norm": 4.4375,
290
+ "learning_rate": 6.18e-05,
291
+ "loss": 0.6133,
292
+ "step": 310
293
+ },
294
+ {
295
+ "epoch": 0.25117739403453687,
296
+ "grad_norm": 4.4375,
297
+ "learning_rate": 6.38e-05,
298
+ "loss": 0.6203,
299
+ "step": 320
300
+ },
301
+ {
302
+ "epoch": 0.25902668759811615,
303
+ "grad_norm": 6.3125,
304
+ "learning_rate": 6.58e-05,
305
+ "loss": 0.7164,
306
+ "step": 330
307
+ },
308
+ {
309
+ "epoch": 0.2668759811616955,
310
+ "grad_norm": 4.84375,
311
+ "learning_rate": 6.780000000000001e-05,
312
+ "loss": 0.7157,
313
+ "step": 340
314
+ },
315
+ {
316
+ "epoch": 0.27472527472527475,
317
+ "grad_norm": 5.09375,
318
+ "learning_rate": 6.98e-05,
319
+ "loss": 0.7176,
320
+ "step": 350
321
+ },
322
+ {
323
+ "epoch": 0.282574568288854,
324
+ "grad_norm": 5.1875,
325
+ "learning_rate": 7.18e-05,
326
+ "loss": 0.702,
327
+ "step": 360
328
+ },
329
+ {
330
+ "epoch": 0.2904238618524333,
331
+ "grad_norm": 5.25,
332
+ "learning_rate": 7.38e-05,
333
+ "loss": 0.5834,
334
+ "step": 370
335
+ },
336
+ {
337
+ "epoch": 0.29827315541601257,
338
+ "grad_norm": 5.46875,
339
+ "learning_rate": 7.58e-05,
340
+ "loss": 0.594,
341
+ "step": 380
342
+ },
343
+ {
344
+ "epoch": 0.30612244897959184,
345
+ "grad_norm": 6.625,
346
+ "learning_rate": 7.780000000000001e-05,
347
+ "loss": 0.7667,
348
+ "step": 390
349
+ },
350
+ {
351
+ "epoch": 0.3139717425431711,
352
+ "grad_norm": 6.8125,
353
+ "learning_rate": 7.98e-05,
354
+ "loss": 0.6477,
355
+ "step": 400
356
+ },
357
+ {
358
+ "epoch": 0.3218210361067504,
359
+ "grad_norm": 5.21875,
360
+ "learning_rate": 8.18e-05,
361
+ "loss": 0.6104,
362
+ "step": 410
363
+ },
364
+ {
365
+ "epoch": 0.32967032967032966,
366
+ "grad_norm": 6.03125,
367
+ "learning_rate": 8.38e-05,
368
+ "loss": 0.6879,
369
+ "step": 420
370
+ },
371
+ {
372
+ "epoch": 0.33751962323390894,
373
+ "grad_norm": 5.90625,
374
+ "learning_rate": 8.58e-05,
375
+ "loss": 0.6514,
376
+ "step": 430
377
+ },
378
+ {
379
+ "epoch": 0.3453689167974882,
380
+ "grad_norm": 6.03125,
381
+ "learning_rate": 8.78e-05,
382
+ "loss": 0.649,
383
+ "step": 440
384
+ },
385
+ {
386
+ "epoch": 0.3532182103610675,
387
+ "grad_norm": 4.75,
388
+ "learning_rate": 8.98e-05,
389
+ "loss": 0.6204,
390
+ "step": 450
391
+ },
392
+ {
393
+ "epoch": 0.36106750392464676,
394
+ "grad_norm": 5.125,
395
+ "learning_rate": 9.180000000000001e-05,
396
+ "loss": 0.6633,
397
+ "step": 460
398
+ },
399
+ {
400
+ "epoch": 0.36891679748822603,
401
+ "grad_norm": 4.59375,
402
+ "learning_rate": 9.38e-05,
403
+ "loss": 0.7058,
404
+ "step": 470
405
+ },
406
+ {
407
+ "epoch": 0.37676609105180536,
408
+ "grad_norm": 5.9375,
409
+ "learning_rate": 9.58e-05,
410
+ "loss": 0.5653,
411
+ "step": 480
412
+ },
413
+ {
414
+ "epoch": 0.38461538461538464,
415
+ "grad_norm": 5.6875,
416
+ "learning_rate": 9.78e-05,
417
+ "loss": 0.6143,
418
+ "step": 490
419
+ },
420
+ {
421
+ "epoch": 0.3924646781789639,
422
+ "grad_norm": 5.28125,
423
+ "learning_rate": 9.98e-05,
424
+ "loss": 0.6218,
425
+ "step": 500
426
+ },
427
+ {
428
+ "epoch": 0.4003139717425432,
429
+ "grad_norm": 6.03125,
430
+ "learning_rate": 9.999901304280685e-05,
431
+ "loss": 0.6115,
432
+ "step": 510
433
+ },
434
+ {
435
+ "epoch": 0.40816326530612246,
436
+ "grad_norm": 4.53125,
437
+ "learning_rate": 9.999560138895238e-05,
438
+ "loss": 0.6206,
439
+ "step": 520
440
+ },
441
+ {
442
+ "epoch": 0.41601255886970173,
443
+ "grad_norm": 5.15625,
444
+ "learning_rate": 9.99897530200195e-05,
445
+ "loss": 0.629,
446
+ "step": 530
447
+ },
448
+ {
449
+ "epoch": 0.423861852433281,
450
+ "grad_norm": 4.6875,
451
+ "learning_rate": 9.998146822104943e-05,
452
+ "loss": 0.6492,
453
+ "step": 540
454
+ },
455
+ {
456
+ "epoch": 0.4317111459968603,
457
+ "grad_norm": 5.1875,
458
+ "learning_rate": 9.997074739583162e-05,
459
+ "loss": 0.585,
460
+ "step": 550
461
+ },
462
+ {
463
+ "epoch": 0.43956043956043955,
464
+ "grad_norm": 7.1875,
465
+ "learning_rate": 9.995759106688393e-05,
466
+ "loss": 0.558,
467
+ "step": 560
468
+ },
469
+ {
470
+ "epoch": 0.4474097331240188,
471
+ "grad_norm": 5.28125,
472
+ "learning_rate": 9.99419998754273e-05,
473
+ "loss": 0.6242,
474
+ "step": 570
475
+ },
476
+ {
477
+ "epoch": 0.4552590266875981,
478
+ "grad_norm": 4.53125,
479
+ "learning_rate": 9.992397458135438e-05,
480
+ "loss": 0.682,
481
+ "step": 580
482
+ },
483
+ {
484
+ "epoch": 0.4631083202511774,
485
+ "grad_norm": 4.3125,
486
+ "learning_rate": 9.990351606319261e-05,
487
+ "loss": 0.6194,
488
+ "step": 590
489
+ },
490
+ {
491
+ "epoch": 0.47095761381475665,
492
+ "grad_norm": 5.90625,
493
+ "learning_rate": 9.988062531806126e-05,
494
+ "loss": 0.6211,
495
+ "step": 600
496
+ },
497
+ {
498
+ "epoch": 0.478806907378336,
499
+ "grad_norm": 4.5625,
500
+ "learning_rate": 9.985530346162299e-05,
501
+ "loss": 0.617,
502
+ "step": 610
503
+ },
504
+ {
505
+ "epoch": 0.48665620094191525,
506
+ "grad_norm": 5.125,
507
+ "learning_rate": 9.982755172802933e-05,
508
+ "loss": 0.5901,
509
+ "step": 620
510
+ },
511
+ {
512
+ "epoch": 0.4945054945054945,
513
+ "grad_norm": 4.15625,
514
+ "learning_rate": 9.979737146986064e-05,
515
+ "loss": 0.6067,
516
+ "step": 630
517
+ },
518
+ {
519
+ "epoch": 0.5023547880690737,
520
+ "grad_norm": 5.4375,
521
+ "learning_rate": 9.976476415806013e-05,
522
+ "loss": 0.5976,
523
+ "step": 640
524
+ },
525
+ {
526
+ "epoch": 0.5102040816326531,
527
+ "grad_norm": 5.75,
528
+ "learning_rate": 9.972973138186217e-05,
529
+ "loss": 0.6467,
530
+ "step": 650
531
+ },
532
+ {
533
+ "epoch": 0.5180533751962323,
534
+ "grad_norm": 7.25,
535
+ "learning_rate": 9.969227484871484e-05,
536
+ "loss": 0.5904,
537
+ "step": 660
538
+ },
539
+ {
540
+ "epoch": 0.5259026687598116,
541
+ "grad_norm": 4.65625,
542
+ "learning_rate": 9.965239638419672e-05,
543
+ "loss": 0.6814,
544
+ "step": 670
545
+ },
546
+ {
547
+ "epoch": 0.533751962323391,
548
+ "grad_norm": 6.125,
549
+ "learning_rate": 9.961009793192792e-05,
550
+ "loss": 0.6684,
551
+ "step": 680
552
+ },
553
+ {
554
+ "epoch": 0.5416012558869702,
555
+ "grad_norm": 4.40625,
556
+ "learning_rate": 9.956538155347534e-05,
557
+ "loss": 0.6206,
558
+ "step": 690
559
+ },
560
+ {
561
+ "epoch": 0.5494505494505495,
562
+ "grad_norm": 5.875,
563
+ "learning_rate": 9.951824942825215e-05,
564
+ "loss": 0.6096,
565
+ "step": 700
566
+ },
567
+ {
568
+ "epoch": 0.5572998430141287,
569
+ "grad_norm": 5.71875,
570
+ "learning_rate": 9.946870385341167e-05,
571
+ "loss": 0.5468,
572
+ "step": 710
573
+ },
574
+ {
575
+ "epoch": 0.565149136577708,
576
+ "grad_norm": 4.84375,
577
+ "learning_rate": 9.941674724373531e-05,
578
+ "loss": 0.5883,
579
+ "step": 720
580
+ },
581
+ {
582
+ "epoch": 0.5729984301412873,
583
+ "grad_norm": 4.75,
584
+ "learning_rate": 9.936238213151491e-05,
585
+ "loss": 0.596,
586
+ "step": 730
587
+ },
588
+ {
589
+ "epoch": 0.5808477237048666,
590
+ "grad_norm": 4.9375,
591
+ "learning_rate": 9.930561116642935e-05,
592
+ "loss": 0.5178,
593
+ "step": 740
594
+ },
595
+ {
596
+ "epoch": 0.5886970172684458,
597
+ "grad_norm": 5.59375,
598
+ "learning_rate": 9.924643711541539e-05,
599
+ "loss": 0.6438,
600
+ "step": 750
601
+ },
602
+ {
603
+ "epoch": 0.5965463108320251,
604
+ "grad_norm": 4.96875,
605
+ "learning_rate": 9.918486286253279e-05,
606
+ "loss": 0.6975,
607
+ "step": 760
608
+ },
609
+ {
610
+ "epoch": 0.6043956043956044,
611
+ "grad_norm": 4.375,
612
+ "learning_rate": 9.912089140882376e-05,
613
+ "loss": 0.5828,
614
+ "step": 770
615
+ },
616
+ {
617
+ "epoch": 0.6122448979591837,
618
+ "grad_norm": 7.0625,
619
+ "learning_rate": 9.90545258721667e-05,
620
+ "loss": 0.5019,
621
+ "step": 780
622
+ },
623
+ {
624
+ "epoch": 0.6200941915227629,
625
+ "grad_norm": 6.75,
626
+ "learning_rate": 9.898576948712426e-05,
627
+ "loss": 0.6489,
628
+ "step": 790
629
+ },
630
+ {
631
+ "epoch": 0.6279434850863422,
632
+ "grad_norm": 5.28125,
633
+ "learning_rate": 9.891462560478562e-05,
634
+ "loss": 0.6364,
635
+ "step": 800
636
+ },
637
+ {
638
+ "epoch": 0.6357927786499215,
639
+ "grad_norm": 4.6875,
640
+ "learning_rate": 9.884109769260325e-05,
641
+ "loss": 0.5458,
642
+ "step": 810
643
+ },
644
+ {
645
+ "epoch": 0.6436420722135008,
646
+ "grad_norm": 3.703125,
647
+ "learning_rate": 9.876518933422386e-05,
648
+ "loss": 0.5859,
649
+ "step": 820
650
+ },
651
+ {
652
+ "epoch": 0.6514913657770801,
653
+ "grad_norm": 4.375,
654
+ "learning_rate": 9.868690422931372e-05,
655
+ "loss": 0.6369,
656
+ "step": 830
657
+ },
658
+ {
659
+ "epoch": 0.6593406593406593,
660
+ "grad_norm": 5.40625,
661
+ "learning_rate": 9.860624619337844e-05,
662
+ "loss": 0.7126,
663
+ "step": 840
664
+ },
665
+ {
666
+ "epoch": 0.6671899529042387,
667
+ "grad_norm": 4.9375,
668
+ "learning_rate": 9.852321915757687e-05,
669
+ "loss": 0.5857,
670
+ "step": 850
671
+ },
672
+ {
673
+ "epoch": 0.6750392464678179,
674
+ "grad_norm": 5.0,
675
+ "learning_rate": 9.843782716852963e-05,
676
+ "loss": 0.645,
677
+ "step": 860
678
+ },
679
+ {
680
+ "epoch": 0.6828885400313972,
681
+ "grad_norm": 3.78125,
682
+ "learning_rate": 9.835007438812177e-05,
683
+ "loss": 0.6538,
684
+ "step": 870
685
+ },
686
+ {
687
+ "epoch": 0.6907378335949764,
688
+ "grad_norm": 5.15625,
689
+ "learning_rate": 9.82599650933e-05,
690
+ "loss": 0.6045,
691
+ "step": 880
692
+ },
693
+ {
694
+ "epoch": 0.6985871271585558,
695
+ "grad_norm": 3.59375,
696
+ "learning_rate": 9.816750367586425e-05,
697
+ "loss": 0.6282,
698
+ "step": 890
699
+ },
700
+ {
701
+ "epoch": 0.706436420722135,
702
+ "grad_norm": 5.1875,
703
+ "learning_rate": 9.807269464225355e-05,
704
+ "loss": 0.5509,
705
+ "step": 900
706
+ },
707
+ {
708
+ "epoch": 0.7142857142857143,
709
+ "grad_norm": 7.125,
710
+ "learning_rate": 9.797554261332645e-05,
711
+ "loss": 0.6392,
712
+ "step": 910
713
+ },
714
+ {
715
+ "epoch": 0.7221350078492935,
716
+ "grad_norm": 5.34375,
717
+ "learning_rate": 9.787605232413574e-05,
718
+ "loss": 0.6152,
719
+ "step": 920
720
+ },
721
+ {
722
+ "epoch": 0.7299843014128728,
723
+ "grad_norm": 4.28125,
724
+ "learning_rate": 9.777422862369783e-05,
725
+ "loss": 0.7706,
726
+ "step": 930
727
+ },
728
+ {
729
+ "epoch": 0.7378335949764521,
730
+ "grad_norm": 4.59375,
731
+ "learning_rate": 9.767007647475617e-05,
732
+ "loss": 0.6656,
733
+ "step": 940
734
+ },
735
+ {
736
+ "epoch": 0.7456828885400314,
737
+ "grad_norm": 4.53125,
738
+ "learning_rate": 9.756360095353956e-05,
739
+ "loss": 0.6592,
740
+ "step": 950
741
+ },
742
+ {
743
+ "epoch": 0.7535321821036107,
744
+ "grad_norm": 4.40625,
745
+ "learning_rate": 9.745480724951473e-05,
746
+ "loss": 0.6141,
747
+ "step": 960
748
+ },
749
+ {
750
+ "epoch": 0.7613814756671899,
751
+ "grad_norm": 4.96875,
752
+ "learning_rate": 9.73437006651333e-05,
753
+ "loss": 0.6161,
754
+ "step": 970
755
+ },
756
+ {
757
+ "epoch": 0.7692307692307693,
758
+ "grad_norm": 4.25,
759
+ "learning_rate": 9.723028661557345e-05,
760
+ "loss": 0.7166,
761
+ "step": 980
762
+ },
763
+ {
764
+ "epoch": 0.7770800627943485,
765
+ "grad_norm": 5.34375,
766
+ "learning_rate": 9.711457062847595e-05,
767
+ "loss": 0.6028,
768
+ "step": 990
769
+ },
770
+ {
771
+ "epoch": 0.7849293563579278,
772
+ "grad_norm": 4.34375,
773
+ "learning_rate": 9.699655834367478e-05,
774
+ "loss": 0.5262,
775
+ "step": 1000
776
+ },
777
+ {
778
+ "epoch": 0.7849293563579278,
779
+ "eval/nano_beir.arguana": 0.4712918167441636,
780
+ "step": 1000
781
+ },
782
+ {
783
+ "epoch": 0.7849293563579278,
784
+ "eval/nano_beir.climate_fever": 0.23854189010584295,
785
+ "step": 1000
786
+ },
787
+ {
788
+ "epoch": 0.7849293563579278,
789
+ "eval/nano_beir.dbpedia_entity": 0.33317100014903445,
790
+ "step": 1000
791
+ },
792
+ {
793
+ "epoch": 0.7849293563579278,
794
+ "eval/nano_beir.fever": 0.6283057972805869,
795
+ "step": 1000
796
+ },
797
+ {
798
+ "epoch": 0.7849293563579278,
799
+ "eval/nano_beir.fiqa": 0.4448076388918781,
800
+ "step": 1000
801
+ },
802
+ {
803
+ "epoch": 0.7849293563579278,
804
+ "eval/nano_beir.hotpotqa": 0.6717658203813929,
805
+ "step": 1000
806
+ },
807
+ {
808
+ "epoch": 0.7849293563579278,
809
+ "eval/nano_beir.nfcorpus": 0.35468777200298285,
810
+ "step": 1000
811
+ },
812
+ {
813
+ "epoch": 0.7849293563579278,
814
+ "eval/nano_beir.nq": 0.41666515563087825,
815
+ "step": 1000
816
+ },
817
+ {
818
+ "epoch": 0.7849293563579278,
819
+ "eval/nano_beir.quora": 0.8413355846974869,
820
+ "step": 1000
821
+ },
822
+ {
823
+ "epoch": 0.7849293563579278,
824
+ "eval/nano_beir.scidocs": 0.25553825348355963,
825
+ "step": 1000
826
+ },
827
+ {
828
+ "epoch": 0.7849293563579278,
829
+ "eval/nano_beir.scifact": 0.7484102747744907,
830
+ "step": 1000
831
+ },
832
+ {
833
+ "epoch": 0.7849293563579278,
834
+ "eval/nano_beir.webis_touche2020": 0.2895097071551271,
835
+ "step": 1000
836
+ },
837
+ {
838
+ "epoch": 0.7849293563579278,
839
+ "eval/avg": 0.47450255927478535,
840
+ "step": 1000
841
+ },
842
+ {
843
+ "epoch": 0.792778649921507,
844
+ "grad_norm": 4.90625,
845
+ "learning_rate": 9.68762555129222e-05,
846
+ "loss": 0.5077,
847
+ "step": 1010
848
+ },
849
+ {
850
+ "epoch": 0.8006279434850864,
851
+ "grad_norm": 5.0,
852
+ "learning_rate": 9.675366799960841e-05,
853
+ "loss": 0.7276,
854
+ "step": 1020
855
+ },
856
+ {
857
+ "epoch": 0.8084772370486656,
858
+ "grad_norm": 6.59375,
859
+ "learning_rate": 9.662880177847594e-05,
860
+ "loss": 0.5999,
861
+ "step": 1030
862
+ },
863
+ {
864
+ "epoch": 0.8163265306122449,
865
+ "grad_norm": 3.875,
866
+ "learning_rate": 9.65016629353282e-05,
867
+ "loss": 0.6675,
868
+ "step": 1040
869
+ },
870
+ {
871
+ "epoch": 0.8241758241758241,
872
+ "grad_norm": 3.875,
873
+ "learning_rate": 9.637225766673307e-05,
874
+ "loss": 0.6682,
875
+ "step": 1050
876
+ },
877
+ {
878
+ "epoch": 0.8320251177394035,
879
+ "grad_norm": 5.125,
880
+ "learning_rate": 9.624059227972075e-05,
881
+ "loss": 0.6381,
882
+ "step": 1060
883
+ },
884
+ {
885
+ "epoch": 0.8398744113029827,
886
+ "grad_norm": 4.8125,
887
+ "learning_rate": 9.610667319147647e-05,
888
+ "loss": 0.607,
889
+ "step": 1070
890
+ },
891
+ {
892
+ "epoch": 0.847723704866562,
893
+ "grad_norm": 4.75,
894
+ "learning_rate": 9.597050692902765e-05,
895
+ "loss": 0.5786,
896
+ "step": 1080
897
+ },
898
+ {
899
+ "epoch": 0.8555729984301413,
900
+ "grad_norm": 6.125,
901
+ "learning_rate": 9.583210012892582e-05,
902
+ "loss": 0.5928,
903
+ "step": 1090
904
+ },
905
+ {
906
+ "epoch": 0.8634222919937206,
907
+ "grad_norm": 3.734375,
908
+ "learning_rate": 9.569145953692315e-05,
909
+ "loss": 0.595,
910
+ "step": 1100
911
+ },
912
+ {
913
+ "epoch": 0.8712715855572999,
914
+ "grad_norm": 4.8125,
915
+ "learning_rate": 9.55485920076437e-05,
916
+ "loss": 0.604,
917
+ "step": 1110
918
+ },
919
+ {
920
+ "epoch": 0.8791208791208791,
921
+ "grad_norm": 5.0625,
922
+ "learning_rate": 9.540350450424926e-05,
923
+ "loss": 0.6252,
924
+ "step": 1120
925
+ },
926
+ {
927
+ "epoch": 0.8869701726844584,
928
+ "grad_norm": 5.46875,
929
+ "learning_rate": 9.525620409810007e-05,
930
+ "loss": 0.5765,
931
+ "step": 1130
932
+ },
933
+ {
934
+ "epoch": 0.8948194662480377,
935
+ "grad_norm": 4.40625,
936
+ "learning_rate": 9.510669796841014e-05,
937
+ "loss": 0.5604,
938
+ "step": 1140
939
+ },
940
+ {
941
+ "epoch": 0.902668759811617,
942
+ "grad_norm": 6.34375,
943
+ "learning_rate": 9.495499340189728e-05,
944
+ "loss": 0.555,
945
+ "step": 1150
946
+ },
947
+ {
948
+ "epoch": 0.9105180533751962,
949
+ "grad_norm": 4.71875,
950
+ "learning_rate": 9.480109779242805e-05,
951
+ "loss": 0.6496,
952
+ "step": 1160
953
+ },
954
+ {
955
+ "epoch": 0.9183673469387755,
956
+ "grad_norm": 4.71875,
957
+ "learning_rate": 9.464501864065735e-05,
958
+ "loss": 0.584,
959
+ "step": 1170
960
+ },
961
+ {
962
+ "epoch": 0.9262166405023547,
963
+ "grad_norm": 4.1875,
964
+ "learning_rate": 9.448676355366282e-05,
965
+ "loss": 0.6644,
966
+ "step": 1180
967
+ },
968
+ {
969
+ "epoch": 0.9340659340659341,
970
+ "grad_norm": 3.890625,
971
+ "learning_rate": 9.432634024457414e-05,
972
+ "loss": 0.6077,
973
+ "step": 1190
974
+ },
975
+ {
976
+ "epoch": 0.9419152276295133,
977
+ "grad_norm": 6.1875,
978
+ "learning_rate": 9.416375653219709e-05,
979
+ "loss": 0.5531,
980
+ "step": 1200
981
+ },
982
+ {
983
+ "epoch": 0.9497645211930926,
984
+ "grad_norm": 3.921875,
985
+ "learning_rate": 9.399902034063242e-05,
986
+ "loss": 0.6593,
987
+ "step": 1210
988
+ },
989
+ {
990
+ "epoch": 0.957613814756672,
991
+ "grad_norm": 4.15625,
992
+ "learning_rate": 9.383213969888972e-05,
993
+ "loss": 0.5477,
994
+ "step": 1220
995
+ },
996
+ {
997
+ "epoch": 0.9654631083202512,
998
+ "grad_norm": 4.09375,
999
+ "learning_rate": 9.366312274049602e-05,
1000
+ "loss": 0.6071,
1001
+ "step": 1230
1002
+ },
1003
+ {
1004
+ "epoch": 0.9733124018838305,
1005
+ "grad_norm": 4.34375,
1006
+ "learning_rate": 9.34919777030994e-05,
1007
+ "loss": 0.5946,
1008
+ "step": 1240
1009
+ },
1010
+ {
1011
+ "epoch": 0.9811616954474097,
1012
+ "grad_norm": 4.8125,
1013
+ "learning_rate": 9.331871292806759e-05,
1014
+ "loss": 0.5837,
1015
+ "step": 1250
1016
+ },
1017
+ {
1018
+ "epoch": 0.989010989010989,
1019
+ "grad_norm": 5.1875,
1020
+ "learning_rate": 9.314333686008125e-05,
1021
+ "loss": 0.6056,
1022
+ "step": 1260
1023
+ },
1024
+ {
1025
+ "epoch": 0.9968602825745683,
1026
+ "grad_norm": 4.90625,
1027
+ "learning_rate": 9.296585804672252e-05,
1028
+ "loss": 0.5941,
1029
+ "step": 1270
1030
+ },
1031
+ {
1032
+ "epoch": 1.0047095761381475,
1033
+ "grad_norm": 5.03125,
1034
+ "learning_rate": 9.278628513805838e-05,
1035
+ "loss": 0.6723,
1036
+ "step": 1280
1037
+ },
1038
+ {
1039
+ "epoch": 1.012558869701727,
1040
+ "grad_norm": 3.875,
1041
+ "learning_rate": 9.260462688621905e-05,
1042
+ "loss": 0.6221,
1043
+ "step": 1290
1044
+ },
1045
+ {
1046
+ "epoch": 1.0204081632653061,
1047
+ "grad_norm": 4.3125,
1048
+ "learning_rate": 9.242089214497145e-05,
1049
+ "loss": 0.5627,
1050
+ "step": 1300
1051
+ },
1052
+ {
1053
+ "epoch": 1.0282574568288854,
1054
+ "grad_norm": 4.6875,
1055
+ "learning_rate": 9.223508986928766e-05,
1056
+ "loss": 0.6045,
1057
+ "step": 1310
1058
+ },
1059
+ {
1060
+ "epoch": 1.0361067503924646,
1061
+ "grad_norm": 3.515625,
1062
+ "learning_rate": 9.204722911490846e-05,
1063
+ "loss": 0.5723,
1064
+ "step": 1320
1065
+ },
1066
+ {
1067
+ "epoch": 1.043956043956044,
1068
+ "grad_norm": 5.625,
1069
+ "learning_rate": 9.1857319037902e-05,
1070
+ "loss": 0.5616,
1071
+ "step": 1330
1072
+ },
1073
+ {
1074
+ "epoch": 1.0518053375196232,
1075
+ "grad_norm": 5.5625,
1076
+ "learning_rate": 9.16653688942175e-05,
1077
+ "loss": 0.5994,
1078
+ "step": 1340
1079
+ },
1080
+ {
1081
+ "epoch": 1.0596546310832025,
1082
+ "grad_norm": 4.65625,
1083
+ "learning_rate": 9.147138803923416e-05,
1084
+ "loss": 0.6163,
1085
+ "step": 1350
1086
+ },
1087
+ {
1088
+ "epoch": 1.0675039246467817,
1089
+ "grad_norm": 8.0,
1090
+ "learning_rate": 9.127538592730519e-05,
1091
+ "loss": 0.6001,
1092
+ "step": 1360
1093
+ },
1094
+ {
1095
+ "epoch": 1.0753532182103611,
1096
+ "grad_norm": 4.4375,
1097
+ "learning_rate": 9.107737211129701e-05,
1098
+ "loss": 0.55,
1099
+ "step": 1370
1100
+ },
1101
+ {
1102
+ "epoch": 1.0832025117739403,
1103
+ "grad_norm": 5.21875,
1104
+ "learning_rate": 9.087735624212365e-05,
1105
+ "loss": 0.5109,
1106
+ "step": 1380
1107
+ },
1108
+ {
1109
+ "epoch": 1.0910518053375196,
1110
+ "grad_norm": 4.59375,
1111
+ "learning_rate": 9.06753480682764e-05,
1112
+ "loss": 0.6688,
1113
+ "step": 1390
1114
+ },
1115
+ {
1116
+ "epoch": 1.098901098901099,
1117
+ "grad_norm": 5.3125,
1118
+ "learning_rate": 9.047135743534868e-05,
1119
+ "loss": 0.6504,
1120
+ "step": 1400
1121
+ },
1122
+ {
1123
+ "epoch": 1.1067503924646782,
1124
+ "grad_norm": 4.625,
1125
+ "learning_rate": 9.02653942855561e-05,
1126
+ "loss": 0.6315,
1127
+ "step": 1410
1128
+ },
1129
+ {
1130
+ "epoch": 1.1145996860282574,
1131
+ "grad_norm": 6.25,
1132
+ "learning_rate": 9.005746865725207e-05,
1133
+ "loss": 0.6297,
1134
+ "step": 1420
1135
+ },
1136
+ {
1137
+ "epoch": 1.1224489795918366,
1138
+ "grad_norm": 5.5625,
1139
+ "learning_rate": 8.984759068443831e-05,
1140
+ "loss": 0.5962,
1141
+ "step": 1430
1142
+ },
1143
+ {
1144
+ "epoch": 1.130298273155416,
1145
+ "grad_norm": 5.0625,
1146
+ "learning_rate": 8.963577059627118e-05,
1147
+ "loss": 0.5681,
1148
+ "step": 1440
1149
+ },
1150
+ {
1151
+ "epoch": 1.1381475667189953,
1152
+ "grad_norm": 4.78125,
1153
+ "learning_rate": 8.942201871656291e-05,
1154
+ "loss": 0.6486,
1155
+ "step": 1450
1156
+ },
1157
+ {
1158
+ "epoch": 1.1459968602825745,
1159
+ "grad_norm": 5.40625,
1160
+ "learning_rate": 8.920634546327858e-05,
1161
+ "loss": 0.5462,
1162
+ "step": 1460
1163
+ },
1164
+ {
1165
+ "epoch": 1.1538461538461537,
1166
+ "grad_norm": 5.6875,
1167
+ "learning_rate": 8.898876134802826e-05,
1168
+ "loss": 0.5653,
1169
+ "step": 1470
1170
+ },
1171
+ {
1172
+ "epoch": 1.1616954474097332,
1173
+ "grad_norm": 4.5625,
1174
+ "learning_rate": 8.87692769755548e-05,
1175
+ "loss": 0.492,
1176
+ "step": 1480
1177
+ },
1178
+ {
1179
+ "epoch": 1.1695447409733124,
1180
+ "grad_norm": 4.34375,
1181
+ "learning_rate": 8.854790304321681e-05,
1182
+ "loss": 0.5613,
1183
+ "step": 1490
1184
+ },
1185
+ {
1186
+ "epoch": 1.1773940345368916,
1187
+ "grad_norm": 5.15625,
1188
+ "learning_rate": 8.832465034046749e-05,
1189
+ "loss": 0.5728,
1190
+ "step": 1500
1191
+ },
1192
+ {
1193
+ "epoch": 1.185243328100471,
1194
+ "grad_norm": 4.34375,
1195
+ "learning_rate": 8.80995297483286e-05,
1196
+ "loss": 0.5405,
1197
+ "step": 1510
1198
+ },
1199
+ {
1200
+ "epoch": 1.1930926216640503,
1201
+ "grad_norm": 4.53125,
1202
+ "learning_rate": 8.787255223886019e-05,
1203
+ "loss": 0.5618,
1204
+ "step": 1520
1205
+ },
1206
+ {
1207
+ "epoch": 1.2009419152276295,
1208
+ "grad_norm": 4.875,
1209
+ "learning_rate": 8.764372887462586e-05,
1210
+ "loss": 0.5731,
1211
+ "step": 1530
1212
+ },
1213
+ {
1214
+ "epoch": 1.2087912087912087,
1215
+ "grad_norm": 4.71875,
1216
+ "learning_rate": 8.741307080815357e-05,
1217
+ "loss": 0.5745,
1218
+ "step": 1540
1219
+ },
1220
+ {
1221
+ "epoch": 1.2166405023547882,
1222
+ "grad_norm": 5.15625,
1223
+ "learning_rate": 8.718058928139205e-05,
1224
+ "loss": 0.5266,
1225
+ "step": 1550
1226
+ },
1227
+ {
1228
+ "epoch": 1.2244897959183674,
1229
+ "grad_norm": 4.53125,
1230
+ "learning_rate": 8.694629562516294e-05,
1231
+ "loss": 0.6069,
1232
+ "step": 1560
1233
+ },
1234
+ {
1235
+ "epoch": 1.2323390894819466,
1236
+ "grad_norm": 3.5,
1237
+ "learning_rate": 8.671020125860851e-05,
1238
+ "loss": 0.5433,
1239
+ "step": 1570
1240
+ },
1241
+ {
1242
+ "epoch": 1.2401883830455258,
1243
+ "grad_norm": 4.0,
1244
+ "learning_rate": 8.647231768863512e-05,
1245
+ "loss": 0.4842,
1246
+ "step": 1580
1247
+ },
1248
+ {
1249
+ "epoch": 1.2480376766091053,
1250
+ "grad_norm": 4.6875,
1251
+ "learning_rate": 8.623265650935234e-05,
1252
+ "loss": 0.58,
1253
+ "step": 1590
1254
+ },
1255
+ {
1256
+ "epoch": 1.2558869701726845,
1257
+ "grad_norm": 4.125,
1258
+ "learning_rate": 8.599122940150795e-05,
1259
+ "loss": 0.5331,
1260
+ "step": 1600
1261
+ },
1262
+ {
1263
+ "epoch": 1.2637362637362637,
1264
+ "grad_norm": 6.0,
1265
+ "learning_rate": 8.574804813191858e-05,
1266
+ "loss": 0.6408,
1267
+ "step": 1610
1268
+ },
1269
+ {
1270
+ "epoch": 1.2715855572998431,
1271
+ "grad_norm": 4.90625,
1272
+ "learning_rate": 8.550312455289625e-05,
1273
+ "loss": 0.5044,
1274
+ "step": 1620
1275
+ },
1276
+ {
1277
+ "epoch": 1.2794348508634223,
1278
+ "grad_norm": 4.96875,
1279
+ "learning_rate": 8.525647060167063e-05,
1280
+ "loss": 0.6098,
1281
+ "step": 1630
1282
+ },
1283
+ {
1284
+ "epoch": 1.2872841444270016,
1285
+ "grad_norm": 4.1875,
1286
+ "learning_rate": 8.500809829980734e-05,
1287
+ "loss": 0.5911,
1288
+ "step": 1640
1289
+ },
1290
+ {
1291
+ "epoch": 1.2951334379905808,
1292
+ "grad_norm": 5.3125,
1293
+ "learning_rate": 8.4758019752622e-05,
1294
+ "loss": 0.5876,
1295
+ "step": 1650
1296
+ },
1297
+ {
1298
+ "epoch": 1.30298273155416,
1299
+ "grad_norm": 5.09375,
1300
+ "learning_rate": 8.450624714859016e-05,
1301
+ "loss": 0.5618,
1302
+ "step": 1660
1303
+ },
1304
+ {
1305
+ "epoch": 1.3108320251177394,
1306
+ "grad_norm": 3.78125,
1307
+ "learning_rate": 8.425279275875336e-05,
1308
+ "loss": 0.5629,
1309
+ "step": 1670
1310
+ },
1311
+ {
1312
+ "epoch": 1.3186813186813187,
1313
+ "grad_norm": 3.953125,
1314
+ "learning_rate": 8.399766893612096e-05,
1315
+ "loss": 0.5672,
1316
+ "step": 1680
1317
+ },
1318
+ {
1319
+ "epoch": 1.3265306122448979,
1320
+ "grad_norm": 5.5,
1321
+ "learning_rate": 8.374088811506819e-05,
1322
+ "loss": 0.6229,
1323
+ "step": 1690
1324
+ },
1325
+ {
1326
+ "epoch": 1.3343799058084773,
1327
+ "grad_norm": 7.78125,
1328
+ "learning_rate": 8.348246281072998e-05,
1329
+ "loss": 0.5945,
1330
+ "step": 1700
1331
+ },
1332
+ {
1333
+ "epoch": 1.3422291993720565,
1334
+ "grad_norm": 4.125,
1335
+ "learning_rate": 8.322240561839109e-05,
1336
+ "loss": 0.5755,
1337
+ "step": 1710
1338
+ },
1339
+ {
1340
+ "epoch": 1.3500784929356358,
1341
+ "grad_norm": 5.90625,
1342
+ "learning_rate": 8.296072921287217e-05,
1343
+ "loss": 0.5472,
1344
+ "step": 1720
1345
+ },
1346
+ {
1347
+ "epoch": 1.3579277864992152,
1348
+ "grad_norm": 6.34375,
1349
+ "learning_rate": 8.269744634791208e-05,
1350
+ "loss": 0.6287,
1351
+ "step": 1730
1352
+ },
1353
+ {
1354
+ "epoch": 1.3657770800627944,
1355
+ "grad_norm": 4.34375,
1356
+ "learning_rate": 8.243256985554621e-05,
1357
+ "loss": 0.5684,
1358
+ "step": 1740
1359
+ },
1360
+ {
1361
+ "epoch": 1.3736263736263736,
1362
+ "grad_norm": 5.875,
1363
+ "learning_rate": 8.21661126454811e-05,
1364
+ "loss": 0.5455,
1365
+ "step": 1750
1366
+ },
1367
+ {
1368
+ "epoch": 1.3814756671899528,
1369
+ "grad_norm": 4.3125,
1370
+ "learning_rate": 8.189808770446528e-05,
1371
+ "loss": 0.5403,
1372
+ "step": 1760
1373
+ },
1374
+ {
1375
+ "epoch": 1.389324960753532,
1376
+ "grad_norm": 5.03125,
1377
+ "learning_rate": 8.162850809565623e-05,
1378
+ "loss": 0.5343,
1379
+ "step": 1770
1380
+ },
1381
+ {
1382
+ "epoch": 1.3971742543171115,
1383
+ "grad_norm": 4.53125,
1384
+ "learning_rate": 8.135738695798376e-05,
1385
+ "loss": 0.5692,
1386
+ "step": 1780
1387
+ },
1388
+ {
1389
+ "epoch": 1.4050235478806907,
1390
+ "grad_norm": 5.125,
1391
+ "learning_rate": 8.108473750550965e-05,
1392
+ "loss": 0.5382,
1393
+ "step": 1790
1394
+ },
1395
+ {
1396
+ "epoch": 1.41287284144427,
1397
+ "grad_norm": 5.6875,
1398
+ "learning_rate": 8.081057302678352e-05,
1399
+ "loss": 0.5316,
1400
+ "step": 1800
1401
+ },
1402
+ {
1403
+ "epoch": 1.4207221350078494,
1404
+ "grad_norm": 5.15625,
1405
+ "learning_rate": 8.053490688419533e-05,
1406
+ "loss": 0.5221,
1407
+ "step": 1810
1408
+ },
1409
+ {
1410
+ "epoch": 1.4285714285714286,
1411
+ "grad_norm": 6.6875,
1412
+ "learning_rate": 8.02577525133239e-05,
1413
+ "loss": 0.6501,
1414
+ "step": 1820
1415
+ },
1416
+ {
1417
+ "epoch": 1.4364207221350078,
1418
+ "grad_norm": 7.34375,
1419
+ "learning_rate": 7.997912342228232e-05,
1420
+ "loss": 0.5745,
1421
+ "step": 1830
1422
+ },
1423
+ {
1424
+ "epoch": 1.4442700156985873,
1425
+ "grad_norm": 4.15625,
1426
+ "learning_rate": 7.969903319105935e-05,
1427
+ "loss": 0.5797,
1428
+ "step": 1840
1429
+ },
1430
+ {
1431
+ "epoch": 1.4521193092621665,
1432
+ "grad_norm": 6.21875,
1433
+ "learning_rate": 7.941749547085777e-05,
1434
+ "loss": 0.6899,
1435
+ "step": 1850
1436
+ },
1437
+ {
1438
+ "epoch": 1.4599686028257457,
1439
+ "grad_norm": 4.625,
1440
+ "learning_rate": 7.913452398342881e-05,
1441
+ "loss": 0.5891,
1442
+ "step": 1860
1443
+ },
1444
+ {
1445
+ "epoch": 1.467817896389325,
1446
+ "grad_norm": 7.1875,
1447
+ "learning_rate": 7.885013252040359e-05,
1448
+ "loss": 0.6118,
1449
+ "step": 1870
1450
+ },
1451
+ {
1452
+ "epoch": 1.4756671899529041,
1453
+ "grad_norm": 6.8125,
1454
+ "learning_rate": 7.856433494262077e-05,
1455
+ "loss": 0.6102,
1456
+ "step": 1880
1457
+ },
1458
+ {
1459
+ "epoch": 1.4835164835164836,
1460
+ "grad_norm": 5.0,
1461
+ "learning_rate": 7.827714517945115e-05,
1462
+ "loss": 0.5459,
1463
+ "step": 1890
1464
+ },
1465
+ {
1466
+ "epoch": 1.4913657770800628,
1467
+ "grad_norm": 4.40625,
1468
+ "learning_rate": 7.798857722811856e-05,
1469
+ "loss": 0.5717,
1470
+ "step": 1900
1471
+ },
1472
+ {
1473
+ "epoch": 1.499215070643642,
1474
+ "grad_norm": 3.734375,
1475
+ "learning_rate": 7.769864515301787e-05,
1476
+ "loss": 0.5279,
1477
+ "step": 1910
1478
+ },
1479
+ {
1480
+ "epoch": 1.5070643642072215,
1481
+ "grad_norm": 4.6875,
1482
+ "learning_rate": 7.740736308502938e-05,
1483
+ "loss": 0.66,
1484
+ "step": 1920
1485
+ },
1486
+ {
1487
+ "epoch": 1.5149136577708007,
1488
+ "grad_norm": 6.5,
1489
+ "learning_rate": 7.711474522083015e-05,
1490
+ "loss": 0.546,
1491
+ "step": 1930
1492
+ },
1493
+ {
1494
+ "epoch": 1.5227629513343799,
1495
+ "grad_norm": 4.625,
1496
+ "learning_rate": 7.682080582220206e-05,
1497
+ "loss": 0.5207,
1498
+ "step": 1940
1499
+ },
1500
+ {
1501
+ "epoch": 1.5306122448979593,
1502
+ "grad_norm": 6.4375,
1503
+ "learning_rate": 7.65255592153367e-05,
1504
+ "loss": 0.6005,
1505
+ "step": 1950
1506
+ },
1507
+ {
1508
+ "epoch": 1.5384615384615383,
1509
+ "grad_norm": 4.15625,
1510
+ "learning_rate": 7.622901979013717e-05,
1511
+ "loss": 0.5897,
1512
+ "step": 1960
1513
+ },
1514
+ {
1515
+ "epoch": 1.5463108320251178,
1516
+ "grad_norm": 4.40625,
1517
+ "learning_rate": 7.593120199951671e-05,
1518
+ "loss": 0.5639,
1519
+ "step": 1970
1520
+ },
1521
+ {
1522
+ "epoch": 1.554160125588697,
1523
+ "grad_norm": 4.59375,
1524
+ "learning_rate": 7.563212035869425e-05,
1525
+ "loss": 0.5479,
1526
+ "step": 1980
1527
+ },
1528
+ {
1529
+ "epoch": 1.5620094191522762,
1530
+ "grad_norm": 4.8125,
1531
+ "learning_rate": 7.533178944448705e-05,
1532
+ "loss": 0.5539,
1533
+ "step": 1990
1534
+ },
1535
+ {
1536
+ "epoch": 1.5698587127158556,
1537
+ "grad_norm": 6.09375,
1538
+ "learning_rate": 7.503022389460014e-05,
1539
+ "loss": 0.625,
1540
+ "step": 2000
1541
+ },
1542
+ {
1543
+ "epoch": 1.5698587127158556,
1544
+ "eval/nano_beir.arguana": 0.4666002720912347,
1545
+ "step": 2000
1546
+ },
1547
+ {
1548
+ "epoch": 1.5698587127158556,
1549
+ "eval/nano_beir.climate_fever": 0.22462604477746737,
1550
+ "step": 2000
1551
+ },
1552
+ {
1553
+ "epoch": 1.5698587127158556,
1554
+ "eval/nano_beir.dbpedia_entity": 0.3175867497382319,
1555
+ "step": 2000
1556
+ },
1557
+ {
1558
+ "epoch": 1.5698587127158556,
1559
+ "eval/nano_beir.fever": 0.6140103747506924,
1560
+ "step": 2000
1561
+ },
1562
+ {
1563
+ "epoch": 1.5698587127158556,
1564
+ "eval/nano_beir.fiqa": 0.43978199849769206,
1565
+ "step": 2000
1566
+ },
1567
+ {
1568
+ "epoch": 1.5698587127158556,
1569
+ "eval/nano_beir.hotpotqa": 0.6788521964877974,
1570
+ "step": 2000
1571
+ },
1572
+ {
1573
+ "epoch": 1.5698587127158556,
1574
+ "eval/nano_beir.nfcorpus": 0.3485625465712446,
1575
+ "step": 2000
1576
+ },
1577
+ {
1578
+ "epoch": 1.5698587127158556,
1579
+ "eval/nano_beir.nq": 0.42598841107613383,
1580
+ "step": 2000
1581
+ },
1582
+ {
1583
+ "epoch": 1.5698587127158556,
1584
+ "eval/nano_beir.quora": 0.833039807676506,
1585
+ "step": 2000
1586
+ },
1587
+ {
1588
+ "epoch": 1.5698587127158556,
1589
+ "eval/nano_beir.scidocs": 0.24848467409467515,
1590
+ "step": 2000
1591
+ },
1592
+ {
1593
+ "epoch": 1.5698587127158556,
1594
+ "eval/nano_beir.scifact": 0.7682290881646785,
1595
+ "step": 2000
1596
+ },
1597
+ {
1598
+ "epoch": 1.5698587127158556,
1599
+ "eval/nano_beir.webis_touche2020": 0.2836464597787185,
1600
+ "step": 2000
1601
+ },
1602
+ {
1603
+ "epoch": 1.5698587127158556,
1604
+ "eval/avg": 0.47078405197542267,
1605
+ "step": 2000
1606
+ },
1607
+ {
1608
+ "epoch": 1.5777080062794349,
1609
+ "grad_norm": 5.15625,
1610
+ "learning_rate": 7.472743840691299e-05,
1611
+ "loss": 0.4956,
1612
+ "step": 2010
1613
+ },
1614
+ {
1615
+ "epoch": 1.585557299843014,
1616
+ "grad_norm": 4.34375,
1617
+ "learning_rate": 7.44234477387631e-05,
1618
+ "loss": 0.4504,
1619
+ "step": 2020
1620
+ },
1621
+ {
1622
+ "epoch": 1.5934065934065935,
1623
+ "grad_norm": 3.0625,
1624
+ "learning_rate": 7.411826670622676e-05,
1625
+ "loss": 0.5722,
1626
+ "step": 2030
1627
+ },
1628
+ {
1629
+ "epoch": 1.6012558869701727,
1630
+ "grad_norm": 4.875,
1631
+ "learning_rate": 7.381191018339696e-05,
1632
+ "loss": 0.5777,
1633
+ "step": 2040
1634
+ },
1635
+ {
1636
+ "epoch": 1.609105180533752,
1637
+ "grad_norm": 5.34375,
1638
+ "learning_rate": 7.350439310165842e-05,
1639
+ "loss": 0.5645,
1640
+ "step": 2050
1641
+ },
1642
+ {
1643
+ "epoch": 1.6169544740973314,
1644
+ "grad_norm": 5.125,
1645
+ "learning_rate": 7.319573044895986e-05,
1646
+ "loss": 0.5454,
1647
+ "step": 2060
1648
+ },
1649
+ {
1650
+ "epoch": 1.6248037676609104,
1651
+ "grad_norm": 5.4375,
1652
+ "learning_rate": 7.28859372690835e-05,
1653
+ "loss": 0.5681,
1654
+ "step": 2070
1655
+ },
1656
+ {
1657
+ "epoch": 1.6326530612244898,
1658
+ "grad_norm": 4.96875,
1659
+ "learning_rate": 7.257502866091192e-05,
1660
+ "loss": 0.5132,
1661
+ "step": 2080
1662
+ },
1663
+ {
1664
+ "epoch": 1.640502354788069,
1665
+ "grad_norm": 5.125,
1666
+ "learning_rate": 7.226301977769199e-05,
1667
+ "loss": 0.5463,
1668
+ "step": 2090
1669
+ },
1670
+ {
1671
+ "epoch": 1.6483516483516483,
1672
+ "grad_norm": 4.71875,
1673
+ "learning_rate": 7.194992582629654e-05,
1674
+ "loss": 0.6572,
1675
+ "step": 2100
1676
+ },
1677
+ {
1678
+ "epoch": 1.6562009419152277,
1679
+ "grad_norm": 4.625,
1680
+ "learning_rate": 7.163576206648304e-05,
1681
+ "loss": 0.5608,
1682
+ "step": 2110
1683
+ },
1684
+ {
1685
+ "epoch": 1.664050235478807,
1686
+ "grad_norm": 4.8125,
1687
+ "learning_rate": 7.132054381014995e-05,
1688
+ "loss": 0.5352,
1689
+ "step": 2120
1690
+ },
1691
+ {
1692
+ "epoch": 1.6718995290423861,
1693
+ "grad_norm": 6.25,
1694
+ "learning_rate": 7.100428642059033e-05,
1695
+ "loss": 0.6148,
1696
+ "step": 2130
1697
+ },
1698
+ {
1699
+ "epoch": 1.6797488226059656,
1700
+ "grad_norm": 4.75,
1701
+ "learning_rate": 7.068700531174319e-05,
1702
+ "loss": 0.6296,
1703
+ "step": 2140
1704
+ },
1705
+ {
1706
+ "epoch": 1.6875981161695446,
1707
+ "grad_norm": 5.125,
1708
+ "learning_rate": 7.036871594744217e-05,
1709
+ "loss": 0.5741,
1710
+ "step": 2150
1711
+ },
1712
+ {
1713
+ "epoch": 1.695447409733124,
1714
+ "grad_norm": 4.78125,
1715
+ "learning_rate": 7.004943384066187e-05,
1716
+ "loss": 0.5457,
1717
+ "step": 2160
1718
+ },
1719
+ {
1720
+ "epoch": 1.7032967032967035,
1721
+ "grad_norm": 5.1875,
1722
+ "learning_rate": 6.97291745527617e-05,
1723
+ "loss": 0.6556,
1724
+ "step": 2170
1725
+ },
1726
+ {
1727
+ "epoch": 1.7111459968602825,
1728
+ "grad_norm": 5.875,
1729
+ "learning_rate": 6.940795369272754e-05,
1730
+ "loss": 0.6284,
1731
+ "step": 2180
1732
+ },
1733
+ {
1734
+ "epoch": 1.718995290423862,
1735
+ "grad_norm": 4.40625,
1736
+ "learning_rate": 6.908578691641092e-05,
1737
+ "loss": 0.5214,
1738
+ "step": 2190
1739
+ },
1740
+ {
1741
+ "epoch": 1.7268445839874411,
1742
+ "grad_norm": 5.96875,
1743
+ "learning_rate": 6.876268992576604e-05,
1744
+ "loss": 0.5484,
1745
+ "step": 2200
1746
+ },
1747
+ {
1748
+ "epoch": 1.7346938775510203,
1749
+ "grad_norm": 5.75,
1750
+ "learning_rate": 6.843867846808438e-05,
1751
+ "loss": 0.5804,
1752
+ "step": 2210
1753
+ },
1754
+ {
1755
+ "epoch": 1.7425431711145998,
1756
+ "grad_norm": 5.59375,
1757
+ "learning_rate": 6.811376833522729e-05,
1758
+ "loss": 0.5653,
1759
+ "step": 2220
1760
+ },
1761
+ {
1762
+ "epoch": 1.750392464678179,
1763
+ "grad_norm": 4.09375,
1764
+ "learning_rate": 6.778797536285624e-05,
1765
+ "loss": 0.5328,
1766
+ "step": 2230
1767
+ },
1768
+ {
1769
+ "epoch": 1.7582417582417582,
1770
+ "grad_norm": 4.0,
1771
+ "learning_rate": 6.746131542966112e-05,
1772
+ "loss": 0.5909,
1773
+ "step": 2240
1774
+ },
1775
+ {
1776
+ "epoch": 1.7660910518053377,
1777
+ "grad_norm": 5.6875,
1778
+ "learning_rate": 6.713380445658618e-05,
1779
+ "loss": 0.5715,
1780
+ "step": 2250
1781
+ },
1782
+ {
1783
+ "epoch": 1.7739403453689166,
1784
+ "grad_norm": 5.09375,
1785
+ "learning_rate": 6.680545840605423e-05,
1786
+ "loss": 0.4487,
1787
+ "step": 2260
1788
+ },
1789
+ {
1790
+ "epoch": 1.781789638932496,
1791
+ "grad_norm": 4.9375,
1792
+ "learning_rate": 6.647629328118851e-05,
1793
+ "loss": 0.5661,
1794
+ "step": 2270
1795
+ },
1796
+ {
1797
+ "epoch": 1.7896389324960753,
1798
+ "grad_norm": 4.65625,
1799
+ "learning_rate": 6.614632512503288e-05,
1800
+ "loss": 0.5829,
1801
+ "step": 2280
1802
+ },
1803
+ {
1804
+ "epoch": 1.7974882260596545,
1805
+ "grad_norm": 5.1875,
1806
+ "learning_rate": 6.58155700197697e-05,
1807
+ "loss": 0.5994,
1808
+ "step": 2290
1809
+ },
1810
+ {
1811
+ "epoch": 1.805337519623234,
1812
+ "grad_norm": 4.71875,
1813
+ "learning_rate": 6.548404408593621e-05,
1814
+ "loss": 0.579,
1815
+ "step": 2300
1816
+ },
1817
+ {
1818
+ "epoch": 1.8131868131868132,
1819
+ "grad_norm": 4.25,
1820
+ "learning_rate": 6.515176348163871e-05,
1821
+ "loss": 0.488,
1822
+ "step": 2310
1823
+ },
1824
+ {
1825
+ "epoch": 1.8210361067503924,
1826
+ "grad_norm": 4.25,
1827
+ "learning_rate": 6.481874440176506e-05,
1828
+ "loss": 0.5083,
1829
+ "step": 2320
1830
+ },
1831
+ {
1832
+ "epoch": 1.8288854003139718,
1833
+ "grad_norm": 5.6875,
1834
+ "learning_rate": 6.448500307719538e-05,
1835
+ "loss": 0.6211,
1836
+ "step": 2330
1837
+ },
1838
+ {
1839
+ "epoch": 1.836734693877551,
1840
+ "grad_norm": 4.96875,
1841
+ "learning_rate": 6.415055577401102e-05,
1842
+ "loss": 0.5724,
1843
+ "step": 2340
1844
+ },
1845
+ {
1846
+ "epoch": 1.8445839874411303,
1847
+ "grad_norm": 3.90625,
1848
+ "learning_rate": 6.381541879270169e-05,
1849
+ "loss": 0.5523,
1850
+ "step": 2350
1851
+ },
1852
+ {
1853
+ "epoch": 1.8524332810047097,
1854
+ "grad_norm": 4.03125,
1855
+ "learning_rate": 6.347960846737106e-05,
1856
+ "loss": 0.5615,
1857
+ "step": 2360
1858
+ },
1859
+ {
1860
+ "epoch": 1.8602825745682887,
1861
+ "grad_norm": 4.375,
1862
+ "learning_rate": 6.31431411649406e-05,
1863
+ "loss": 0.4643,
1864
+ "step": 2370
1865
+ },
1866
+ {
1867
+ "epoch": 1.8681318681318682,
1868
+ "grad_norm": 5.09375,
1869
+ "learning_rate": 6.280603328435198e-05,
1870
+ "loss": 0.5643,
1871
+ "step": 2380
1872
+ },
1873
+ {
1874
+ "epoch": 1.8759811616954474,
1875
+ "grad_norm": 5.375,
1876
+ "learning_rate": 6.24683012557677e-05,
1877
+ "loss": 0.5446,
1878
+ "step": 2390
1879
+ },
1880
+ {
1881
+ "epoch": 1.8838304552590266,
1882
+ "grad_norm": 4.84375,
1883
+ "learning_rate": 6.212996153977037e-05,
1884
+ "loss": 0.5759,
1885
+ "step": 2400
1886
+ },
1887
+ {
1888
+ "epoch": 1.891679748822606,
1889
+ "grad_norm": 5.46875,
1890
+ "learning_rate": 6.179103062656042e-05,
1891
+ "loss": 0.5231,
1892
+ "step": 2410
1893
+ },
1894
+ {
1895
+ "epoch": 1.8995290423861853,
1896
+ "grad_norm": 5.5625,
1897
+ "learning_rate": 6.145152503515239e-05,
1898
+ "loss": 0.5679,
1899
+ "step": 2420
1900
+ },
1901
+ {
1902
+ "epoch": 1.9073783359497645,
1903
+ "grad_norm": 4.78125,
1904
+ "learning_rate": 6.111146131256983e-05,
1905
+ "loss": 0.5482,
1906
+ "step": 2430
1907
+ },
1908
+ {
1909
+ "epoch": 1.915227629513344,
1910
+ "grad_norm": 5.1875,
1911
+ "learning_rate": 6.077085603303882e-05,
1912
+ "loss": 0.5287,
1913
+ "step": 2440
1914
+ },
1915
+ {
1916
+ "epoch": 1.9230769230769231,
1917
+ "grad_norm": 5.8125,
1918
+ "learning_rate": 6.042972579718019e-05,
1919
+ "loss": 0.6531,
1920
+ "step": 2450
1921
+ },
1922
+ {
1923
+ "epoch": 1.9309262166405023,
1924
+ "grad_norm": 4.75,
1925
+ "learning_rate": 6.008808723120035e-05,
1926
+ "loss": 0.5463,
1927
+ "step": 2460
1928
+ },
1929
+ {
1930
+ "epoch": 1.9387755102040818,
1931
+ "grad_norm": 4.0625,
1932
+ "learning_rate": 5.974595698608103e-05,
1933
+ "loss": 0.5132,
1934
+ "step": 2470
1935
+ },
1936
+ {
1937
+ "epoch": 1.9466248037676608,
1938
+ "grad_norm": 4.6875,
1939
+ "learning_rate": 5.9403351736767696e-05,
1940
+ "loss": 0.6158,
1941
+ "step": 2480
1942
+ },
1943
+ {
1944
+ "epoch": 1.9544740973312402,
1945
+ "grad_norm": 4.84375,
1946
+ "learning_rate": 5.906028818135687e-05,
1947
+ "loss": 0.5588,
1948
+ "step": 2490
1949
+ },
1950
+ {
1951
+ "epoch": 1.9623233908948194,
1952
+ "grad_norm": 5.125,
1953
+ "learning_rate": 5.8716783040282244e-05,
1954
+ "loss": 0.5456,
1955
+ "step": 2500
1956
+ },
1957
+ {
1958
+ "epoch": 1.9701726844583987,
1959
+ "grad_norm": 4.65625,
1960
+ "learning_rate": 5.837285305549978e-05,
1961
+ "loss": 0.6038,
1962
+ "step": 2510
1963
+ },
1964
+ {
1965
+ "epoch": 1.978021978021978,
1966
+ "grad_norm": 5.25,
1967
+ "learning_rate": 5.8028514989671724e-05,
1968
+ "loss": 0.5504,
1969
+ "step": 2520
1970
+ },
1971
+ {
1972
+ "epoch": 1.9858712715855573,
1973
+ "grad_norm": 5.0,
1974
+ "learning_rate": 5.7683785625349616e-05,
1975
+ "loss": 0.5705,
1976
+ "step": 2530
1977
+ },
1978
+ {
1979
+ "epoch": 1.9937205651491365,
1980
+ "grad_norm": 6.21875,
1981
+ "learning_rate": 5.733868176415633e-05,
1982
+ "loss": 0.5866,
1983
+ "step": 2540
1984
+ },
1985
+ {
1986
+ "epoch": 2.001569858712716,
1987
+ "grad_norm": 5.125,
1988
+ "learning_rate": 5.699322022596722e-05,
1989
+ "loss": 0.6031,
1990
+ "step": 2550
1991
+ },
1992
+ {
1993
+ "epoch": 2.009419152276295,
1994
+ "grad_norm": 4.5625,
1995
+ "learning_rate": 5.664741784809022e-05,
1996
+ "loss": 0.5695,
1997
+ "step": 2560
1998
+ },
1999
+ {
2000
+ "epoch": 2.0172684458398744,
2001
+ "grad_norm": 5.65625,
2002
+ "learning_rate": 5.630129148444543e-05,
2003
+ "loss": 0.5491,
2004
+ "step": 2570
2005
+ },
2006
+ {
2007
+ "epoch": 2.025117739403454,
2008
+ "grad_norm": 3.78125,
2009
+ "learning_rate": 5.595485800474349e-05,
2010
+ "loss": 0.5192,
2011
+ "step": 2580
2012
+ },
2013
+ {
2014
+ "epoch": 2.032967032967033,
2015
+ "grad_norm": 6.25,
2016
+ "learning_rate": 5.560813429366345e-05,
2017
+ "loss": 0.5584,
2018
+ "step": 2590
2019
+ },
2020
+ {
2021
+ "epoch": 2.0408163265306123,
2022
+ "grad_norm": 4.78125,
2023
+ "learning_rate": 5.5261137250029835e-05,
2024
+ "loss": 0.4623,
2025
+ "step": 2600
2026
+ },
2027
+ {
2028
+ "epoch": 2.0486656200941917,
2029
+ "grad_norm": 4.53125,
2030
+ "learning_rate": 5.4913883785988993e-05,
2031
+ "loss": 0.5125,
2032
+ "step": 2610
2033
+ },
2034
+ {
2035
+ "epoch": 2.0565149136577707,
2036
+ "grad_norm": 4.5,
2037
+ "learning_rate": 5.456639082618489e-05,
2038
+ "loss": 0.4937,
2039
+ "step": 2620
2040
+ },
2041
+ {
2042
+ "epoch": 2.06436420722135,
2043
+ "grad_norm": 4.65625,
2044
+ "learning_rate": 5.421867530693414e-05,
2045
+ "loss": 0.5231,
2046
+ "step": 2630
2047
+ },
2048
+ {
2049
+ "epoch": 2.072213500784929,
2050
+ "grad_norm": 3.546875,
2051
+ "learning_rate": 5.3870754175400595e-05,
2052
+ "loss": 0.422,
2053
+ "step": 2640
2054
+ },
2055
+ {
2056
+ "epoch": 2.0800627943485086,
2057
+ "grad_norm": 4.40625,
2058
+ "learning_rate": 5.352264438876935e-05,
2059
+ "loss": 0.5353,
2060
+ "step": 2650
2061
+ },
2062
+ {
2063
+ "epoch": 2.087912087912088,
2064
+ "grad_norm": 4.34375,
2065
+ "learning_rate": 5.3174362913420306e-05,
2066
+ "loss": 0.4964,
2067
+ "step": 2660
2068
+ },
2069
+ {
2070
+ "epoch": 2.095761381475667,
2071
+ "grad_norm": 4.28125,
2072
+ "learning_rate": 5.2825926724101236e-05,
2073
+ "loss": 0.5701,
2074
+ "step": 2670
2075
+ },
2076
+ {
2077
+ "epoch": 2.1036106750392465,
2078
+ "grad_norm": 5.5,
2079
+ "learning_rate": 5.247735280310041e-05,
2080
+ "loss": 0.5823,
2081
+ "step": 2680
2082
+ },
2083
+ {
2084
+ "epoch": 2.111459968602826,
2085
+ "grad_norm": 4.125,
2086
+ "learning_rate": 5.212865813941899e-05,
2087
+ "loss": 0.4692,
2088
+ "step": 2690
2089
+ },
2090
+ {
2091
+ "epoch": 2.119309262166405,
2092
+ "grad_norm": 6.9375,
2093
+ "learning_rate": 5.1779859727942924e-05,
2094
+ "loss": 0.5522,
2095
+ "step": 2700
2096
+ },
2097
+ {
2098
+ "epoch": 2.1271585557299844,
2099
+ "grad_norm": 5.78125,
2100
+ "learning_rate": 5.143097456861474e-05,
2101
+ "loss": 0.5413,
2102
+ "step": 2710
2103
+ },
2104
+ {
2105
+ "epoch": 2.1350078492935634,
2106
+ "grad_norm": 4.625,
2107
+ "learning_rate": 5.1082019665604895e-05,
2108
+ "loss": 0.4929,
2109
+ "step": 2720
2110
+ },
2111
+ {
2112
+ "epoch": 2.142857142857143,
2113
+ "grad_norm": 5.5625,
2114
+ "learning_rate": 5.073301202648304e-05,
2115
+ "loss": 0.5183,
2116
+ "step": 2730
2117
+ },
2118
+ {
2119
+ "epoch": 2.1507064364207222,
2120
+ "grad_norm": 6.0,
2121
+ "learning_rate": 5.0383968661389146e-05,
2122
+ "loss": 0.5698,
2123
+ "step": 2740
2124
+ },
2125
+ {
2126
+ "epoch": 2.1585557299843012,
2127
+ "grad_norm": 5.0,
2128
+ "learning_rate": 5.003490658220438e-05,
2129
+ "loss": 0.4615,
2130
+ "step": 2750
2131
+ },
2132
+ {
2133
+ "epoch": 2.1664050235478807,
2134
+ "grad_norm": 6.09375,
2135
+ "learning_rate": 4.968584280172206e-05,
2136
+ "loss": 0.5037,
2137
+ "step": 2760
2138
+ },
2139
+ {
2140
+ "epoch": 2.17425431711146,
2141
+ "grad_norm": 7.5625,
2142
+ "learning_rate": 4.933679433281836e-05,
2143
+ "loss": 0.5184,
2144
+ "step": 2770
2145
+ },
2146
+ {
2147
+ "epoch": 2.182103610675039,
2148
+ "grad_norm": 5.8125,
2149
+ "learning_rate": 4.8987778187623245e-05,
2150
+ "loss": 0.5307,
2151
+ "step": 2780
2152
+ },
2153
+ {
2154
+ "epoch": 2.1899529042386185,
2155
+ "grad_norm": 5.28125,
2156
+ "learning_rate": 4.863881137669123e-05,
2157
+ "loss": 0.4549,
2158
+ "step": 2790
2159
+ },
2160
+ {
2161
+ "epoch": 2.197802197802198,
2162
+ "grad_norm": 4.03125,
2163
+ "learning_rate": 4.8289910908172376e-05,
2164
+ "loss": 0.5787,
2165
+ "step": 2800
2166
+ },
2167
+ {
2168
+ "epoch": 2.205651491365777,
2169
+ "grad_norm": 4.8125,
2170
+ "learning_rate": 4.794109378698327e-05,
2171
+ "loss": 0.6173,
2172
+ "step": 2810
2173
+ },
2174
+ {
2175
+ "epoch": 2.2135007849293564,
2176
+ "grad_norm": 5.6875,
2177
+ "learning_rate": 4.7592377013978306e-05,
2178
+ "loss": 0.5365,
2179
+ "step": 2820
2180
+ },
2181
+ {
2182
+ "epoch": 2.221350078492936,
2183
+ "grad_norm": 5.125,
2184
+ "learning_rate": 4.724377758512103e-05,
2185
+ "loss": 0.4651,
2186
+ "step": 2830
2187
+ },
2188
+ {
2189
+ "epoch": 2.229199372056515,
2190
+ "grad_norm": 6.15625,
2191
+ "learning_rate": 4.68953124906558e-05,
2192
+ "loss": 0.5659,
2193
+ "step": 2840
2194
+ },
2195
+ {
2196
+ "epoch": 2.2370486656200943,
2197
+ "grad_norm": 4.71875,
2198
+ "learning_rate": 4.654699871427971e-05,
2199
+ "loss": 0.6009,
2200
+ "step": 2850
2201
+ },
2202
+ {
2203
+ "epoch": 2.2448979591836733,
2204
+ "grad_norm": 6.65625,
2205
+ "learning_rate": 4.619885323231484e-05,
2206
+ "loss": 0.5581,
2207
+ "step": 2860
2208
+ },
2209
+ {
2210
+ "epoch": 2.2527472527472527,
2211
+ "grad_norm": 5.25,
2212
+ "learning_rate": 4.58508930128808e-05,
2213
+ "loss": 0.5357,
2214
+ "step": 2870
2215
+ },
2216
+ {
2217
+ "epoch": 2.260596546310832,
2218
+ "grad_norm": 5.625,
2219
+ "learning_rate": 4.550313501506781e-05,
2220
+ "loss": 0.5235,
2221
+ "step": 2880
2222
+ },
2223
+ {
2224
+ "epoch": 2.268445839874411,
2225
+ "grad_norm": 4.875,
2226
+ "learning_rate": 4.515559618811005e-05,
2227
+ "loss": 0.442,
2228
+ "step": 2890
2229
+ },
2230
+ {
2231
+ "epoch": 2.2762951334379906,
2232
+ "grad_norm": 5.375,
2233
+ "learning_rate": 4.4808293470559643e-05,
2234
+ "loss": 0.4889,
2235
+ "step": 2900
2236
+ },
2237
+ {
2238
+ "epoch": 2.2841444270015696,
2239
+ "grad_norm": 5.28125,
2240
+ "learning_rate": 4.446124378946107e-05,
2241
+ "loss": 0.6102,
2242
+ "step": 2910
2243
+ },
2244
+ {
2245
+ "epoch": 2.291993720565149,
2246
+ "grad_norm": 6.78125,
2247
+ "learning_rate": 4.411446405952618e-05,
2248
+ "loss": 0.5462,
2249
+ "step": 2920
2250
+ },
2251
+ {
2252
+ "epoch": 2.2998430141287285,
2253
+ "grad_norm": 5.21875,
2254
+ "learning_rate": 4.376797118230979e-05,
2255
+ "loss": 0.4926,
2256
+ "step": 2930
2257
+ },
2258
+ {
2259
+ "epoch": 2.3076923076923075,
2260
+ "grad_norm": 5.5,
2261
+ "learning_rate": 4.342178204538588e-05,
2262
+ "loss": 0.5301,
2263
+ "step": 2940
2264
+ },
2265
+ {
2266
+ "epoch": 2.315541601255887,
2267
+ "grad_norm": 5.84375,
2268
+ "learning_rate": 4.307591352152459e-05,
2269
+ "loss": 0.5053,
2270
+ "step": 2950
2271
+ },
2272
+ {
2273
+ "epoch": 2.3233908948194664,
2274
+ "grad_norm": 5.0,
2275
+ "learning_rate": 4.273038246786986e-05,
2276
+ "loss": 0.5717,
2277
+ "step": 2960
2278
+ },
2279
+ {
2280
+ "epoch": 2.3312401883830454,
2281
+ "grad_norm": 5.0625,
2282
+ "learning_rate": 4.238520572511773e-05,
2283
+ "loss": 0.5054,
2284
+ "step": 2970
2285
+ },
2286
+ {
2287
+ "epoch": 2.339089481946625,
2288
+ "grad_norm": 4.53125,
2289
+ "learning_rate": 4.2040400116695666e-05,
2290
+ "loss": 0.4606,
2291
+ "step": 2980
2292
+ },
2293
+ {
2294
+ "epoch": 2.3469387755102042,
2295
+ "grad_norm": 4.0625,
2296
+ "learning_rate": 4.169598244794261e-05,
2297
+ "loss": 0.4655,
2298
+ "step": 2990
2299
+ },
2300
+ {
2301
+ "epoch": 2.3547880690737832,
2302
+ "grad_norm": 5.6875,
2303
+ "learning_rate": 4.135196950528982e-05,
2304
+ "loss": 0.4932,
2305
+ "step": 3000
2306
+ }
2307
+ ],
2308
+ "logging_steps": 10,
2309
+ "max_steps": 5000,
2310
+ "num_input_tokens_seen": 0,
2311
+ "num_train_epochs": 4,
2312
+ "save_steps": 1000,
2313
+ "stateful_callbacks": {
2314
+ "TrainerControl": {
2315
+ "args": {
2316
+ "should_epoch_stop": false,
2317
+ "should_evaluate": false,
2318
+ "should_log": false,
2319
+ "should_save": true,
2320
+ "should_training_stop": false
2321
+ },
2322
+ "attributes": {}
2323
+ }
2324
+ },
2325
+ "total_flos": 0.0,
2326
+ "train_batch_size": 16,
2327
+ "trial_name": null,
2328
+ "trial_params": null
2329
+ }
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-3000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9365177436ad70bb9d6000d96d3ccdc5c5ca38536defdcfec6b548401edbc07f
3
+ size 6481
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "gradient_checkpointing": false,
20
+ "hidden_activation": "gelu",
21
+ "hidden_size": 768,
22
+ "initializer_cutoff_factor": 2.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 1152,
25
+ "layer_norm_eps": 1e-05,
26
+ "local_attention": 128,
27
+ "local_rope_theta": 10000.0,
28
+ "max_position_embeddings": 8192,
29
+ "mlp_bias": false,
30
+ "mlp_dropout": 0.0,
31
+ "model_type": "modernbert",
32
+ "norm_bias": false,
33
+ "norm_eps": 1e-05,
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 22,
36
+ "pad_token_id": 50283,
37
+ "position_embedding_type": "absolute",
38
+ "repad_logits_with_grad": false,
39
+ "sep_token_id": 50282,
40
+ "sparse_pred_ignore_index": -100,
41
+ "sparse_prediction": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.53.0",
44
+ "vocab_size": 50368
45
+ }
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1dd3dcdc59285c708a09271b560d3abb25a372173e901184f4c94aa67938defe
3
+ size 298041696
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:421aa92cd21b5f582224f33575aab4f4b0cb5e42700e3b3279f086383cb122a0
3
+ size 596170443
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24c12cceaffe8acec04189ba71dbb81f100dd5cd03b86390253117b3fab2911d
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69adbd9997461473344beb9c44d2e496e24fbc4d6fe69245ab0bd127882efd96
3
+ size 15429
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:baaf7026d9e588ca2ba5b4de8768a379982b7530f39e0b88fa44af9bef8e8bc9
3
+ size 15429
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ec81cbdf599c90b49fb13998f6e46e2492b55345216231b6ce078f88cf04eae
3
+ size 15429
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9218a8f161ba45a13a24f24cd020f32b0abec4bcc53a2764e48b480568dd2d13
3
+ size 1465
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/trainer_state.json ADDED
@@ -0,0 +1,3094 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 3.1397174254317113,
6
+ "eval_steps": 1000.0,
7
+ "global_step": 4000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0,
14
+ "eval/nano_beir.arguana": 0.4400549254974839,
15
+ "step": 0
16
+ },
17
+ {
18
+ "epoch": 0,
19
+ "eval/nano_beir.climate_fever": 0.22880053057901228,
20
+ "step": 0
21
+ },
22
+ {
23
+ "epoch": 0,
24
+ "eval/nano_beir.dbpedia_entity": 0.30026716538531234,
25
+ "step": 0
26
+ },
27
+ {
28
+ "epoch": 0,
29
+ "eval/nano_beir.fever": 0.6419178788751609,
30
+ "step": 0
31
+ },
32
+ {
33
+ "epoch": 0,
34
+ "eval/nano_beir.fiqa": 0.4370593595651809,
35
+ "step": 0
36
+ },
37
+ {
38
+ "epoch": 0,
39
+ "eval/nano_beir.hotpotqa": 0.6723047477224562,
40
+ "step": 0
41
+ },
42
+ {
43
+ "epoch": 0,
44
+ "eval/nano_beir.nfcorpus": 0.3437446614516137,
45
+ "step": 0
46
+ },
47
+ {
48
+ "epoch": 0,
49
+ "eval/nano_beir.nq": 0.436711543163819,
50
+ "step": 0
51
+ },
52
+ {
53
+ "epoch": 0,
54
+ "eval/nano_beir.quora": 0.8439997970269911,
55
+ "step": 0
56
+ },
57
+ {
58
+ "epoch": 0,
59
+ "eval/nano_beir.scidocs": 0.26304331393922176,
60
+ "step": 0
61
+ },
62
+ {
63
+ "epoch": 0,
64
+ "eval/nano_beir.scifact": 0.7599341408912704,
65
+ "step": 0
66
+ },
67
+ {
68
+ "epoch": 0,
69
+ "eval/nano_beir.webis_touche2020": 0.29696490432942557,
70
+ "step": 0
71
+ },
72
+ {
73
+ "epoch": 0,
74
+ "eval/avg": 0.47206691403557904,
75
+ "step": 0
76
+ },
77
+ {
78
+ "epoch": 0.007849293563579277,
79
+ "grad_norm": 8.25,
80
+ "learning_rate": 1.8e-06,
81
+ "loss": 0.7314,
82
+ "step": 10
83
+ },
84
+ {
85
+ "epoch": 0.015698587127158554,
86
+ "grad_norm": 7.03125,
87
+ "learning_rate": 3.8e-06,
88
+ "loss": 0.8796,
89
+ "step": 20
90
+ },
91
+ {
92
+ "epoch": 0.023547880690737835,
93
+ "grad_norm": 7.3125,
94
+ "learning_rate": 5.8e-06,
95
+ "loss": 0.7311,
96
+ "step": 30
97
+ },
98
+ {
99
+ "epoch": 0.03139717425431711,
100
+ "grad_norm": 7.15625,
101
+ "learning_rate": 7.8e-06,
102
+ "loss": 0.8214,
103
+ "step": 40
104
+ },
105
+ {
106
+ "epoch": 0.03924646781789639,
107
+ "grad_norm": 9.1875,
108
+ "learning_rate": 9.800000000000001e-06,
109
+ "loss": 0.8489,
110
+ "step": 50
111
+ },
112
+ {
113
+ "epoch": 0.04709576138147567,
114
+ "grad_norm": 6.84375,
115
+ "learning_rate": 1.18e-05,
116
+ "loss": 0.7661,
117
+ "step": 60
118
+ },
119
+ {
120
+ "epoch": 0.054945054945054944,
121
+ "grad_norm": 6.6875,
122
+ "learning_rate": 1.3800000000000002e-05,
123
+ "loss": 0.7525,
124
+ "step": 70
125
+ },
126
+ {
127
+ "epoch": 0.06279434850863422,
128
+ "grad_norm": 7.5625,
129
+ "learning_rate": 1.58e-05,
130
+ "loss": 0.8653,
131
+ "step": 80
132
+ },
133
+ {
134
+ "epoch": 0.0706436420722135,
135
+ "grad_norm": 6.25,
136
+ "learning_rate": 1.78e-05,
137
+ "loss": 0.7607,
138
+ "step": 90
139
+ },
140
+ {
141
+ "epoch": 0.07849293563579278,
142
+ "grad_norm": 5.28125,
143
+ "learning_rate": 1.9800000000000004e-05,
144
+ "loss": 0.7259,
145
+ "step": 100
146
+ },
147
+ {
148
+ "epoch": 0.08634222919937205,
149
+ "grad_norm": 5.59375,
150
+ "learning_rate": 2.18e-05,
151
+ "loss": 0.7862,
152
+ "step": 110
153
+ },
154
+ {
155
+ "epoch": 0.09419152276295134,
156
+ "grad_norm": 5.8125,
157
+ "learning_rate": 2.38e-05,
158
+ "loss": 0.8212,
159
+ "step": 120
160
+ },
161
+ {
162
+ "epoch": 0.10204081632653061,
163
+ "grad_norm": 6.0,
164
+ "learning_rate": 2.58e-05,
165
+ "loss": 0.6884,
166
+ "step": 130
167
+ },
168
+ {
169
+ "epoch": 0.10989010989010989,
170
+ "grad_norm": 5.9375,
171
+ "learning_rate": 2.7800000000000005e-05,
172
+ "loss": 0.712,
173
+ "step": 140
174
+ },
175
+ {
176
+ "epoch": 0.11773940345368916,
177
+ "grad_norm": 5.53125,
178
+ "learning_rate": 2.98e-05,
179
+ "loss": 0.6231,
180
+ "step": 150
181
+ },
182
+ {
183
+ "epoch": 0.12558869701726844,
184
+ "grad_norm": 5.59375,
185
+ "learning_rate": 3.18e-05,
186
+ "loss": 0.6733,
187
+ "step": 160
188
+ },
189
+ {
190
+ "epoch": 0.13343799058084774,
191
+ "grad_norm": 6.84375,
192
+ "learning_rate": 3.38e-05,
193
+ "loss": 0.6973,
194
+ "step": 170
195
+ },
196
+ {
197
+ "epoch": 0.141287284144427,
198
+ "grad_norm": 5.4375,
199
+ "learning_rate": 3.58e-05,
200
+ "loss": 0.5928,
201
+ "step": 180
202
+ },
203
+ {
204
+ "epoch": 0.14913657770800628,
205
+ "grad_norm": 6.375,
206
+ "learning_rate": 3.7800000000000004e-05,
207
+ "loss": 0.6189,
208
+ "step": 190
209
+ },
210
+ {
211
+ "epoch": 0.15698587127158556,
212
+ "grad_norm": 5.28125,
213
+ "learning_rate": 3.9800000000000005e-05,
214
+ "loss": 0.6661,
215
+ "step": 200
216
+ },
217
+ {
218
+ "epoch": 0.16483516483516483,
219
+ "grad_norm": 4.875,
220
+ "learning_rate": 4.18e-05,
221
+ "loss": 0.634,
222
+ "step": 210
223
+ },
224
+ {
225
+ "epoch": 0.1726844583987441,
226
+ "grad_norm": 6.78125,
227
+ "learning_rate": 4.38e-05,
228
+ "loss": 0.6963,
229
+ "step": 220
230
+ },
231
+ {
232
+ "epoch": 0.18053375196232338,
233
+ "grad_norm": 5.875,
234
+ "learning_rate": 4.58e-05,
235
+ "loss": 0.6744,
236
+ "step": 230
237
+ },
238
+ {
239
+ "epoch": 0.18838304552590268,
240
+ "grad_norm": 5.90625,
241
+ "learning_rate": 4.78e-05,
242
+ "loss": 0.6635,
243
+ "step": 240
244
+ },
245
+ {
246
+ "epoch": 0.19623233908948196,
247
+ "grad_norm": 4.9375,
248
+ "learning_rate": 4.9800000000000004e-05,
249
+ "loss": 0.6874,
250
+ "step": 250
251
+ },
252
+ {
253
+ "epoch": 0.20408163265306123,
254
+ "grad_norm": 5.28125,
255
+ "learning_rate": 5.1800000000000005e-05,
256
+ "loss": 0.6516,
257
+ "step": 260
258
+ },
259
+ {
260
+ "epoch": 0.2119309262166405,
261
+ "grad_norm": 5.40625,
262
+ "learning_rate": 5.380000000000001e-05,
263
+ "loss": 0.7579,
264
+ "step": 270
265
+ },
266
+ {
267
+ "epoch": 0.21978021978021978,
268
+ "grad_norm": 4.15625,
269
+ "learning_rate": 5.580000000000001e-05,
270
+ "loss": 0.6867,
271
+ "step": 280
272
+ },
273
+ {
274
+ "epoch": 0.22762951334379905,
275
+ "grad_norm": 6.59375,
276
+ "learning_rate": 5.7799999999999995e-05,
277
+ "loss": 0.7038,
278
+ "step": 290
279
+ },
280
+ {
281
+ "epoch": 0.23547880690737832,
282
+ "grad_norm": 5.9375,
283
+ "learning_rate": 5.9800000000000003e-05,
284
+ "loss": 0.6681,
285
+ "step": 300
286
+ },
287
+ {
288
+ "epoch": 0.24332810047095763,
289
+ "grad_norm": 4.4375,
290
+ "learning_rate": 6.18e-05,
291
+ "loss": 0.6133,
292
+ "step": 310
293
+ },
294
+ {
295
+ "epoch": 0.25117739403453687,
296
+ "grad_norm": 4.4375,
297
+ "learning_rate": 6.38e-05,
298
+ "loss": 0.6203,
299
+ "step": 320
300
+ },
301
+ {
302
+ "epoch": 0.25902668759811615,
303
+ "grad_norm": 6.3125,
304
+ "learning_rate": 6.58e-05,
305
+ "loss": 0.7164,
306
+ "step": 330
307
+ },
308
+ {
309
+ "epoch": 0.2668759811616955,
310
+ "grad_norm": 4.84375,
311
+ "learning_rate": 6.780000000000001e-05,
312
+ "loss": 0.7157,
313
+ "step": 340
314
+ },
315
+ {
316
+ "epoch": 0.27472527472527475,
317
+ "grad_norm": 5.09375,
318
+ "learning_rate": 6.98e-05,
319
+ "loss": 0.7176,
320
+ "step": 350
321
+ },
322
+ {
323
+ "epoch": 0.282574568288854,
324
+ "grad_norm": 5.1875,
325
+ "learning_rate": 7.18e-05,
326
+ "loss": 0.702,
327
+ "step": 360
328
+ },
329
+ {
330
+ "epoch": 0.2904238618524333,
331
+ "grad_norm": 5.25,
332
+ "learning_rate": 7.38e-05,
333
+ "loss": 0.5834,
334
+ "step": 370
335
+ },
336
+ {
337
+ "epoch": 0.29827315541601257,
338
+ "grad_norm": 5.46875,
339
+ "learning_rate": 7.58e-05,
340
+ "loss": 0.594,
341
+ "step": 380
342
+ },
343
+ {
344
+ "epoch": 0.30612244897959184,
345
+ "grad_norm": 6.625,
346
+ "learning_rate": 7.780000000000001e-05,
347
+ "loss": 0.7667,
348
+ "step": 390
349
+ },
350
+ {
351
+ "epoch": 0.3139717425431711,
352
+ "grad_norm": 6.8125,
353
+ "learning_rate": 7.98e-05,
354
+ "loss": 0.6477,
355
+ "step": 400
356
+ },
357
+ {
358
+ "epoch": 0.3218210361067504,
359
+ "grad_norm": 5.21875,
360
+ "learning_rate": 8.18e-05,
361
+ "loss": 0.6104,
362
+ "step": 410
363
+ },
364
+ {
365
+ "epoch": 0.32967032967032966,
366
+ "grad_norm": 6.03125,
367
+ "learning_rate": 8.38e-05,
368
+ "loss": 0.6879,
369
+ "step": 420
370
+ },
371
+ {
372
+ "epoch": 0.33751962323390894,
373
+ "grad_norm": 5.90625,
374
+ "learning_rate": 8.58e-05,
375
+ "loss": 0.6514,
376
+ "step": 430
377
+ },
378
+ {
379
+ "epoch": 0.3453689167974882,
380
+ "grad_norm": 6.03125,
381
+ "learning_rate": 8.78e-05,
382
+ "loss": 0.649,
383
+ "step": 440
384
+ },
385
+ {
386
+ "epoch": 0.3532182103610675,
387
+ "grad_norm": 4.75,
388
+ "learning_rate": 8.98e-05,
389
+ "loss": 0.6204,
390
+ "step": 450
391
+ },
392
+ {
393
+ "epoch": 0.36106750392464676,
394
+ "grad_norm": 5.125,
395
+ "learning_rate": 9.180000000000001e-05,
396
+ "loss": 0.6633,
397
+ "step": 460
398
+ },
399
+ {
400
+ "epoch": 0.36891679748822603,
401
+ "grad_norm": 4.59375,
402
+ "learning_rate": 9.38e-05,
403
+ "loss": 0.7058,
404
+ "step": 470
405
+ },
406
+ {
407
+ "epoch": 0.37676609105180536,
408
+ "grad_norm": 5.9375,
409
+ "learning_rate": 9.58e-05,
410
+ "loss": 0.5653,
411
+ "step": 480
412
+ },
413
+ {
414
+ "epoch": 0.38461538461538464,
415
+ "grad_norm": 5.6875,
416
+ "learning_rate": 9.78e-05,
417
+ "loss": 0.6143,
418
+ "step": 490
419
+ },
420
+ {
421
+ "epoch": 0.3924646781789639,
422
+ "grad_norm": 5.28125,
423
+ "learning_rate": 9.98e-05,
424
+ "loss": 0.6218,
425
+ "step": 500
426
+ },
427
+ {
428
+ "epoch": 0.4003139717425432,
429
+ "grad_norm": 6.03125,
430
+ "learning_rate": 9.999901304280685e-05,
431
+ "loss": 0.6115,
432
+ "step": 510
433
+ },
434
+ {
435
+ "epoch": 0.40816326530612246,
436
+ "grad_norm": 4.53125,
437
+ "learning_rate": 9.999560138895238e-05,
438
+ "loss": 0.6206,
439
+ "step": 520
440
+ },
441
+ {
442
+ "epoch": 0.41601255886970173,
443
+ "grad_norm": 5.15625,
444
+ "learning_rate": 9.99897530200195e-05,
445
+ "loss": 0.629,
446
+ "step": 530
447
+ },
448
+ {
449
+ "epoch": 0.423861852433281,
450
+ "grad_norm": 4.6875,
451
+ "learning_rate": 9.998146822104943e-05,
452
+ "loss": 0.6492,
453
+ "step": 540
454
+ },
455
+ {
456
+ "epoch": 0.4317111459968603,
457
+ "grad_norm": 5.1875,
458
+ "learning_rate": 9.997074739583162e-05,
459
+ "loss": 0.585,
460
+ "step": 550
461
+ },
462
+ {
463
+ "epoch": 0.43956043956043955,
464
+ "grad_norm": 7.1875,
465
+ "learning_rate": 9.995759106688393e-05,
466
+ "loss": 0.558,
467
+ "step": 560
468
+ },
469
+ {
470
+ "epoch": 0.4474097331240188,
471
+ "grad_norm": 5.28125,
472
+ "learning_rate": 9.99419998754273e-05,
473
+ "loss": 0.6242,
474
+ "step": 570
475
+ },
476
+ {
477
+ "epoch": 0.4552590266875981,
478
+ "grad_norm": 4.53125,
479
+ "learning_rate": 9.992397458135438e-05,
480
+ "loss": 0.682,
481
+ "step": 580
482
+ },
483
+ {
484
+ "epoch": 0.4631083202511774,
485
+ "grad_norm": 4.3125,
486
+ "learning_rate": 9.990351606319261e-05,
487
+ "loss": 0.6194,
488
+ "step": 590
489
+ },
490
+ {
491
+ "epoch": 0.47095761381475665,
492
+ "grad_norm": 5.90625,
493
+ "learning_rate": 9.988062531806126e-05,
494
+ "loss": 0.6211,
495
+ "step": 600
496
+ },
497
+ {
498
+ "epoch": 0.478806907378336,
499
+ "grad_norm": 4.5625,
500
+ "learning_rate": 9.985530346162299e-05,
501
+ "loss": 0.617,
502
+ "step": 610
503
+ },
504
+ {
505
+ "epoch": 0.48665620094191525,
506
+ "grad_norm": 5.125,
507
+ "learning_rate": 9.982755172802933e-05,
508
+ "loss": 0.5901,
509
+ "step": 620
510
+ },
511
+ {
512
+ "epoch": 0.4945054945054945,
513
+ "grad_norm": 4.15625,
514
+ "learning_rate": 9.979737146986064e-05,
515
+ "loss": 0.6067,
516
+ "step": 630
517
+ },
518
+ {
519
+ "epoch": 0.5023547880690737,
520
+ "grad_norm": 5.4375,
521
+ "learning_rate": 9.976476415806013e-05,
522
+ "loss": 0.5976,
523
+ "step": 640
524
+ },
525
+ {
526
+ "epoch": 0.5102040816326531,
527
+ "grad_norm": 5.75,
528
+ "learning_rate": 9.972973138186217e-05,
529
+ "loss": 0.6467,
530
+ "step": 650
531
+ },
532
+ {
533
+ "epoch": 0.5180533751962323,
534
+ "grad_norm": 7.25,
535
+ "learning_rate": 9.969227484871484e-05,
536
+ "loss": 0.5904,
537
+ "step": 660
538
+ },
539
+ {
540
+ "epoch": 0.5259026687598116,
541
+ "grad_norm": 4.65625,
542
+ "learning_rate": 9.965239638419672e-05,
543
+ "loss": 0.6814,
544
+ "step": 670
545
+ },
546
+ {
547
+ "epoch": 0.533751962323391,
548
+ "grad_norm": 6.125,
549
+ "learning_rate": 9.961009793192792e-05,
550
+ "loss": 0.6684,
551
+ "step": 680
552
+ },
553
+ {
554
+ "epoch": 0.5416012558869702,
555
+ "grad_norm": 4.40625,
556
+ "learning_rate": 9.956538155347534e-05,
557
+ "loss": 0.6206,
558
+ "step": 690
559
+ },
560
+ {
561
+ "epoch": 0.5494505494505495,
562
+ "grad_norm": 5.875,
563
+ "learning_rate": 9.951824942825215e-05,
564
+ "loss": 0.6096,
565
+ "step": 700
566
+ },
567
+ {
568
+ "epoch": 0.5572998430141287,
569
+ "grad_norm": 5.71875,
570
+ "learning_rate": 9.946870385341167e-05,
571
+ "loss": 0.5468,
572
+ "step": 710
573
+ },
574
+ {
575
+ "epoch": 0.565149136577708,
576
+ "grad_norm": 4.84375,
577
+ "learning_rate": 9.941674724373531e-05,
578
+ "loss": 0.5883,
579
+ "step": 720
580
+ },
581
+ {
582
+ "epoch": 0.5729984301412873,
583
+ "grad_norm": 4.75,
584
+ "learning_rate": 9.936238213151491e-05,
585
+ "loss": 0.596,
586
+ "step": 730
587
+ },
588
+ {
589
+ "epoch": 0.5808477237048666,
590
+ "grad_norm": 4.9375,
591
+ "learning_rate": 9.930561116642935e-05,
592
+ "loss": 0.5178,
593
+ "step": 740
594
+ },
595
+ {
596
+ "epoch": 0.5886970172684458,
597
+ "grad_norm": 5.59375,
598
+ "learning_rate": 9.924643711541539e-05,
599
+ "loss": 0.6438,
600
+ "step": 750
601
+ },
602
+ {
603
+ "epoch": 0.5965463108320251,
604
+ "grad_norm": 4.96875,
605
+ "learning_rate": 9.918486286253279e-05,
606
+ "loss": 0.6975,
607
+ "step": 760
608
+ },
609
+ {
610
+ "epoch": 0.6043956043956044,
611
+ "grad_norm": 4.375,
612
+ "learning_rate": 9.912089140882376e-05,
613
+ "loss": 0.5828,
614
+ "step": 770
615
+ },
616
+ {
617
+ "epoch": 0.6122448979591837,
618
+ "grad_norm": 7.0625,
619
+ "learning_rate": 9.90545258721667e-05,
620
+ "loss": 0.5019,
621
+ "step": 780
622
+ },
623
+ {
624
+ "epoch": 0.6200941915227629,
625
+ "grad_norm": 6.75,
626
+ "learning_rate": 9.898576948712426e-05,
627
+ "loss": 0.6489,
628
+ "step": 790
629
+ },
630
+ {
631
+ "epoch": 0.6279434850863422,
632
+ "grad_norm": 5.28125,
633
+ "learning_rate": 9.891462560478562e-05,
634
+ "loss": 0.6364,
635
+ "step": 800
636
+ },
637
+ {
638
+ "epoch": 0.6357927786499215,
639
+ "grad_norm": 4.6875,
640
+ "learning_rate": 9.884109769260325e-05,
641
+ "loss": 0.5458,
642
+ "step": 810
643
+ },
644
+ {
645
+ "epoch": 0.6436420722135008,
646
+ "grad_norm": 3.703125,
647
+ "learning_rate": 9.876518933422386e-05,
648
+ "loss": 0.5859,
649
+ "step": 820
650
+ },
651
+ {
652
+ "epoch": 0.6514913657770801,
653
+ "grad_norm": 4.375,
654
+ "learning_rate": 9.868690422931372e-05,
655
+ "loss": 0.6369,
656
+ "step": 830
657
+ },
658
+ {
659
+ "epoch": 0.6593406593406593,
660
+ "grad_norm": 5.40625,
661
+ "learning_rate": 9.860624619337844e-05,
662
+ "loss": 0.7126,
663
+ "step": 840
664
+ },
665
+ {
666
+ "epoch": 0.6671899529042387,
667
+ "grad_norm": 4.9375,
668
+ "learning_rate": 9.852321915757687e-05,
669
+ "loss": 0.5857,
670
+ "step": 850
671
+ },
672
+ {
673
+ "epoch": 0.6750392464678179,
674
+ "grad_norm": 5.0,
675
+ "learning_rate": 9.843782716852963e-05,
676
+ "loss": 0.645,
677
+ "step": 860
678
+ },
679
+ {
680
+ "epoch": 0.6828885400313972,
681
+ "grad_norm": 3.78125,
682
+ "learning_rate": 9.835007438812177e-05,
683
+ "loss": 0.6538,
684
+ "step": 870
685
+ },
686
+ {
687
+ "epoch": 0.6907378335949764,
688
+ "grad_norm": 5.15625,
689
+ "learning_rate": 9.82599650933e-05,
690
+ "loss": 0.6045,
691
+ "step": 880
692
+ },
693
+ {
694
+ "epoch": 0.6985871271585558,
695
+ "grad_norm": 3.59375,
696
+ "learning_rate": 9.816750367586425e-05,
697
+ "loss": 0.6282,
698
+ "step": 890
699
+ },
700
+ {
701
+ "epoch": 0.706436420722135,
702
+ "grad_norm": 5.1875,
703
+ "learning_rate": 9.807269464225355e-05,
704
+ "loss": 0.5509,
705
+ "step": 900
706
+ },
707
+ {
708
+ "epoch": 0.7142857142857143,
709
+ "grad_norm": 7.125,
710
+ "learning_rate": 9.797554261332645e-05,
711
+ "loss": 0.6392,
712
+ "step": 910
713
+ },
714
+ {
715
+ "epoch": 0.7221350078492935,
716
+ "grad_norm": 5.34375,
717
+ "learning_rate": 9.787605232413574e-05,
718
+ "loss": 0.6152,
719
+ "step": 920
720
+ },
721
+ {
722
+ "epoch": 0.7299843014128728,
723
+ "grad_norm": 4.28125,
724
+ "learning_rate": 9.777422862369783e-05,
725
+ "loss": 0.7706,
726
+ "step": 930
727
+ },
728
+ {
729
+ "epoch": 0.7378335949764521,
730
+ "grad_norm": 4.59375,
731
+ "learning_rate": 9.767007647475617e-05,
732
+ "loss": 0.6656,
733
+ "step": 940
734
+ },
735
+ {
736
+ "epoch": 0.7456828885400314,
737
+ "grad_norm": 4.53125,
738
+ "learning_rate": 9.756360095353956e-05,
739
+ "loss": 0.6592,
740
+ "step": 950
741
+ },
742
+ {
743
+ "epoch": 0.7535321821036107,
744
+ "grad_norm": 4.40625,
745
+ "learning_rate": 9.745480724951473e-05,
746
+ "loss": 0.6141,
747
+ "step": 960
748
+ },
749
+ {
750
+ "epoch": 0.7613814756671899,
751
+ "grad_norm": 4.96875,
752
+ "learning_rate": 9.73437006651333e-05,
753
+ "loss": 0.6161,
754
+ "step": 970
755
+ },
756
+ {
757
+ "epoch": 0.7692307692307693,
758
+ "grad_norm": 4.25,
759
+ "learning_rate": 9.723028661557345e-05,
760
+ "loss": 0.7166,
761
+ "step": 980
762
+ },
763
+ {
764
+ "epoch": 0.7770800627943485,
765
+ "grad_norm": 5.34375,
766
+ "learning_rate": 9.711457062847595e-05,
767
+ "loss": 0.6028,
768
+ "step": 990
769
+ },
770
+ {
771
+ "epoch": 0.7849293563579278,
772
+ "grad_norm": 4.34375,
773
+ "learning_rate": 9.699655834367478e-05,
774
+ "loss": 0.5262,
775
+ "step": 1000
776
+ },
777
+ {
778
+ "epoch": 0.7849293563579278,
779
+ "eval/nano_beir.arguana": 0.4712918167441636,
780
+ "step": 1000
781
+ },
782
+ {
783
+ "epoch": 0.7849293563579278,
784
+ "eval/nano_beir.climate_fever": 0.23854189010584295,
785
+ "step": 1000
786
+ },
787
+ {
788
+ "epoch": 0.7849293563579278,
789
+ "eval/nano_beir.dbpedia_entity": 0.33317100014903445,
790
+ "step": 1000
791
+ },
792
+ {
793
+ "epoch": 0.7849293563579278,
794
+ "eval/nano_beir.fever": 0.6283057972805869,
795
+ "step": 1000
796
+ },
797
+ {
798
+ "epoch": 0.7849293563579278,
799
+ "eval/nano_beir.fiqa": 0.4448076388918781,
800
+ "step": 1000
801
+ },
802
+ {
803
+ "epoch": 0.7849293563579278,
804
+ "eval/nano_beir.hotpotqa": 0.6717658203813929,
805
+ "step": 1000
806
+ },
807
+ {
808
+ "epoch": 0.7849293563579278,
809
+ "eval/nano_beir.nfcorpus": 0.35468777200298285,
810
+ "step": 1000
811
+ },
812
+ {
813
+ "epoch": 0.7849293563579278,
814
+ "eval/nano_beir.nq": 0.41666515563087825,
815
+ "step": 1000
816
+ },
817
+ {
818
+ "epoch": 0.7849293563579278,
819
+ "eval/nano_beir.quora": 0.8413355846974869,
820
+ "step": 1000
821
+ },
822
+ {
823
+ "epoch": 0.7849293563579278,
824
+ "eval/nano_beir.scidocs": 0.25553825348355963,
825
+ "step": 1000
826
+ },
827
+ {
828
+ "epoch": 0.7849293563579278,
829
+ "eval/nano_beir.scifact": 0.7484102747744907,
830
+ "step": 1000
831
+ },
832
+ {
833
+ "epoch": 0.7849293563579278,
834
+ "eval/nano_beir.webis_touche2020": 0.2895097071551271,
835
+ "step": 1000
836
+ },
837
+ {
838
+ "epoch": 0.7849293563579278,
839
+ "eval/avg": 0.47450255927478535,
840
+ "step": 1000
841
+ },
842
+ {
843
+ "epoch": 0.792778649921507,
844
+ "grad_norm": 4.90625,
845
+ "learning_rate": 9.68762555129222e-05,
846
+ "loss": 0.5077,
847
+ "step": 1010
848
+ },
849
+ {
850
+ "epoch": 0.8006279434850864,
851
+ "grad_norm": 5.0,
852
+ "learning_rate": 9.675366799960841e-05,
853
+ "loss": 0.7276,
854
+ "step": 1020
855
+ },
856
+ {
857
+ "epoch": 0.8084772370486656,
858
+ "grad_norm": 6.59375,
859
+ "learning_rate": 9.662880177847594e-05,
860
+ "loss": 0.5999,
861
+ "step": 1030
862
+ },
863
+ {
864
+ "epoch": 0.8163265306122449,
865
+ "grad_norm": 3.875,
866
+ "learning_rate": 9.65016629353282e-05,
867
+ "loss": 0.6675,
868
+ "step": 1040
869
+ },
870
+ {
871
+ "epoch": 0.8241758241758241,
872
+ "grad_norm": 3.875,
873
+ "learning_rate": 9.637225766673307e-05,
874
+ "loss": 0.6682,
875
+ "step": 1050
876
+ },
877
+ {
878
+ "epoch": 0.8320251177394035,
879
+ "grad_norm": 5.125,
880
+ "learning_rate": 9.624059227972075e-05,
881
+ "loss": 0.6381,
882
+ "step": 1060
883
+ },
884
+ {
885
+ "epoch": 0.8398744113029827,
886
+ "grad_norm": 4.8125,
887
+ "learning_rate": 9.610667319147647e-05,
888
+ "loss": 0.607,
889
+ "step": 1070
890
+ },
891
+ {
892
+ "epoch": 0.847723704866562,
893
+ "grad_norm": 4.75,
894
+ "learning_rate": 9.597050692902765e-05,
895
+ "loss": 0.5786,
896
+ "step": 1080
897
+ },
898
+ {
899
+ "epoch": 0.8555729984301413,
900
+ "grad_norm": 6.125,
901
+ "learning_rate": 9.583210012892582e-05,
902
+ "loss": 0.5928,
903
+ "step": 1090
904
+ },
905
+ {
906
+ "epoch": 0.8634222919937206,
907
+ "grad_norm": 3.734375,
908
+ "learning_rate": 9.569145953692315e-05,
909
+ "loss": 0.595,
910
+ "step": 1100
911
+ },
912
+ {
913
+ "epoch": 0.8712715855572999,
914
+ "grad_norm": 4.8125,
915
+ "learning_rate": 9.55485920076437e-05,
916
+ "loss": 0.604,
917
+ "step": 1110
918
+ },
919
+ {
920
+ "epoch": 0.8791208791208791,
921
+ "grad_norm": 5.0625,
922
+ "learning_rate": 9.540350450424926e-05,
923
+ "loss": 0.6252,
924
+ "step": 1120
925
+ },
926
+ {
927
+ "epoch": 0.8869701726844584,
928
+ "grad_norm": 5.46875,
929
+ "learning_rate": 9.525620409810007e-05,
930
+ "loss": 0.5765,
931
+ "step": 1130
932
+ },
933
+ {
934
+ "epoch": 0.8948194662480377,
935
+ "grad_norm": 4.40625,
936
+ "learning_rate": 9.510669796841014e-05,
937
+ "loss": 0.5604,
938
+ "step": 1140
939
+ },
940
+ {
941
+ "epoch": 0.902668759811617,
942
+ "grad_norm": 6.34375,
943
+ "learning_rate": 9.495499340189728e-05,
944
+ "loss": 0.555,
945
+ "step": 1150
946
+ },
947
+ {
948
+ "epoch": 0.9105180533751962,
949
+ "grad_norm": 4.71875,
950
+ "learning_rate": 9.480109779242805e-05,
951
+ "loss": 0.6496,
952
+ "step": 1160
953
+ },
954
+ {
955
+ "epoch": 0.9183673469387755,
956
+ "grad_norm": 4.71875,
957
+ "learning_rate": 9.464501864065735e-05,
958
+ "loss": 0.584,
959
+ "step": 1170
960
+ },
961
+ {
962
+ "epoch": 0.9262166405023547,
963
+ "grad_norm": 4.1875,
964
+ "learning_rate": 9.448676355366282e-05,
965
+ "loss": 0.6644,
966
+ "step": 1180
967
+ },
968
+ {
969
+ "epoch": 0.9340659340659341,
970
+ "grad_norm": 3.890625,
971
+ "learning_rate": 9.432634024457414e-05,
972
+ "loss": 0.6077,
973
+ "step": 1190
974
+ },
975
+ {
976
+ "epoch": 0.9419152276295133,
977
+ "grad_norm": 6.1875,
978
+ "learning_rate": 9.416375653219709e-05,
979
+ "loss": 0.5531,
980
+ "step": 1200
981
+ },
982
+ {
983
+ "epoch": 0.9497645211930926,
984
+ "grad_norm": 3.921875,
985
+ "learning_rate": 9.399902034063242e-05,
986
+ "loss": 0.6593,
987
+ "step": 1210
988
+ },
989
+ {
990
+ "epoch": 0.957613814756672,
991
+ "grad_norm": 4.15625,
992
+ "learning_rate": 9.383213969888972e-05,
993
+ "loss": 0.5477,
994
+ "step": 1220
995
+ },
996
+ {
997
+ "epoch": 0.9654631083202512,
998
+ "grad_norm": 4.09375,
999
+ "learning_rate": 9.366312274049602e-05,
1000
+ "loss": 0.6071,
1001
+ "step": 1230
1002
+ },
1003
+ {
1004
+ "epoch": 0.9733124018838305,
1005
+ "grad_norm": 4.34375,
1006
+ "learning_rate": 9.34919777030994e-05,
1007
+ "loss": 0.5946,
1008
+ "step": 1240
1009
+ },
1010
+ {
1011
+ "epoch": 0.9811616954474097,
1012
+ "grad_norm": 4.8125,
1013
+ "learning_rate": 9.331871292806759e-05,
1014
+ "loss": 0.5837,
1015
+ "step": 1250
1016
+ },
1017
+ {
1018
+ "epoch": 0.989010989010989,
1019
+ "grad_norm": 5.1875,
1020
+ "learning_rate": 9.314333686008125e-05,
1021
+ "loss": 0.6056,
1022
+ "step": 1260
1023
+ },
1024
+ {
1025
+ "epoch": 0.9968602825745683,
1026
+ "grad_norm": 4.90625,
1027
+ "learning_rate": 9.296585804672252e-05,
1028
+ "loss": 0.5941,
1029
+ "step": 1270
1030
+ },
1031
+ {
1032
+ "epoch": 1.0047095761381475,
1033
+ "grad_norm": 5.03125,
1034
+ "learning_rate": 9.278628513805838e-05,
1035
+ "loss": 0.6723,
1036
+ "step": 1280
1037
+ },
1038
+ {
1039
+ "epoch": 1.012558869701727,
1040
+ "grad_norm": 3.875,
1041
+ "learning_rate": 9.260462688621905e-05,
1042
+ "loss": 0.6221,
1043
+ "step": 1290
1044
+ },
1045
+ {
1046
+ "epoch": 1.0204081632653061,
1047
+ "grad_norm": 4.3125,
1048
+ "learning_rate": 9.242089214497145e-05,
1049
+ "loss": 0.5627,
1050
+ "step": 1300
1051
+ },
1052
+ {
1053
+ "epoch": 1.0282574568288854,
1054
+ "grad_norm": 4.6875,
1055
+ "learning_rate": 9.223508986928766e-05,
1056
+ "loss": 0.6045,
1057
+ "step": 1310
1058
+ },
1059
+ {
1060
+ "epoch": 1.0361067503924646,
1061
+ "grad_norm": 3.515625,
1062
+ "learning_rate": 9.204722911490846e-05,
1063
+ "loss": 0.5723,
1064
+ "step": 1320
1065
+ },
1066
+ {
1067
+ "epoch": 1.043956043956044,
1068
+ "grad_norm": 5.625,
1069
+ "learning_rate": 9.1857319037902e-05,
1070
+ "loss": 0.5616,
1071
+ "step": 1330
1072
+ },
1073
+ {
1074
+ "epoch": 1.0518053375196232,
1075
+ "grad_norm": 5.5625,
1076
+ "learning_rate": 9.16653688942175e-05,
1077
+ "loss": 0.5994,
1078
+ "step": 1340
1079
+ },
1080
+ {
1081
+ "epoch": 1.0596546310832025,
1082
+ "grad_norm": 4.65625,
1083
+ "learning_rate": 9.147138803923416e-05,
1084
+ "loss": 0.6163,
1085
+ "step": 1350
1086
+ },
1087
+ {
1088
+ "epoch": 1.0675039246467817,
1089
+ "grad_norm": 8.0,
1090
+ "learning_rate": 9.127538592730519e-05,
1091
+ "loss": 0.6001,
1092
+ "step": 1360
1093
+ },
1094
+ {
1095
+ "epoch": 1.0753532182103611,
1096
+ "grad_norm": 4.4375,
1097
+ "learning_rate": 9.107737211129701e-05,
1098
+ "loss": 0.55,
1099
+ "step": 1370
1100
+ },
1101
+ {
1102
+ "epoch": 1.0832025117739403,
1103
+ "grad_norm": 5.21875,
1104
+ "learning_rate": 9.087735624212365e-05,
1105
+ "loss": 0.5109,
1106
+ "step": 1380
1107
+ },
1108
+ {
1109
+ "epoch": 1.0910518053375196,
1110
+ "grad_norm": 4.59375,
1111
+ "learning_rate": 9.06753480682764e-05,
1112
+ "loss": 0.6688,
1113
+ "step": 1390
1114
+ },
1115
+ {
1116
+ "epoch": 1.098901098901099,
1117
+ "grad_norm": 5.3125,
1118
+ "learning_rate": 9.047135743534868e-05,
1119
+ "loss": 0.6504,
1120
+ "step": 1400
1121
+ },
1122
+ {
1123
+ "epoch": 1.1067503924646782,
1124
+ "grad_norm": 4.625,
1125
+ "learning_rate": 9.02653942855561e-05,
1126
+ "loss": 0.6315,
1127
+ "step": 1410
1128
+ },
1129
+ {
1130
+ "epoch": 1.1145996860282574,
1131
+ "grad_norm": 6.25,
1132
+ "learning_rate": 9.005746865725207e-05,
1133
+ "loss": 0.6297,
1134
+ "step": 1420
1135
+ },
1136
+ {
1137
+ "epoch": 1.1224489795918366,
1138
+ "grad_norm": 5.5625,
1139
+ "learning_rate": 8.984759068443831e-05,
1140
+ "loss": 0.5962,
1141
+ "step": 1430
1142
+ },
1143
+ {
1144
+ "epoch": 1.130298273155416,
1145
+ "grad_norm": 5.0625,
1146
+ "learning_rate": 8.963577059627118e-05,
1147
+ "loss": 0.5681,
1148
+ "step": 1440
1149
+ },
1150
+ {
1151
+ "epoch": 1.1381475667189953,
1152
+ "grad_norm": 4.78125,
1153
+ "learning_rate": 8.942201871656291e-05,
1154
+ "loss": 0.6486,
1155
+ "step": 1450
1156
+ },
1157
+ {
1158
+ "epoch": 1.1459968602825745,
1159
+ "grad_norm": 5.40625,
1160
+ "learning_rate": 8.920634546327858e-05,
1161
+ "loss": 0.5462,
1162
+ "step": 1460
1163
+ },
1164
+ {
1165
+ "epoch": 1.1538461538461537,
1166
+ "grad_norm": 5.6875,
1167
+ "learning_rate": 8.898876134802826e-05,
1168
+ "loss": 0.5653,
1169
+ "step": 1470
1170
+ },
1171
+ {
1172
+ "epoch": 1.1616954474097332,
1173
+ "grad_norm": 4.5625,
1174
+ "learning_rate": 8.87692769755548e-05,
1175
+ "loss": 0.492,
1176
+ "step": 1480
1177
+ },
1178
+ {
1179
+ "epoch": 1.1695447409733124,
1180
+ "grad_norm": 4.34375,
1181
+ "learning_rate": 8.854790304321681e-05,
1182
+ "loss": 0.5613,
1183
+ "step": 1490
1184
+ },
1185
+ {
1186
+ "epoch": 1.1773940345368916,
1187
+ "grad_norm": 5.15625,
1188
+ "learning_rate": 8.832465034046749e-05,
1189
+ "loss": 0.5728,
1190
+ "step": 1500
1191
+ },
1192
+ {
1193
+ "epoch": 1.185243328100471,
1194
+ "grad_norm": 4.34375,
1195
+ "learning_rate": 8.80995297483286e-05,
1196
+ "loss": 0.5405,
1197
+ "step": 1510
1198
+ },
1199
+ {
1200
+ "epoch": 1.1930926216640503,
1201
+ "grad_norm": 4.53125,
1202
+ "learning_rate": 8.787255223886019e-05,
1203
+ "loss": 0.5618,
1204
+ "step": 1520
1205
+ },
1206
+ {
1207
+ "epoch": 1.2009419152276295,
1208
+ "grad_norm": 4.875,
1209
+ "learning_rate": 8.764372887462586e-05,
1210
+ "loss": 0.5731,
1211
+ "step": 1530
1212
+ },
1213
+ {
1214
+ "epoch": 1.2087912087912087,
1215
+ "grad_norm": 4.71875,
1216
+ "learning_rate": 8.741307080815357e-05,
1217
+ "loss": 0.5745,
1218
+ "step": 1540
1219
+ },
1220
+ {
1221
+ "epoch": 1.2166405023547882,
1222
+ "grad_norm": 5.15625,
1223
+ "learning_rate": 8.718058928139205e-05,
1224
+ "loss": 0.5266,
1225
+ "step": 1550
1226
+ },
1227
+ {
1228
+ "epoch": 1.2244897959183674,
1229
+ "grad_norm": 4.53125,
1230
+ "learning_rate": 8.694629562516294e-05,
1231
+ "loss": 0.6069,
1232
+ "step": 1560
1233
+ },
1234
+ {
1235
+ "epoch": 1.2323390894819466,
1236
+ "grad_norm": 3.5,
1237
+ "learning_rate": 8.671020125860851e-05,
1238
+ "loss": 0.5433,
1239
+ "step": 1570
1240
+ },
1241
+ {
1242
+ "epoch": 1.2401883830455258,
1243
+ "grad_norm": 4.0,
1244
+ "learning_rate": 8.647231768863512e-05,
1245
+ "loss": 0.4842,
1246
+ "step": 1580
1247
+ },
1248
+ {
1249
+ "epoch": 1.2480376766091053,
1250
+ "grad_norm": 4.6875,
1251
+ "learning_rate": 8.623265650935234e-05,
1252
+ "loss": 0.58,
1253
+ "step": 1590
1254
+ },
1255
+ {
1256
+ "epoch": 1.2558869701726845,
1257
+ "grad_norm": 4.125,
1258
+ "learning_rate": 8.599122940150795e-05,
1259
+ "loss": 0.5331,
1260
+ "step": 1600
1261
+ },
1262
+ {
1263
+ "epoch": 1.2637362637362637,
1264
+ "grad_norm": 6.0,
1265
+ "learning_rate": 8.574804813191858e-05,
1266
+ "loss": 0.6408,
1267
+ "step": 1610
1268
+ },
1269
+ {
1270
+ "epoch": 1.2715855572998431,
1271
+ "grad_norm": 4.90625,
1272
+ "learning_rate": 8.550312455289625e-05,
1273
+ "loss": 0.5044,
1274
+ "step": 1620
1275
+ },
1276
+ {
1277
+ "epoch": 1.2794348508634223,
1278
+ "grad_norm": 4.96875,
1279
+ "learning_rate": 8.525647060167063e-05,
1280
+ "loss": 0.6098,
1281
+ "step": 1630
1282
+ },
1283
+ {
1284
+ "epoch": 1.2872841444270016,
1285
+ "grad_norm": 4.1875,
1286
+ "learning_rate": 8.500809829980734e-05,
1287
+ "loss": 0.5911,
1288
+ "step": 1640
1289
+ },
1290
+ {
1291
+ "epoch": 1.2951334379905808,
1292
+ "grad_norm": 5.3125,
1293
+ "learning_rate": 8.4758019752622e-05,
1294
+ "loss": 0.5876,
1295
+ "step": 1650
1296
+ },
1297
+ {
1298
+ "epoch": 1.30298273155416,
1299
+ "grad_norm": 5.09375,
1300
+ "learning_rate": 8.450624714859016e-05,
1301
+ "loss": 0.5618,
1302
+ "step": 1660
1303
+ },
1304
+ {
1305
+ "epoch": 1.3108320251177394,
1306
+ "grad_norm": 3.78125,
1307
+ "learning_rate": 8.425279275875336e-05,
1308
+ "loss": 0.5629,
1309
+ "step": 1670
1310
+ },
1311
+ {
1312
+ "epoch": 1.3186813186813187,
1313
+ "grad_norm": 3.953125,
1314
+ "learning_rate": 8.399766893612096e-05,
1315
+ "loss": 0.5672,
1316
+ "step": 1680
1317
+ },
1318
+ {
1319
+ "epoch": 1.3265306122448979,
1320
+ "grad_norm": 5.5,
1321
+ "learning_rate": 8.374088811506819e-05,
1322
+ "loss": 0.6229,
1323
+ "step": 1690
1324
+ },
1325
+ {
1326
+ "epoch": 1.3343799058084773,
1327
+ "grad_norm": 7.78125,
1328
+ "learning_rate": 8.348246281072998e-05,
1329
+ "loss": 0.5945,
1330
+ "step": 1700
1331
+ },
1332
+ {
1333
+ "epoch": 1.3422291993720565,
1334
+ "grad_norm": 4.125,
1335
+ "learning_rate": 8.322240561839109e-05,
1336
+ "loss": 0.5755,
1337
+ "step": 1710
1338
+ },
1339
+ {
1340
+ "epoch": 1.3500784929356358,
1341
+ "grad_norm": 5.90625,
1342
+ "learning_rate": 8.296072921287217e-05,
1343
+ "loss": 0.5472,
1344
+ "step": 1720
1345
+ },
1346
+ {
1347
+ "epoch": 1.3579277864992152,
1348
+ "grad_norm": 6.34375,
1349
+ "learning_rate": 8.269744634791208e-05,
1350
+ "loss": 0.6287,
1351
+ "step": 1730
1352
+ },
1353
+ {
1354
+ "epoch": 1.3657770800627944,
1355
+ "grad_norm": 4.34375,
1356
+ "learning_rate": 8.243256985554621e-05,
1357
+ "loss": 0.5684,
1358
+ "step": 1740
1359
+ },
1360
+ {
1361
+ "epoch": 1.3736263736263736,
1362
+ "grad_norm": 5.875,
1363
+ "learning_rate": 8.21661126454811e-05,
1364
+ "loss": 0.5455,
1365
+ "step": 1750
1366
+ },
1367
+ {
1368
+ "epoch": 1.3814756671899528,
1369
+ "grad_norm": 4.3125,
1370
+ "learning_rate": 8.189808770446528e-05,
1371
+ "loss": 0.5403,
1372
+ "step": 1760
1373
+ },
1374
+ {
1375
+ "epoch": 1.389324960753532,
1376
+ "grad_norm": 5.03125,
1377
+ "learning_rate": 8.162850809565623e-05,
1378
+ "loss": 0.5343,
1379
+ "step": 1770
1380
+ },
1381
+ {
1382
+ "epoch": 1.3971742543171115,
1383
+ "grad_norm": 4.53125,
1384
+ "learning_rate": 8.135738695798376e-05,
1385
+ "loss": 0.5692,
1386
+ "step": 1780
1387
+ },
1388
+ {
1389
+ "epoch": 1.4050235478806907,
1390
+ "grad_norm": 5.125,
1391
+ "learning_rate": 8.108473750550965e-05,
1392
+ "loss": 0.5382,
1393
+ "step": 1790
1394
+ },
1395
+ {
1396
+ "epoch": 1.41287284144427,
1397
+ "grad_norm": 5.6875,
1398
+ "learning_rate": 8.081057302678352e-05,
1399
+ "loss": 0.5316,
1400
+ "step": 1800
1401
+ },
1402
+ {
1403
+ "epoch": 1.4207221350078494,
1404
+ "grad_norm": 5.15625,
1405
+ "learning_rate": 8.053490688419533e-05,
1406
+ "loss": 0.5221,
1407
+ "step": 1810
1408
+ },
1409
+ {
1410
+ "epoch": 1.4285714285714286,
1411
+ "grad_norm": 6.6875,
1412
+ "learning_rate": 8.02577525133239e-05,
1413
+ "loss": 0.6501,
1414
+ "step": 1820
1415
+ },
1416
+ {
1417
+ "epoch": 1.4364207221350078,
1418
+ "grad_norm": 7.34375,
1419
+ "learning_rate": 7.997912342228232e-05,
1420
+ "loss": 0.5745,
1421
+ "step": 1830
1422
+ },
1423
+ {
1424
+ "epoch": 1.4442700156985873,
1425
+ "grad_norm": 4.15625,
1426
+ "learning_rate": 7.969903319105935e-05,
1427
+ "loss": 0.5797,
1428
+ "step": 1840
1429
+ },
1430
+ {
1431
+ "epoch": 1.4521193092621665,
1432
+ "grad_norm": 6.21875,
1433
+ "learning_rate": 7.941749547085777e-05,
1434
+ "loss": 0.6899,
1435
+ "step": 1850
1436
+ },
1437
+ {
1438
+ "epoch": 1.4599686028257457,
1439
+ "grad_norm": 4.625,
1440
+ "learning_rate": 7.913452398342881e-05,
1441
+ "loss": 0.5891,
1442
+ "step": 1860
1443
+ },
1444
+ {
1445
+ "epoch": 1.467817896389325,
1446
+ "grad_norm": 7.1875,
1447
+ "learning_rate": 7.885013252040359e-05,
1448
+ "loss": 0.6118,
1449
+ "step": 1870
1450
+ },
1451
+ {
1452
+ "epoch": 1.4756671899529041,
1453
+ "grad_norm": 6.8125,
1454
+ "learning_rate": 7.856433494262077e-05,
1455
+ "loss": 0.6102,
1456
+ "step": 1880
1457
+ },
1458
+ {
1459
+ "epoch": 1.4835164835164836,
1460
+ "grad_norm": 5.0,
1461
+ "learning_rate": 7.827714517945115e-05,
1462
+ "loss": 0.5459,
1463
+ "step": 1890
1464
+ },
1465
+ {
1466
+ "epoch": 1.4913657770800628,
1467
+ "grad_norm": 4.40625,
1468
+ "learning_rate": 7.798857722811856e-05,
1469
+ "loss": 0.5717,
1470
+ "step": 1900
1471
+ },
1472
+ {
1473
+ "epoch": 1.499215070643642,
1474
+ "grad_norm": 3.734375,
1475
+ "learning_rate": 7.769864515301787e-05,
1476
+ "loss": 0.5279,
1477
+ "step": 1910
1478
+ },
1479
+ {
1480
+ "epoch": 1.5070643642072215,
1481
+ "grad_norm": 4.6875,
1482
+ "learning_rate": 7.740736308502938e-05,
1483
+ "loss": 0.66,
1484
+ "step": 1920
1485
+ },
1486
+ {
1487
+ "epoch": 1.5149136577708007,
1488
+ "grad_norm": 6.5,
1489
+ "learning_rate": 7.711474522083015e-05,
1490
+ "loss": 0.546,
1491
+ "step": 1930
1492
+ },
1493
+ {
1494
+ "epoch": 1.5227629513343799,
1495
+ "grad_norm": 4.625,
1496
+ "learning_rate": 7.682080582220206e-05,
1497
+ "loss": 0.5207,
1498
+ "step": 1940
1499
+ },
1500
+ {
1501
+ "epoch": 1.5306122448979593,
1502
+ "grad_norm": 6.4375,
1503
+ "learning_rate": 7.65255592153367e-05,
1504
+ "loss": 0.6005,
1505
+ "step": 1950
1506
+ },
1507
+ {
1508
+ "epoch": 1.5384615384615383,
1509
+ "grad_norm": 4.15625,
1510
+ "learning_rate": 7.622901979013717e-05,
1511
+ "loss": 0.5897,
1512
+ "step": 1960
1513
+ },
1514
+ {
1515
+ "epoch": 1.5463108320251178,
1516
+ "grad_norm": 4.40625,
1517
+ "learning_rate": 7.593120199951671e-05,
1518
+ "loss": 0.5639,
1519
+ "step": 1970
1520
+ },
1521
+ {
1522
+ "epoch": 1.554160125588697,
1523
+ "grad_norm": 4.59375,
1524
+ "learning_rate": 7.563212035869425e-05,
1525
+ "loss": 0.5479,
1526
+ "step": 1980
1527
+ },
1528
+ {
1529
+ "epoch": 1.5620094191522762,
1530
+ "grad_norm": 4.8125,
1531
+ "learning_rate": 7.533178944448705e-05,
1532
+ "loss": 0.5539,
1533
+ "step": 1990
1534
+ },
1535
+ {
1536
+ "epoch": 1.5698587127158556,
1537
+ "grad_norm": 6.09375,
1538
+ "learning_rate": 7.503022389460014e-05,
1539
+ "loss": 0.625,
1540
+ "step": 2000
1541
+ },
1542
+ {
1543
+ "epoch": 1.5698587127158556,
1544
+ "eval/nano_beir.arguana": 0.4666002720912347,
1545
+ "step": 2000
1546
+ },
1547
+ {
1548
+ "epoch": 1.5698587127158556,
1549
+ "eval/nano_beir.climate_fever": 0.22462604477746737,
1550
+ "step": 2000
1551
+ },
1552
+ {
1553
+ "epoch": 1.5698587127158556,
1554
+ "eval/nano_beir.dbpedia_entity": 0.3175867497382319,
1555
+ "step": 2000
1556
+ },
1557
+ {
1558
+ "epoch": 1.5698587127158556,
1559
+ "eval/nano_beir.fever": 0.6140103747506924,
1560
+ "step": 2000
1561
+ },
1562
+ {
1563
+ "epoch": 1.5698587127158556,
1564
+ "eval/nano_beir.fiqa": 0.43978199849769206,
1565
+ "step": 2000
1566
+ },
1567
+ {
1568
+ "epoch": 1.5698587127158556,
1569
+ "eval/nano_beir.hotpotqa": 0.6788521964877974,
1570
+ "step": 2000
1571
+ },
1572
+ {
1573
+ "epoch": 1.5698587127158556,
1574
+ "eval/nano_beir.nfcorpus": 0.3485625465712446,
1575
+ "step": 2000
1576
+ },
1577
+ {
1578
+ "epoch": 1.5698587127158556,
1579
+ "eval/nano_beir.nq": 0.42598841107613383,
1580
+ "step": 2000
1581
+ },
1582
+ {
1583
+ "epoch": 1.5698587127158556,
1584
+ "eval/nano_beir.quora": 0.833039807676506,
1585
+ "step": 2000
1586
+ },
1587
+ {
1588
+ "epoch": 1.5698587127158556,
1589
+ "eval/nano_beir.scidocs": 0.24848467409467515,
1590
+ "step": 2000
1591
+ },
1592
+ {
1593
+ "epoch": 1.5698587127158556,
1594
+ "eval/nano_beir.scifact": 0.7682290881646785,
1595
+ "step": 2000
1596
+ },
1597
+ {
1598
+ "epoch": 1.5698587127158556,
1599
+ "eval/nano_beir.webis_touche2020": 0.2836464597787185,
1600
+ "step": 2000
1601
+ },
1602
+ {
1603
+ "epoch": 1.5698587127158556,
1604
+ "eval/avg": 0.47078405197542267,
1605
+ "step": 2000
1606
+ },
1607
+ {
1608
+ "epoch": 1.5777080062794349,
1609
+ "grad_norm": 5.15625,
1610
+ "learning_rate": 7.472743840691299e-05,
1611
+ "loss": 0.4956,
1612
+ "step": 2010
1613
+ },
1614
+ {
1615
+ "epoch": 1.585557299843014,
1616
+ "grad_norm": 4.34375,
1617
+ "learning_rate": 7.44234477387631e-05,
1618
+ "loss": 0.4504,
1619
+ "step": 2020
1620
+ },
1621
+ {
1622
+ "epoch": 1.5934065934065935,
1623
+ "grad_norm": 3.0625,
1624
+ "learning_rate": 7.411826670622676e-05,
1625
+ "loss": 0.5722,
1626
+ "step": 2030
1627
+ },
1628
+ {
1629
+ "epoch": 1.6012558869701727,
1630
+ "grad_norm": 4.875,
1631
+ "learning_rate": 7.381191018339696e-05,
1632
+ "loss": 0.5777,
1633
+ "step": 2040
1634
+ },
1635
+ {
1636
+ "epoch": 1.609105180533752,
1637
+ "grad_norm": 5.34375,
1638
+ "learning_rate": 7.350439310165842e-05,
1639
+ "loss": 0.5645,
1640
+ "step": 2050
1641
+ },
1642
+ {
1643
+ "epoch": 1.6169544740973314,
1644
+ "grad_norm": 5.125,
1645
+ "learning_rate": 7.319573044895986e-05,
1646
+ "loss": 0.5454,
1647
+ "step": 2060
1648
+ },
1649
+ {
1650
+ "epoch": 1.6248037676609104,
1651
+ "grad_norm": 5.4375,
1652
+ "learning_rate": 7.28859372690835e-05,
1653
+ "loss": 0.5681,
1654
+ "step": 2070
1655
+ },
1656
+ {
1657
+ "epoch": 1.6326530612244898,
1658
+ "grad_norm": 4.96875,
1659
+ "learning_rate": 7.257502866091192e-05,
1660
+ "loss": 0.5132,
1661
+ "step": 2080
1662
+ },
1663
+ {
1664
+ "epoch": 1.640502354788069,
1665
+ "grad_norm": 5.125,
1666
+ "learning_rate": 7.226301977769199e-05,
1667
+ "loss": 0.5463,
1668
+ "step": 2090
1669
+ },
1670
+ {
1671
+ "epoch": 1.6483516483516483,
1672
+ "grad_norm": 4.71875,
1673
+ "learning_rate": 7.194992582629654e-05,
1674
+ "loss": 0.6572,
1675
+ "step": 2100
1676
+ },
1677
+ {
1678
+ "epoch": 1.6562009419152277,
1679
+ "grad_norm": 4.625,
1680
+ "learning_rate": 7.163576206648304e-05,
1681
+ "loss": 0.5608,
1682
+ "step": 2110
1683
+ },
1684
+ {
1685
+ "epoch": 1.664050235478807,
1686
+ "grad_norm": 4.8125,
1687
+ "learning_rate": 7.132054381014995e-05,
1688
+ "loss": 0.5352,
1689
+ "step": 2120
1690
+ },
1691
+ {
1692
+ "epoch": 1.6718995290423861,
1693
+ "grad_norm": 6.25,
1694
+ "learning_rate": 7.100428642059033e-05,
1695
+ "loss": 0.6148,
1696
+ "step": 2130
1697
+ },
1698
+ {
1699
+ "epoch": 1.6797488226059656,
1700
+ "grad_norm": 4.75,
1701
+ "learning_rate": 7.068700531174319e-05,
1702
+ "loss": 0.6296,
1703
+ "step": 2140
1704
+ },
1705
+ {
1706
+ "epoch": 1.6875981161695446,
1707
+ "grad_norm": 5.125,
1708
+ "learning_rate": 7.036871594744217e-05,
1709
+ "loss": 0.5741,
1710
+ "step": 2150
1711
+ },
1712
+ {
1713
+ "epoch": 1.695447409733124,
1714
+ "grad_norm": 4.78125,
1715
+ "learning_rate": 7.004943384066187e-05,
1716
+ "loss": 0.5457,
1717
+ "step": 2160
1718
+ },
1719
+ {
1720
+ "epoch": 1.7032967032967035,
1721
+ "grad_norm": 5.1875,
1722
+ "learning_rate": 6.97291745527617e-05,
1723
+ "loss": 0.6556,
1724
+ "step": 2170
1725
+ },
1726
+ {
1727
+ "epoch": 1.7111459968602825,
1728
+ "grad_norm": 5.875,
1729
+ "learning_rate": 6.940795369272754e-05,
1730
+ "loss": 0.6284,
1731
+ "step": 2180
1732
+ },
1733
+ {
1734
+ "epoch": 1.718995290423862,
1735
+ "grad_norm": 4.40625,
1736
+ "learning_rate": 6.908578691641092e-05,
1737
+ "loss": 0.5214,
1738
+ "step": 2190
1739
+ },
1740
+ {
1741
+ "epoch": 1.7268445839874411,
1742
+ "grad_norm": 5.96875,
1743
+ "learning_rate": 6.876268992576604e-05,
1744
+ "loss": 0.5484,
1745
+ "step": 2200
1746
+ },
1747
+ {
1748
+ "epoch": 1.7346938775510203,
1749
+ "grad_norm": 5.75,
1750
+ "learning_rate": 6.843867846808438e-05,
1751
+ "loss": 0.5804,
1752
+ "step": 2210
1753
+ },
1754
+ {
1755
+ "epoch": 1.7425431711145998,
1756
+ "grad_norm": 5.59375,
1757
+ "learning_rate": 6.811376833522729e-05,
1758
+ "loss": 0.5653,
1759
+ "step": 2220
1760
+ },
1761
+ {
1762
+ "epoch": 1.750392464678179,
1763
+ "grad_norm": 4.09375,
1764
+ "learning_rate": 6.778797536285624e-05,
1765
+ "loss": 0.5328,
1766
+ "step": 2230
1767
+ },
1768
+ {
1769
+ "epoch": 1.7582417582417582,
1770
+ "grad_norm": 4.0,
1771
+ "learning_rate": 6.746131542966112e-05,
1772
+ "loss": 0.5909,
1773
+ "step": 2240
1774
+ },
1775
+ {
1776
+ "epoch": 1.7660910518053377,
1777
+ "grad_norm": 5.6875,
1778
+ "learning_rate": 6.713380445658618e-05,
1779
+ "loss": 0.5715,
1780
+ "step": 2250
1781
+ },
1782
+ {
1783
+ "epoch": 1.7739403453689166,
1784
+ "grad_norm": 5.09375,
1785
+ "learning_rate": 6.680545840605423e-05,
1786
+ "loss": 0.4487,
1787
+ "step": 2260
1788
+ },
1789
+ {
1790
+ "epoch": 1.781789638932496,
1791
+ "grad_norm": 4.9375,
1792
+ "learning_rate": 6.647629328118851e-05,
1793
+ "loss": 0.5661,
1794
+ "step": 2270
1795
+ },
1796
+ {
1797
+ "epoch": 1.7896389324960753,
1798
+ "grad_norm": 4.65625,
1799
+ "learning_rate": 6.614632512503288e-05,
1800
+ "loss": 0.5829,
1801
+ "step": 2280
1802
+ },
1803
+ {
1804
+ "epoch": 1.7974882260596545,
1805
+ "grad_norm": 5.1875,
1806
+ "learning_rate": 6.58155700197697e-05,
1807
+ "loss": 0.5994,
1808
+ "step": 2290
1809
+ },
1810
+ {
1811
+ "epoch": 1.805337519623234,
1812
+ "grad_norm": 4.71875,
1813
+ "learning_rate": 6.548404408593621e-05,
1814
+ "loss": 0.579,
1815
+ "step": 2300
1816
+ },
1817
+ {
1818
+ "epoch": 1.8131868131868132,
1819
+ "grad_norm": 4.25,
1820
+ "learning_rate": 6.515176348163871e-05,
1821
+ "loss": 0.488,
1822
+ "step": 2310
1823
+ },
1824
+ {
1825
+ "epoch": 1.8210361067503924,
1826
+ "grad_norm": 4.25,
1827
+ "learning_rate": 6.481874440176506e-05,
1828
+ "loss": 0.5083,
1829
+ "step": 2320
1830
+ },
1831
+ {
1832
+ "epoch": 1.8288854003139718,
1833
+ "grad_norm": 5.6875,
1834
+ "learning_rate": 6.448500307719538e-05,
1835
+ "loss": 0.6211,
1836
+ "step": 2330
1837
+ },
1838
+ {
1839
+ "epoch": 1.836734693877551,
1840
+ "grad_norm": 4.96875,
1841
+ "learning_rate": 6.415055577401102e-05,
1842
+ "loss": 0.5724,
1843
+ "step": 2340
1844
+ },
1845
+ {
1846
+ "epoch": 1.8445839874411303,
1847
+ "grad_norm": 3.90625,
1848
+ "learning_rate": 6.381541879270169e-05,
1849
+ "loss": 0.5523,
1850
+ "step": 2350
1851
+ },
1852
+ {
1853
+ "epoch": 1.8524332810047097,
1854
+ "grad_norm": 4.03125,
1855
+ "learning_rate": 6.347960846737106e-05,
1856
+ "loss": 0.5615,
1857
+ "step": 2360
1858
+ },
1859
+ {
1860
+ "epoch": 1.8602825745682887,
1861
+ "grad_norm": 4.375,
1862
+ "learning_rate": 6.31431411649406e-05,
1863
+ "loss": 0.4643,
1864
+ "step": 2370
1865
+ },
1866
+ {
1867
+ "epoch": 1.8681318681318682,
1868
+ "grad_norm": 5.09375,
1869
+ "learning_rate": 6.280603328435198e-05,
1870
+ "loss": 0.5643,
1871
+ "step": 2380
1872
+ },
1873
+ {
1874
+ "epoch": 1.8759811616954474,
1875
+ "grad_norm": 5.375,
1876
+ "learning_rate": 6.24683012557677e-05,
1877
+ "loss": 0.5446,
1878
+ "step": 2390
1879
+ },
1880
+ {
1881
+ "epoch": 1.8838304552590266,
1882
+ "grad_norm": 4.84375,
1883
+ "learning_rate": 6.212996153977037e-05,
1884
+ "loss": 0.5759,
1885
+ "step": 2400
1886
+ },
1887
+ {
1888
+ "epoch": 1.891679748822606,
1889
+ "grad_norm": 5.46875,
1890
+ "learning_rate": 6.179103062656042e-05,
1891
+ "loss": 0.5231,
1892
+ "step": 2410
1893
+ },
1894
+ {
1895
+ "epoch": 1.8995290423861853,
1896
+ "grad_norm": 5.5625,
1897
+ "learning_rate": 6.145152503515239e-05,
1898
+ "loss": 0.5679,
1899
+ "step": 2420
1900
+ },
1901
+ {
1902
+ "epoch": 1.9073783359497645,
1903
+ "grad_norm": 4.78125,
1904
+ "learning_rate": 6.111146131256983e-05,
1905
+ "loss": 0.5482,
1906
+ "step": 2430
1907
+ },
1908
+ {
1909
+ "epoch": 1.915227629513344,
1910
+ "grad_norm": 5.1875,
1911
+ "learning_rate": 6.077085603303882e-05,
1912
+ "loss": 0.5287,
1913
+ "step": 2440
1914
+ },
1915
+ {
1916
+ "epoch": 1.9230769230769231,
1917
+ "grad_norm": 5.8125,
1918
+ "learning_rate": 6.042972579718019e-05,
1919
+ "loss": 0.6531,
1920
+ "step": 2450
1921
+ },
1922
+ {
1923
+ "epoch": 1.9309262166405023,
1924
+ "grad_norm": 4.75,
1925
+ "learning_rate": 6.008808723120035e-05,
1926
+ "loss": 0.5463,
1927
+ "step": 2460
1928
+ },
1929
+ {
1930
+ "epoch": 1.9387755102040818,
1931
+ "grad_norm": 4.0625,
1932
+ "learning_rate": 5.974595698608103e-05,
1933
+ "loss": 0.5132,
1934
+ "step": 2470
1935
+ },
1936
+ {
1937
+ "epoch": 1.9466248037676608,
1938
+ "grad_norm": 4.6875,
1939
+ "learning_rate": 5.9403351736767696e-05,
1940
+ "loss": 0.6158,
1941
+ "step": 2480
1942
+ },
1943
+ {
1944
+ "epoch": 1.9544740973312402,
1945
+ "grad_norm": 4.84375,
1946
+ "learning_rate": 5.906028818135687e-05,
1947
+ "loss": 0.5588,
1948
+ "step": 2490
1949
+ },
1950
+ {
1951
+ "epoch": 1.9623233908948194,
1952
+ "grad_norm": 5.125,
1953
+ "learning_rate": 5.8716783040282244e-05,
1954
+ "loss": 0.5456,
1955
+ "step": 2500
1956
+ },
1957
+ {
1958
+ "epoch": 1.9701726844583987,
1959
+ "grad_norm": 4.65625,
1960
+ "learning_rate": 5.837285305549978e-05,
1961
+ "loss": 0.6038,
1962
+ "step": 2510
1963
+ },
1964
+ {
1965
+ "epoch": 1.978021978021978,
1966
+ "grad_norm": 5.25,
1967
+ "learning_rate": 5.8028514989671724e-05,
1968
+ "loss": 0.5504,
1969
+ "step": 2520
1970
+ },
1971
+ {
1972
+ "epoch": 1.9858712715855573,
1973
+ "grad_norm": 5.0,
1974
+ "learning_rate": 5.7683785625349616e-05,
1975
+ "loss": 0.5705,
1976
+ "step": 2530
1977
+ },
1978
+ {
1979
+ "epoch": 1.9937205651491365,
1980
+ "grad_norm": 6.21875,
1981
+ "learning_rate": 5.733868176415633e-05,
1982
+ "loss": 0.5866,
1983
+ "step": 2540
1984
+ },
1985
+ {
1986
+ "epoch": 2.001569858712716,
1987
+ "grad_norm": 5.125,
1988
+ "learning_rate": 5.699322022596722e-05,
1989
+ "loss": 0.6031,
1990
+ "step": 2550
1991
+ },
1992
+ {
1993
+ "epoch": 2.009419152276295,
1994
+ "grad_norm": 4.5625,
1995
+ "learning_rate": 5.664741784809022e-05,
1996
+ "loss": 0.5695,
1997
+ "step": 2560
1998
+ },
1999
+ {
2000
+ "epoch": 2.0172684458398744,
2001
+ "grad_norm": 5.65625,
2002
+ "learning_rate": 5.630129148444543e-05,
2003
+ "loss": 0.5491,
2004
+ "step": 2570
2005
+ },
2006
+ {
2007
+ "epoch": 2.025117739403454,
2008
+ "grad_norm": 3.78125,
2009
+ "learning_rate": 5.595485800474349e-05,
2010
+ "loss": 0.5192,
2011
+ "step": 2580
2012
+ },
2013
+ {
2014
+ "epoch": 2.032967032967033,
2015
+ "grad_norm": 6.25,
2016
+ "learning_rate": 5.560813429366345e-05,
2017
+ "loss": 0.5584,
2018
+ "step": 2590
2019
+ },
2020
+ {
2021
+ "epoch": 2.0408163265306123,
2022
+ "grad_norm": 4.78125,
2023
+ "learning_rate": 5.5261137250029835e-05,
2024
+ "loss": 0.4623,
2025
+ "step": 2600
2026
+ },
2027
+ {
2028
+ "epoch": 2.0486656200941917,
2029
+ "grad_norm": 4.53125,
2030
+ "learning_rate": 5.4913883785988993e-05,
2031
+ "loss": 0.5125,
2032
+ "step": 2610
2033
+ },
2034
+ {
2035
+ "epoch": 2.0565149136577707,
2036
+ "grad_norm": 4.5,
2037
+ "learning_rate": 5.456639082618489e-05,
2038
+ "loss": 0.4937,
2039
+ "step": 2620
2040
+ },
2041
+ {
2042
+ "epoch": 2.06436420722135,
2043
+ "grad_norm": 4.65625,
2044
+ "learning_rate": 5.421867530693414e-05,
2045
+ "loss": 0.5231,
2046
+ "step": 2630
2047
+ },
2048
+ {
2049
+ "epoch": 2.072213500784929,
2050
+ "grad_norm": 3.546875,
2051
+ "learning_rate": 5.3870754175400595e-05,
2052
+ "loss": 0.422,
2053
+ "step": 2640
2054
+ },
2055
+ {
2056
+ "epoch": 2.0800627943485086,
2057
+ "grad_norm": 4.40625,
2058
+ "learning_rate": 5.352264438876935e-05,
2059
+ "loss": 0.5353,
2060
+ "step": 2650
2061
+ },
2062
+ {
2063
+ "epoch": 2.087912087912088,
2064
+ "grad_norm": 4.34375,
2065
+ "learning_rate": 5.3174362913420306e-05,
2066
+ "loss": 0.4964,
2067
+ "step": 2660
2068
+ },
2069
+ {
2070
+ "epoch": 2.095761381475667,
2071
+ "grad_norm": 4.28125,
2072
+ "learning_rate": 5.2825926724101236e-05,
2073
+ "loss": 0.5701,
2074
+ "step": 2670
2075
+ },
2076
+ {
2077
+ "epoch": 2.1036106750392465,
2078
+ "grad_norm": 5.5,
2079
+ "learning_rate": 5.247735280310041e-05,
2080
+ "loss": 0.5823,
2081
+ "step": 2680
2082
+ },
2083
+ {
2084
+ "epoch": 2.111459968602826,
2085
+ "grad_norm": 4.125,
2086
+ "learning_rate": 5.212865813941899e-05,
2087
+ "loss": 0.4692,
2088
+ "step": 2690
2089
+ },
2090
+ {
2091
+ "epoch": 2.119309262166405,
2092
+ "grad_norm": 6.9375,
2093
+ "learning_rate": 5.1779859727942924e-05,
2094
+ "loss": 0.5522,
2095
+ "step": 2700
2096
+ },
2097
+ {
2098
+ "epoch": 2.1271585557299844,
2099
+ "grad_norm": 5.78125,
2100
+ "learning_rate": 5.143097456861474e-05,
2101
+ "loss": 0.5413,
2102
+ "step": 2710
2103
+ },
2104
+ {
2105
+ "epoch": 2.1350078492935634,
2106
+ "grad_norm": 4.625,
2107
+ "learning_rate": 5.1082019665604895e-05,
2108
+ "loss": 0.4929,
2109
+ "step": 2720
2110
+ },
2111
+ {
2112
+ "epoch": 2.142857142857143,
2113
+ "grad_norm": 5.5625,
2114
+ "learning_rate": 5.073301202648304e-05,
2115
+ "loss": 0.5183,
2116
+ "step": 2730
2117
+ },
2118
+ {
2119
+ "epoch": 2.1507064364207222,
2120
+ "grad_norm": 6.0,
2121
+ "learning_rate": 5.0383968661389146e-05,
2122
+ "loss": 0.5698,
2123
+ "step": 2740
2124
+ },
2125
+ {
2126
+ "epoch": 2.1585557299843012,
2127
+ "grad_norm": 5.0,
2128
+ "learning_rate": 5.003490658220438e-05,
2129
+ "loss": 0.4615,
2130
+ "step": 2750
2131
+ },
2132
+ {
2133
+ "epoch": 2.1664050235478807,
2134
+ "grad_norm": 6.09375,
2135
+ "learning_rate": 4.968584280172206e-05,
2136
+ "loss": 0.5037,
2137
+ "step": 2760
2138
+ },
2139
+ {
2140
+ "epoch": 2.17425431711146,
2141
+ "grad_norm": 7.5625,
2142
+ "learning_rate": 4.933679433281836e-05,
2143
+ "loss": 0.5184,
2144
+ "step": 2770
2145
+ },
2146
+ {
2147
+ "epoch": 2.182103610675039,
2148
+ "grad_norm": 5.8125,
2149
+ "learning_rate": 4.8987778187623245e-05,
2150
+ "loss": 0.5307,
2151
+ "step": 2780
2152
+ },
2153
+ {
2154
+ "epoch": 2.1899529042386185,
2155
+ "grad_norm": 5.28125,
2156
+ "learning_rate": 4.863881137669123e-05,
2157
+ "loss": 0.4549,
2158
+ "step": 2790
2159
+ },
2160
+ {
2161
+ "epoch": 2.197802197802198,
2162
+ "grad_norm": 4.03125,
2163
+ "learning_rate": 4.8289910908172376e-05,
2164
+ "loss": 0.5787,
2165
+ "step": 2800
2166
+ },
2167
+ {
2168
+ "epoch": 2.205651491365777,
2169
+ "grad_norm": 4.8125,
2170
+ "learning_rate": 4.794109378698327e-05,
2171
+ "loss": 0.6173,
2172
+ "step": 2810
2173
+ },
2174
+ {
2175
+ "epoch": 2.2135007849293564,
2176
+ "grad_norm": 5.6875,
2177
+ "learning_rate": 4.7592377013978306e-05,
2178
+ "loss": 0.5365,
2179
+ "step": 2820
2180
+ },
2181
+ {
2182
+ "epoch": 2.221350078492936,
2183
+ "grad_norm": 5.125,
2184
+ "learning_rate": 4.724377758512103e-05,
2185
+ "loss": 0.4651,
2186
+ "step": 2830
2187
+ },
2188
+ {
2189
+ "epoch": 2.229199372056515,
2190
+ "grad_norm": 6.15625,
2191
+ "learning_rate": 4.68953124906558e-05,
2192
+ "loss": 0.5659,
2193
+ "step": 2840
2194
+ },
2195
+ {
2196
+ "epoch": 2.2370486656200943,
2197
+ "grad_norm": 4.71875,
2198
+ "learning_rate": 4.654699871427971e-05,
2199
+ "loss": 0.6009,
2200
+ "step": 2850
2201
+ },
2202
+ {
2203
+ "epoch": 2.2448979591836733,
2204
+ "grad_norm": 6.65625,
2205
+ "learning_rate": 4.619885323231484e-05,
2206
+ "loss": 0.5581,
2207
+ "step": 2860
2208
+ },
2209
+ {
2210
+ "epoch": 2.2527472527472527,
2211
+ "grad_norm": 5.25,
2212
+ "learning_rate": 4.58508930128808e-05,
2213
+ "loss": 0.5357,
2214
+ "step": 2870
2215
+ },
2216
+ {
2217
+ "epoch": 2.260596546310832,
2218
+ "grad_norm": 5.625,
2219
+ "learning_rate": 4.550313501506781e-05,
2220
+ "loss": 0.5235,
2221
+ "step": 2880
2222
+ },
2223
+ {
2224
+ "epoch": 2.268445839874411,
2225
+ "grad_norm": 4.875,
2226
+ "learning_rate": 4.515559618811005e-05,
2227
+ "loss": 0.442,
2228
+ "step": 2890
2229
+ },
2230
+ {
2231
+ "epoch": 2.2762951334379906,
2232
+ "grad_norm": 5.375,
2233
+ "learning_rate": 4.4808293470559643e-05,
2234
+ "loss": 0.4889,
2235
+ "step": 2900
2236
+ },
2237
+ {
2238
+ "epoch": 2.2841444270015696,
2239
+ "grad_norm": 5.28125,
2240
+ "learning_rate": 4.446124378946107e-05,
2241
+ "loss": 0.6102,
2242
+ "step": 2910
2243
+ },
2244
+ {
2245
+ "epoch": 2.291993720565149,
2246
+ "grad_norm": 6.78125,
2247
+ "learning_rate": 4.411446405952618e-05,
2248
+ "loss": 0.5462,
2249
+ "step": 2920
2250
+ },
2251
+ {
2252
+ "epoch": 2.2998430141287285,
2253
+ "grad_norm": 5.21875,
2254
+ "learning_rate": 4.376797118230979e-05,
2255
+ "loss": 0.4926,
2256
+ "step": 2930
2257
+ },
2258
+ {
2259
+ "epoch": 2.3076923076923075,
2260
+ "grad_norm": 5.5,
2261
+ "learning_rate": 4.342178204538588e-05,
2262
+ "loss": 0.5301,
2263
+ "step": 2940
2264
+ },
2265
+ {
2266
+ "epoch": 2.315541601255887,
2267
+ "grad_norm": 5.84375,
2268
+ "learning_rate": 4.307591352152459e-05,
2269
+ "loss": 0.5053,
2270
+ "step": 2950
2271
+ },
2272
+ {
2273
+ "epoch": 2.3233908948194664,
2274
+ "grad_norm": 5.0,
2275
+ "learning_rate": 4.273038246786986e-05,
2276
+ "loss": 0.5717,
2277
+ "step": 2960
2278
+ },
2279
+ {
2280
+ "epoch": 2.3312401883830454,
2281
+ "grad_norm": 5.0625,
2282
+ "learning_rate": 4.238520572511773e-05,
2283
+ "loss": 0.5054,
2284
+ "step": 2970
2285
+ },
2286
+ {
2287
+ "epoch": 2.339089481946625,
2288
+ "grad_norm": 4.53125,
2289
+ "learning_rate": 4.2040400116695666e-05,
2290
+ "loss": 0.4606,
2291
+ "step": 2980
2292
+ },
2293
+ {
2294
+ "epoch": 2.3469387755102042,
2295
+ "grad_norm": 4.0625,
2296
+ "learning_rate": 4.169598244794261e-05,
2297
+ "loss": 0.4655,
2298
+ "step": 2990
2299
+ },
2300
+ {
2301
+ "epoch": 2.3547880690737832,
2302
+ "grad_norm": 5.6875,
2303
+ "learning_rate": 4.135196950528982e-05,
2304
+ "loss": 0.4932,
2305
+ "step": 3000
2306
+ },
2307
+ {
2308
+ "epoch": 2.3547880690737832,
2309
+ "eval/nano_beir.arguana": 0.4735606937464546,
2310
+ "step": 3000
2311
+ },
2312
+ {
2313
+ "epoch": 2.3547880690737832,
2314
+ "eval/nano_beir.climate_fever": 0.23710577392038215,
2315
+ "step": 3000
2316
+ },
2317
+ {
2318
+ "epoch": 2.3547880690737832,
2319
+ "eval/nano_beir.dbpedia_entity": 0.32334770186384726,
2320
+ "step": 3000
2321
+ },
2322
+ {
2323
+ "epoch": 2.3547880690737832,
2324
+ "eval/nano_beir.fever": 0.6341234371233305,
2325
+ "step": 3000
2326
+ },
2327
+ {
2328
+ "epoch": 2.3547880690737832,
2329
+ "eval/nano_beir.fiqa": 0.44598842357063445,
2330
+ "step": 3000
2331
+ },
2332
+ {
2333
+ "epoch": 2.3547880690737832,
2334
+ "eval/nano_beir.hotpotqa": 0.6704826316013306,
2335
+ "step": 3000
2336
+ },
2337
+ {
2338
+ "epoch": 2.3547880690737832,
2339
+ "eval/nano_beir.nfcorpus": 0.35000164335019307,
2340
+ "step": 3000
2341
+ },
2342
+ {
2343
+ "epoch": 2.3547880690737832,
2344
+ "eval/nano_beir.nq": 0.40244906153552973,
2345
+ "step": 3000
2346
+ },
2347
+ {
2348
+ "epoch": 2.3547880690737832,
2349
+ "eval/nano_beir.quora": 0.822792613681568,
2350
+ "step": 3000
2351
+ },
2352
+ {
2353
+ "epoch": 2.3547880690737832,
2354
+ "eval/nano_beir.scidocs": 0.24785649115277483,
2355
+ "step": 3000
2356
+ },
2357
+ {
2358
+ "epoch": 2.3547880690737832,
2359
+ "eval/nano_beir.scifact": 0.7593341539154295,
2360
+ "step": 3000
2361
+ },
2362
+ {
2363
+ "epoch": 2.3547880690737832,
2364
+ "eval/nano_beir.webis_touche2020": 0.28737508267615175,
2365
+ "step": 3000
2366
+ },
2367
+ {
2368
+ "epoch": 2.3547880690737832,
2369
+ "eval/avg": 0.47120147567813553,
2370
+ "step": 3000
2371
+ },
2372
+ {
2373
+ "epoch": 2.3626373626373627,
2374
+ "grad_norm": 4.625,
2375
+ "learning_rate": 4.100837805544279e-05,
2376
+ "loss": 0.5136,
2377
+ "step": 3010
2378
+ },
2379
+ {
2380
+ "epoch": 2.370486656200942,
2381
+ "grad_norm": 4.84375,
2382
+ "learning_rate": 4.0665224844564053e-05,
2383
+ "loss": 0.5163,
2384
+ "step": 3020
2385
+ },
2386
+ {
2387
+ "epoch": 2.378335949764521,
2388
+ "grad_norm": 4.71875,
2389
+ "learning_rate": 4.032252659745699e-05,
2390
+ "loss": 0.4874,
2391
+ "step": 3030
2392
+ },
2393
+ {
2394
+ "epoch": 2.3861852433281006,
2395
+ "grad_norm": 5.90625,
2396
+ "learning_rate": 3.9980300016750694e-05,
2397
+ "loss": 0.5288,
2398
+ "step": 3040
2399
+ },
2400
+ {
2401
+ "epoch": 2.39403453689168,
2402
+ "grad_norm": 6.28125,
2403
+ "learning_rate": 3.963856178208588e-05,
2404
+ "loss": 0.5765,
2405
+ "step": 3050
2406
+ },
2407
+ {
2408
+ "epoch": 2.401883830455259,
2409
+ "grad_norm": 5.75,
2410
+ "learning_rate": 3.9297328549302e-05,
2411
+ "loss": 0.5288,
2412
+ "step": 3060
2413
+ },
2414
+ {
2415
+ "epoch": 2.4097331240188384,
2416
+ "grad_norm": 6.1875,
2417
+ "learning_rate": 3.895661694962542e-05,
2418
+ "loss": 0.6263,
2419
+ "step": 3070
2420
+ },
2421
+ {
2422
+ "epoch": 2.4175824175824174,
2423
+ "grad_norm": 5.65625,
2424
+ "learning_rate": 3.8616443588858805e-05,
2425
+ "loss": 0.5104,
2426
+ "step": 3080
2427
+ },
2428
+ {
2429
+ "epoch": 2.425431711145997,
2430
+ "grad_norm": 5.9375,
2431
+ "learning_rate": 3.827682504657187e-05,
2432
+ "loss": 0.5545,
2433
+ "step": 3090
2434
+ },
2435
+ {
2436
+ "epoch": 2.4332810047095763,
2437
+ "grad_norm": 4.65625,
2438
+ "learning_rate": 3.7937777875293244e-05,
2439
+ "loss": 0.4756,
2440
+ "step": 3100
2441
+ },
2442
+ {
2443
+ "epoch": 2.4411302982731553,
2444
+ "grad_norm": 4.5625,
2445
+ "learning_rate": 3.759931859970374e-05,
2446
+ "loss": 0.568,
2447
+ "step": 3110
2448
+ },
2449
+ {
2450
+ "epoch": 2.4489795918367347,
2451
+ "grad_norm": 5.1875,
2452
+ "learning_rate": 3.72614637158309e-05,
2453
+ "loss": 0.4515,
2454
+ "step": 3120
2455
+ },
2456
+ {
2457
+ "epoch": 2.4568288854003137,
2458
+ "grad_norm": 4.3125,
2459
+ "learning_rate": 3.692422969024516e-05,
2460
+ "loss": 0.4946,
2461
+ "step": 3130
2462
+ },
2463
+ {
2464
+ "epoch": 2.464678178963893,
2465
+ "grad_norm": 6.8125,
2466
+ "learning_rate": 3.6587632959257165e-05,
2467
+ "loss": 0.553,
2468
+ "step": 3140
2469
+ },
2470
+ {
2471
+ "epoch": 2.4725274725274726,
2472
+ "grad_norm": 5.78125,
2473
+ "learning_rate": 3.625168992811671e-05,
2474
+ "loss": 0.4509,
2475
+ "step": 3150
2476
+ },
2477
+ {
2478
+ "epoch": 2.4803767660910516,
2479
+ "grad_norm": 4.15625,
2480
+ "learning_rate": 3.591641697021317e-05,
2481
+ "loss": 0.4393,
2482
+ "step": 3160
2483
+ },
2484
+ {
2485
+ "epoch": 2.488226059654631,
2486
+ "grad_norm": 5.59375,
2487
+ "learning_rate": 3.558183042627755e-05,
2488
+ "loss": 0.5044,
2489
+ "step": 3170
2490
+ },
2491
+ {
2492
+ "epoch": 2.4960753532182105,
2493
+ "grad_norm": 4.15625,
2494
+ "learning_rate": 3.524794660358593e-05,
2495
+ "loss": 0.4536,
2496
+ "step": 3180
2497
+ },
2498
+ {
2499
+ "epoch": 2.5039246467817895,
2500
+ "grad_norm": 3.9375,
2501
+ "learning_rate": 3.491478177516484e-05,
2502
+ "loss": 0.5021,
2503
+ "step": 3190
2504
+ },
2505
+ {
2506
+ "epoch": 2.511773940345369,
2507
+ "grad_norm": 4.0,
2508
+ "learning_rate": 3.4582352178997935e-05,
2509
+ "loss": 0.5415,
2510
+ "step": 3200
2511
+ },
2512
+ {
2513
+ "epoch": 2.5196232339089484,
2514
+ "grad_norm": 6.125,
2515
+ "learning_rate": 3.425067401723477e-05,
2516
+ "loss": 0.542,
2517
+ "step": 3210
2518
+ },
2519
+ {
2520
+ "epoch": 2.5274725274725274,
2521
+ "grad_norm": 5.3125,
2522
+ "learning_rate": 3.3919763455401015e-05,
2523
+ "loss": 0.5314,
2524
+ "step": 3220
2525
+ },
2526
+ {
2527
+ "epoch": 2.535321821036107,
2528
+ "grad_norm": 4.1875,
2529
+ "learning_rate": 3.358963662161062e-05,
2530
+ "loss": 0.4573,
2531
+ "step": 3230
2532
+ },
2533
+ {
2534
+ "epoch": 2.5431711145996863,
2535
+ "grad_norm": 4.9375,
2536
+ "learning_rate": 3.326030960577972e-05,
2537
+ "loss": 0.5061,
2538
+ "step": 3240
2539
+ },
2540
+ {
2541
+ "epoch": 2.5510204081632653,
2542
+ "grad_norm": 3.25,
2543
+ "learning_rate": 3.293179845884245e-05,
2544
+ "loss": 0.4469,
2545
+ "step": 3250
2546
+ },
2547
+ {
2548
+ "epoch": 2.5588697017268447,
2549
+ "grad_norm": 6.15625,
2550
+ "learning_rate": 3.2604119191968654e-05,
2551
+ "loss": 0.4709,
2552
+ "step": 3260
2553
+ },
2554
+ {
2555
+ "epoch": 2.566718995290424,
2556
+ "grad_norm": 4.3125,
2557
+ "learning_rate": 3.2277287775783525e-05,
2558
+ "loss": 0.5695,
2559
+ "step": 3270
2560
+ },
2561
+ {
2562
+ "epoch": 2.574568288854003,
2563
+ "grad_norm": 5.5625,
2564
+ "learning_rate": 3.195132013958918e-05,
2565
+ "loss": 0.5323,
2566
+ "step": 3280
2567
+ },
2568
+ {
2569
+ "epoch": 2.5824175824175826,
2570
+ "grad_norm": 7.09375,
2571
+ "learning_rate": 3.162623217058834e-05,
2572
+ "loss": 0.5312,
2573
+ "step": 3290
2574
+ },
2575
+ {
2576
+ "epoch": 2.5902668759811616,
2577
+ "grad_norm": 6.34375,
2578
+ "learning_rate": 3.130203971310999e-05,
2579
+ "loss": 0.5346,
2580
+ "step": 3300
2581
+ },
2582
+ {
2583
+ "epoch": 2.598116169544741,
2584
+ "grad_norm": 5.53125,
2585
+ "learning_rate": 3.097875856783713e-05,
2586
+ "loss": 0.4832,
2587
+ "step": 3310
2588
+ },
2589
+ {
2590
+ "epoch": 2.60596546310832,
2591
+ "grad_norm": 4.6875,
2592
+ "learning_rate": 3.065640449103669e-05,
2593
+ "loss": 0.5371,
2594
+ "step": 3320
2595
+ },
2596
+ {
2597
+ "epoch": 2.6138147566718994,
2598
+ "grad_norm": 4.84375,
2599
+ "learning_rate": 3.033499319379163e-05,
2600
+ "loss": 0.4707,
2601
+ "step": 3330
2602
+ },
2603
+ {
2604
+ "epoch": 2.621664050235479,
2605
+ "grad_norm": 5.125,
2606
+ "learning_rate": 3.001454034123512e-05,
2607
+ "loss": 0.5356,
2608
+ "step": 3340
2609
+ },
2610
+ {
2611
+ "epoch": 2.629513343799058,
2612
+ "grad_norm": 6.65625,
2613
+ "learning_rate": 2.9695061551787108e-05,
2614
+ "loss": 0.5273,
2615
+ "step": 3350
2616
+ },
2617
+ {
2618
+ "epoch": 2.6373626373626373,
2619
+ "grad_norm": 4.46875,
2620
+ "learning_rate": 2.9376572396393048e-05,
2621
+ "loss": 0.5746,
2622
+ "step": 3360
2623
+ },
2624
+ {
2625
+ "epoch": 2.6452119309262168,
2626
+ "grad_norm": 5.34375,
2627
+ "learning_rate": 2.9059088397765088e-05,
2628
+ "loss": 0.5145,
2629
+ "step": 3370
2630
+ },
2631
+ {
2632
+ "epoch": 2.6530612244897958,
2633
+ "grad_norm": 5.59375,
2634
+ "learning_rate": 2.874262502962537e-05,
2635
+ "loss": 0.4906,
2636
+ "step": 3380
2637
+ },
2638
+ {
2639
+ "epoch": 2.660910518053375,
2640
+ "grad_norm": 4.625,
2641
+ "learning_rate": 2.8427197715952047e-05,
2642
+ "loss": 0.4909,
2643
+ "step": 3390
2644
+ },
2645
+ {
2646
+ "epoch": 2.6687598116169546,
2647
+ "grad_norm": 5.28125,
2648
+ "learning_rate": 2.811282183022736e-05,
2649
+ "loss": 0.535,
2650
+ "step": 3400
2651
+ },
2652
+ {
2653
+ "epoch": 2.6766091051805336,
2654
+ "grad_norm": 5.3125,
2655
+ "learning_rate": 2.7799512694688468e-05,
2656
+ "loss": 0.524,
2657
+ "step": 3410
2658
+ },
2659
+ {
2660
+ "epoch": 2.684458398744113,
2661
+ "grad_norm": 2.984375,
2662
+ "learning_rate": 2.7487285579580637e-05,
2663
+ "loss": 0.4084,
2664
+ "step": 3420
2665
+ },
2666
+ {
2667
+ "epoch": 2.6923076923076925,
2668
+ "grad_norm": 5.0,
2669
+ "learning_rate": 2.7176155702412937e-05,
2670
+ "loss": 0.5571,
2671
+ "step": 3430
2672
+ },
2673
+ {
2674
+ "epoch": 2.7001569858712715,
2675
+ "grad_norm": 4.21875,
2676
+ "learning_rate": 2.686613822721666e-05,
2677
+ "loss": 0.4846,
2678
+ "step": 3440
2679
+ },
2680
+ {
2681
+ "epoch": 2.708006279434851,
2682
+ "grad_norm": 4.53125,
2683
+ "learning_rate": 2.6557248263806174e-05,
2684
+ "loss": 0.5363,
2685
+ "step": 3450
2686
+ },
2687
+ {
2688
+ "epoch": 2.7158555729984304,
2689
+ "grad_norm": 5.3125,
2690
+ "learning_rate": 2.6249500867042522e-05,
2691
+ "loss": 0.5401,
2692
+ "step": 3460
2693
+ },
2694
+ {
2695
+ "epoch": 2.7237048665620094,
2696
+ "grad_norm": 5.53125,
2697
+ "learning_rate": 2.5942911036099655e-05,
2698
+ "loss": 0.5538,
2699
+ "step": 3470
2700
+ },
2701
+ {
2702
+ "epoch": 2.731554160125589,
2703
+ "grad_norm": 4.75,
2704
+ "learning_rate": 2.5637493713733374e-05,
2705
+ "loss": 0.5288,
2706
+ "step": 3480
2707
+ },
2708
+ {
2709
+ "epoch": 2.739403453689168,
2710
+ "grad_norm": 5.0625,
2711
+ "learning_rate": 2.533326378555314e-05,
2712
+ "loss": 0.5771,
2713
+ "step": 3490
2714
+ },
2715
+ {
2716
+ "epoch": 2.7472527472527473,
2717
+ "grad_norm": 3.953125,
2718
+ "learning_rate": 2.5030236079296444e-05,
2719
+ "loss": 0.478,
2720
+ "step": 3500
2721
+ },
2722
+ {
2723
+ "epoch": 2.7551020408163263,
2724
+ "grad_norm": 4.34375,
2725
+ "learning_rate": 2.4728425364106135e-05,
2726
+ "loss": 0.4903,
2727
+ "step": 3510
2728
+ },
2729
+ {
2730
+ "epoch": 2.7629513343799057,
2731
+ "grad_norm": 4.09375,
2732
+ "learning_rate": 2.442784634981071e-05,
2733
+ "loss": 0.4784,
2734
+ "step": 3520
2735
+ },
2736
+ {
2737
+ "epoch": 2.770800627943485,
2738
+ "grad_norm": 4.84375,
2739
+ "learning_rate": 2.412851368620726e-05,
2740
+ "loss": 0.5052,
2741
+ "step": 3530
2742
+ },
2743
+ {
2744
+ "epoch": 2.778649921507064,
2745
+ "grad_norm": 5.0,
2746
+ "learning_rate": 2.383044196234753e-05,
2747
+ "loss": 0.5316,
2748
+ "step": 3540
2749
+ },
2750
+ {
2751
+ "epoch": 2.7864992150706436,
2752
+ "grad_norm": 5.96875,
2753
+ "learning_rate": 2.3533645705826806e-05,
2754
+ "loss": 0.4896,
2755
+ "step": 3550
2756
+ },
2757
+ {
2758
+ "epoch": 2.794348508634223,
2759
+ "grad_norm": 5.53125,
2760
+ "learning_rate": 2.323813938207593e-05,
2761
+ "loss": 0.4697,
2762
+ "step": 3560
2763
+ },
2764
+ {
2765
+ "epoch": 2.802197802197802,
2766
+ "grad_norm": 5.125,
2767
+ "learning_rate": 2.294393739365621e-05,
2768
+ "loss": 0.5038,
2769
+ "step": 3570
2770
+ },
2771
+ {
2772
+ "epoch": 2.8100470957613815,
2773
+ "grad_norm": 5.21875,
2774
+ "learning_rate": 2.265105407955752e-05,
2775
+ "loss": 0.5708,
2776
+ "step": 3580
2777
+ },
2778
+ {
2779
+ "epoch": 2.817896389324961,
2780
+ "grad_norm": 5.34375,
2781
+ "learning_rate": 2.235950371449938e-05,
2782
+ "loss": 0.5062,
2783
+ "step": 3590
2784
+ },
2785
+ {
2786
+ "epoch": 2.82574568288854,
2787
+ "grad_norm": 6.53125,
2788
+ "learning_rate": 2.2069300508235275e-05,
2789
+ "loss": 0.5216,
2790
+ "step": 3600
2791
+ },
2792
+ {
2793
+ "epoch": 2.8335949764521193,
2794
+ "grad_norm": 4.25,
2795
+ "learning_rate": 2.1780458604860056e-05,
2796
+ "loss": 0.4725,
2797
+ "step": 3610
2798
+ },
2799
+ {
2800
+ "epoch": 2.8414442700156988,
2801
+ "grad_norm": 6.5625,
2802
+ "learning_rate": 2.14929920821206e-05,
2803
+ "loss": 0.5209,
2804
+ "step": 3620
2805
+ },
2806
+ {
2807
+ "epoch": 2.8492935635792778,
2808
+ "grad_norm": 5.8125,
2809
+ "learning_rate": 2.1206914950729672e-05,
2810
+ "loss": 0.4435,
2811
+ "step": 3630
2812
+ },
2813
+ {
2814
+ "epoch": 2.857142857142857,
2815
+ "grad_norm": 5.09375,
2816
+ "learning_rate": 2.0922241153683066e-05,
2817
+ "loss": 0.5311,
2818
+ "step": 3640
2819
+ },
2820
+ {
2821
+ "epoch": 2.8649921507064366,
2822
+ "grad_norm": 5.9375,
2823
+ "learning_rate": 2.0638984565580023e-05,
2824
+ "loss": 0.5394,
2825
+ "step": 3650
2826
+ },
2827
+ {
2828
+ "epoch": 2.8728414442700156,
2829
+ "grad_norm": 5.09375,
2830
+ "learning_rate": 2.035715899194704e-05,
2831
+ "loss": 0.496,
2832
+ "step": 3660
2833
+ },
2834
+ {
2835
+ "epoch": 2.880690737833595,
2836
+ "grad_norm": 5.3125,
2837
+ "learning_rate": 2.007677816856498e-05,
2838
+ "loss": 0.5404,
2839
+ "step": 3670
2840
+ },
2841
+ {
2842
+ "epoch": 2.8885400313971745,
2843
+ "grad_norm": 4.4375,
2844
+ "learning_rate": 1.9797855760799612e-05,
2845
+ "loss": 0.4908,
2846
+ "step": 3680
2847
+ },
2848
+ {
2849
+ "epoch": 2.8963893249607535,
2850
+ "grad_norm": 4.34375,
2851
+ "learning_rate": 1.9520405362935594e-05,
2852
+ "loss": 0.4916,
2853
+ "step": 3690
2854
+ },
2855
+ {
2856
+ "epoch": 2.904238618524333,
2857
+ "grad_norm": 5.1875,
2858
+ "learning_rate": 1.9244440497513893e-05,
2859
+ "loss": 0.4959,
2860
+ "step": 3700
2861
+ },
2862
+ {
2863
+ "epoch": 2.912087912087912,
2864
+ "grad_norm": 5.15625,
2865
+ "learning_rate": 1.896997461467272e-05,
2866
+ "loss": 0.5591,
2867
+ "step": 3710
2868
+ },
2869
+ {
2870
+ "epoch": 2.9199372056514914,
2871
+ "grad_norm": 6.84375,
2872
+ "learning_rate": 1.869702109149199e-05,
2873
+ "loss": 0.4443,
2874
+ "step": 3720
2875
+ },
2876
+ {
2877
+ "epoch": 2.9277864992150704,
2878
+ "grad_norm": 5.9375,
2879
+ "learning_rate": 1.842559323134136e-05,
2880
+ "loss": 0.5501,
2881
+ "step": 3730
2882
+ },
2883
+ {
2884
+ "epoch": 2.93563579277865,
2885
+ "grad_norm": 5.625,
2886
+ "learning_rate": 1.8155704263231776e-05,
2887
+ "loss": 0.5145,
2888
+ "step": 3740
2889
+ },
2890
+ {
2891
+ "epoch": 2.9434850863422293,
2892
+ "grad_norm": 5.53125,
2893
+ "learning_rate": 1.788736734117078e-05,
2894
+ "loss": 0.6076,
2895
+ "step": 3750
2896
+ },
2897
+ {
2898
+ "epoch": 2.9513343799058083,
2899
+ "grad_norm": 4.78125,
2900
+ "learning_rate": 1.762059554352143e-05,
2901
+ "loss": 0.4832,
2902
+ "step": 3760
2903
+ },
2904
+ {
2905
+ "epoch": 2.9591836734693877,
2906
+ "grad_norm": 4.96875,
2907
+ "learning_rate": 1.7355401872364758e-05,
2908
+ "loss": 0.5021,
2909
+ "step": 3770
2910
+ },
2911
+ {
2912
+ "epoch": 2.967032967032967,
2913
+ "grad_norm": 3.703125,
2914
+ "learning_rate": 1.709179925286617e-05,
2915
+ "loss": 0.616,
2916
+ "step": 3780
2917
+ },
2918
+ {
2919
+ "epoch": 2.974882260596546,
2920
+ "grad_norm": 5.78125,
2921
+ "learning_rate": 1.6829800532645447e-05,
2922
+ "loss": 0.4569,
2923
+ "step": 3790
2924
+ },
2925
+ {
2926
+ "epoch": 2.9827315541601256,
2927
+ "grad_norm": 3.515625,
2928
+ "learning_rate": 1.6569418481150595e-05,
2929
+ "loss": 0.464,
2930
+ "step": 3800
2931
+ },
2932
+ {
2933
+ "epoch": 2.990580847723705,
2934
+ "grad_norm": 4.21875,
2935
+ "learning_rate": 1.6310665789035467e-05,
2936
+ "loss": 0.5054,
2937
+ "step": 3810
2938
+ },
2939
+ {
2940
+ "epoch": 2.998430141287284,
2941
+ "grad_norm": 4.3125,
2942
+ "learning_rate": 1.605355506754121e-05,
2943
+ "loss": 0.4882,
2944
+ "step": 3820
2945
+ },
2946
+ {
2947
+ "epoch": 3.0062794348508635,
2948
+ "grad_norm": 5.1875,
2949
+ "learning_rate": 1.5798098847881666e-05,
2950
+ "loss": 0.4937,
2951
+ "step": 3830
2952
+ },
2953
+ {
2954
+ "epoch": 3.014128728414443,
2955
+ "grad_norm": 4.21875,
2956
+ "learning_rate": 1.554430958063259e-05,
2957
+ "loss": 0.4828,
2958
+ "step": 3840
2959
+ },
2960
+ {
2961
+ "epoch": 3.021978021978022,
2962
+ "grad_norm": 5.65625,
2963
+ "learning_rate": 1.529219963512481e-05,
2964
+ "loss": 0.4813,
2965
+ "step": 3850
2966
+ },
2967
+ {
2968
+ "epoch": 3.0298273155416013,
2969
+ "grad_norm": 4.84375,
2970
+ "learning_rate": 1.5041781298841423e-05,
2971
+ "loss": 0.4902,
2972
+ "step": 3860
2973
+ },
2974
+ {
2975
+ "epoch": 3.0376766091051803,
2976
+ "grad_norm": 6.21875,
2977
+ "learning_rate": 1.4793066776818842e-05,
2978
+ "loss": 0.5018,
2979
+ "step": 3870
2980
+ },
2981
+ {
2982
+ "epoch": 3.0455259026687598,
2983
+ "grad_norm": 4.71875,
2984
+ "learning_rate": 1.4546068191051987e-05,
2985
+ "loss": 0.5047,
2986
+ "step": 3880
2987
+ },
2988
+ {
2989
+ "epoch": 3.053375196232339,
2990
+ "grad_norm": 6.1875,
2991
+ "learning_rate": 1.4300797579903474e-05,
2992
+ "loss": 0.5615,
2993
+ "step": 3890
2994
+ },
2995
+ {
2996
+ "epoch": 3.061224489795918,
2997
+ "grad_norm": 4.125,
2998
+ "learning_rate": 1.4057266897516841e-05,
2999
+ "loss": 0.4684,
3000
+ "step": 3900
3001
+ },
3002
+ {
3003
+ "epoch": 3.0690737833594977,
3004
+ "grad_norm": 4.15625,
3005
+ "learning_rate": 1.3815488013233984e-05,
3006
+ "loss": 0.5193,
3007
+ "step": 3910
3008
+ },
3009
+ {
3010
+ "epoch": 3.076923076923077,
3011
+ "grad_norm": 3.78125,
3012
+ "learning_rate": 1.3575472711016635e-05,
3013
+ "loss": 0.4829,
3014
+ "step": 3920
3015
+ },
3016
+ {
3017
+ "epoch": 3.084772370486656,
3018
+ "grad_norm": 5.40625,
3019
+ "learning_rate": 1.3337232688872009e-05,
3020
+ "loss": 0.5745,
3021
+ "step": 3930
3022
+ },
3023
+ {
3024
+ "epoch": 3.0926216640502355,
3025
+ "grad_norm": 7.6875,
3026
+ "learning_rate": 1.3100779558282672e-05,
3027
+ "loss": 0.4937,
3028
+ "step": 3940
3029
+ },
3030
+ {
3031
+ "epoch": 3.100470957613815,
3032
+ "grad_norm": 4.34375,
3033
+ "learning_rate": 1.2866124843640615e-05,
3034
+ "loss": 0.5324,
3035
+ "step": 3950
3036
+ },
3037
+ {
3038
+ "epoch": 3.108320251177394,
3039
+ "grad_norm": 5.9375,
3040
+ "learning_rate": 1.2633279981685608e-05,
3041
+ "loss": 0.5407,
3042
+ "step": 3960
3043
+ },
3044
+ {
3045
+ "epoch": 3.1161695447409734,
3046
+ "grad_norm": 4.96875,
3047
+ "learning_rate": 1.240225632094773e-05,
3048
+ "loss": 0.574,
3049
+ "step": 3970
3050
+ },
3051
+ {
3052
+ "epoch": 3.1240188383045524,
3053
+ "grad_norm": 5.59375,
3054
+ "learning_rate": 1.217306512119425e-05,
3055
+ "loss": 0.5532,
3056
+ "step": 3980
3057
+ },
3058
+ {
3059
+ "epoch": 3.131868131868132,
3060
+ "grad_norm": 4.59375,
3061
+ "learning_rate": 1.194571755288092e-05,
3062
+ "loss": 0.4584,
3063
+ "step": 3990
3064
+ },
3065
+ {
3066
+ "epoch": 3.1397174254317113,
3067
+ "grad_norm": 4.9375,
3068
+ "learning_rate": 1.1720224696607474e-05,
3069
+ "loss": 0.5262,
3070
+ "step": 4000
3071
+ }
3072
+ ],
3073
+ "logging_steps": 10,
3074
+ "max_steps": 5000,
3075
+ "num_input_tokens_seen": 0,
3076
+ "num_train_epochs": 4,
3077
+ "save_steps": 1000,
3078
+ "stateful_callbacks": {
3079
+ "TrainerControl": {
3080
+ "args": {
3081
+ "should_epoch_stop": false,
3082
+ "should_evaluate": false,
3083
+ "should_log": false,
3084
+ "should_save": true,
3085
+ "should_training_stop": false
3086
+ },
3087
+ "attributes": {}
3088
+ }
3089
+ },
3090
+ "total_flos": 0.0,
3091
+ "train_batch_size": 16,
3092
+ "trial_name": null,
3093
+ "trial_params": null
3094
+ }
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-4000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9365177436ad70bb9d6000d96d3ccdc5c5ca38536defdcfec6b548401edbc07f
3
+ size 6481
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/checkpoint-5000/moved ADDED
File without changes
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "gradient_checkpointing": false,
20
+ "hidden_activation": "gelu",
21
+ "hidden_size": 768,
22
+ "initializer_cutoff_factor": 2.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 1152,
25
+ "layer_norm_eps": 1e-05,
26
+ "local_attention": 128,
27
+ "local_rope_theta": 10000.0,
28
+ "max_position_embeddings": 8192,
29
+ "mlp_bias": false,
30
+ "mlp_dropout": 0.0,
31
+ "model_type": "modernbert",
32
+ "norm_bias": false,
33
+ "norm_eps": 1e-05,
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 22,
36
+ "pad_token_id": 50283,
37
+ "position_embedding_type": "absolute",
38
+ "repad_logits_with_grad": false,
39
+ "sep_token_id": 50282,
40
+ "sparse_pred_ignore_index": -100,
41
+ "sparse_prediction": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.53.0",
44
+ "vocab_size": 50368
45
+ }
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:461d9cd6f69078fd798ee95a1eb88776ce5adc5748c48ce1477e6be2d2377af0
3
+ size 298041696
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab613bcea54280917982d51c432e28d3ebba6d4e6765fa600dfbc4db6e1f64d5
3
+ size 596170443
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24c12cceaffe8acec04189ba71dbb81f100dd5cd03b86390253117b3fab2911d
3
+ size 15365
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69adbd9997461473344beb9c44d2e496e24fbc4d6fe69245ab0bd127882efd96
3
+ size 15429
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:baaf7026d9e588ca2ba5b4de8768a379982b7530f39e0b88fa44af9bef8e8bc9
3
+ size 15429
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ec81cbdf599c90b49fb13998f6e46e2492b55345216231b6ce078f88cf04eae
3
+ size 15429
modernbert-two-stage-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.crux-researchy.request/runs/Jan01_23-09-47_nid005111/events.out.tfevents.1767301973.nid005111.95408.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a1266b963e5671de48195e4a46d80d1eb6a4537b58b81ada0e638328d7bf13b
3
+ size 119511