Datasets:

Modalities:
Text
ArXiv:
Maple222 commited on
Commit
53e0dae
·
verified ·
1 Parent(s): 30996b3

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. out/pretrain/ppl_metrics.jsonl +20 -0
  2. out/pretrain/qwen2_7b_question_focus_lr_plus/teelog.txt +229 -0
  3. out/pretrain/tinyllama/teelogs/2407.txt +321 -0
  4. out/pretrain/tinyllama/teelogs/2407_lr4e-5.txt +334 -0
  5. out/pretrain/tinyllama/teelogs/2408.txt +288 -0
  6. out/pretrain/tinyllama/teelogs/2408_full.txt +291 -0
  7. out/pretrain/tinyllama/teelogs/2408_lr4e-5.txt +298 -0
  8. out/pretrain/tinyllama/teelogs/2409.txt +0 -0
  9. out/pretrain/tinyllama/teelogs/2409_full.txt +354 -0
  10. out/pretrain/tinyllama/teelogs/2409_lr4e-5.txt +361 -0
  11. out/pretrain/tinyllama/teelogs/2410.txt +368 -0
  12. out/pretrain/tinyllama/teelogs/2410_full.txt +380 -0
  13. out/pretrain/tinyllama/teelogs/2410_lr4e-5.txt +379 -0
  14. out/pretrain/tinyllama/teelogs/2411.txt +360 -0
  15. out/pretrain/tinyllama/teelogs/2411_full.txt +372 -0
  16. out/pretrain/tinyllama/teelogs/2411_lr4e-5.txt +371 -0
  17. out/pretrain/tinyllama/teelogs/2412.txt +414 -0
  18. out/pretrain/tinyllama/teelogs/2412_full.txt +425 -0
  19. out/pretrain/tinyllama/teelogs/2412_lr4e-5.txt +424 -0
  20. out/pretrain/tinyllama/teelogs/2501.txt +378 -0
  21. out/pretrain/tinyllama/teelogs/2501_full.txt +389 -0
  22. out/pretrain/tinyllama_3_epoch/2407/final/config.json +24 -0
  23. out/pretrain/tinyllama_3_epoch/2407/final/generation_config.json +7 -0
  24. out/pretrain/tinyllama_3_epoch/2407/final/hyperparameters.yaml +44 -0
  25. out/pretrain/tinyllama_3_epoch/2407/final/model_config.yaml +44 -0
  26. out/pretrain/tinyllama_3_epoch/2407/final/tokenizer.json +0 -0
  27. out/pretrain/tinyllama_3_epoch/2407/final/tokenizer_config.json +35 -0
  28. out/pretrain/tinyllama_3_epoch/2408/final/config.json +24 -0
  29. out/pretrain/tinyllama_3_epoch/2408/final/generation_config.json +7 -0
  30. out/pretrain/tinyllama_3_epoch/2408/final/hyperparameters.yaml +44 -0
  31. out/pretrain/tinyllama_3_epoch/2408/final/model_config.yaml +44 -0
  32. out/pretrain/tinyllama_3_epoch/2408/final/tokenizer.json +0 -0
  33. out/pretrain/tinyllama_3_epoch/2408/final/tokenizer_config.json +35 -0
  34. out/pretrain/tinyllama_3_epoch/2409/final/tokenizer.json +0 -0
  35. out/pretrain/tinyllama_3_epoch/2409/final/tokenizer_config.json +35 -0
  36. out/pretrain/tinyllama_lr_plus/2501/final/config.json +24 -0
  37. out/pretrain/tinyllama_lr_plus/2502/final/config.json +24 -0
  38. out/pretrain/tinyllama_lr_plus/2502/final/generation_config.json +7 -0
  39. out/pretrain/tinyllama_lr_plus/2502/final/hyperparameters.yaml +44 -0
  40. out/pretrain/tinyllama_lr_plus/2502/final/model_config.yaml +44 -0
  41. out/pretrain/tinyllama_lr_plus/2502/final/tokenizer.json +0 -0
  42. out/pretrain/tinyllama_lr_plus/2502/final/tokenizer_config.json +35 -0
  43. out/pretrain/tinyllama_lr_plus/2503/final/config.json +24 -0
  44. out/pretrain/tinyllama_lr_plus/2503/final/generation_config.json +7 -0
  45. out/pretrain/tinyllama_lr_plus/2503/final/hyperparameters.yaml +44 -0
  46. out/pretrain/tinyllama_lr_plus/2503/final/model_config.yaml +44 -0
  47. out/pretrain/tinyllama_lr_plus/2503/final/tokenizer.json +0 -0
  48. out/pretrain/tinyllama_lr_plus/2503/final/tokenizer_config.json +35 -0
  49. out/pretrain/tinyllama_lr_plus/2504/final/config.json +24 -0
  50. out/pretrain/tinyllama_lr_plus/2504/final/generation_config.json +7 -0
out/pretrain/ppl_metrics.jsonl ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"val_loss": 1.2241098880767822, "val_ppl": 3.4011373421309727}
2
+ {"val_loss": 1.4328051805496216, "val_ppl": 4.190437654174611}
3
+ {"val_loss": 1.2608414888381958, "val_ppl": 3.528389338728705}
4
+ {"val_loss": 1.3117088079452515, "val_ppl": 3.7125122654387708}
5
+ {"val_loss": 1.3475964069366455, "val_ppl": 3.848164983201228}
6
+ {"val_loss": 1.3134140968322754, "val_ppl": 3.718848572429408}
7
+ {"val_loss": 1.249189853668213, "val_ppl": 3.4875164140296686}
8
+ {"val_loss": 1.334436058998108, "val_ppl": 3.797853577039805}
9
+ {"val_loss": 1.3168798685073853, "val_ppl": 3.731759612911466}
10
+ {"val_loss": 1.3145214319229126, "val_ppl": 3.722968864801566}
11
+ {"val_loss": 1.2239998579025269, "val_ppl": 3.400763134983968}
12
+ {"val_loss": 1.334436058998108, "val_ppl": 3.797853577039805}
13
+ {"val_loss": 1.334533929824829, "val_ppl": 3.798225294298997}
14
+ {"val_loss": 1.4382433891296387, "val_ppl": 4.213288204894694}
15
+ {"val_loss": 1.3327378034591675, "val_ppl": 3.7914093247089276}
16
+ {"val_loss": 0.47974008321762085, "val_ppl": 1.615654411917815}
17
+ {"val_loss": 1.4382433891296387, "val_ppl": 4.213288204894694}
18
+ {"val_loss": 1.3327378034591675, "val_ppl": 3.7914093247089276}
19
+ {"val_loss": 1.334533929824829, "val_ppl": 3.798225294298997}
20
+ {"val_loss": 0.47974008321762085, "val_ppl": 1.615654411917815}
out/pretrain/qwen2_7b_question_focus_lr_plus/teelog.txt ADDED
@@ -0,0 +1,229 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Initializing distributed: GLOBAL_RANK: 0, MEMBER: 1/4
2
+ Initializing distributed: GLOBAL_RANK: 1, MEMBER: 2/4
3
+ [rank: 1] Seed set to 42
4
+ Initializing distributed: GLOBAL_RANK: 3, MEMBER: 4/4
5
+ [rank: 3] Seed set to 42
6
+ Initializing distributed: GLOBAL_RANK: 2, MEMBER: 3/4
7
+ [rank: 2] Seed set to 42
8
+ ----------------------------------------------------------------------------------------------------
9
+ distributed_backend=nccl
10
+ All distributed processes registered. Starting with 4 processes
11
+ ----------------------------------------------------------------------------------------------------
12
+
13
+ [rank: 0] Seed set to 42
14
+ /mnt/data/litgpt/litgpt/pretrain.py:495: UserWarning: A newer version of litdata is available (0.2.58). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
15
+ train_dataloader = data.train_dataloader()
16
+ /mnt/data/litgpt/litgpt/pretrain.py:495: UserWarning: A newer version of litdata is available (0.2.58). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
17
+ train_dataloader = data.train_dataloader()
18
+ /mnt/data/litgpt/litgpt/pretrain.py:495: UserWarning: A newer version of litdata is available (0.2.58). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
19
+ train_dataloader = data.train_dataloader()
20
+ All GPUs are fully connected via NVLink.
21
+ {'data': {'batch_size': 1,
22
+ 'data_path': PosixPath('litgpt/data/arxiv'),
23
+ 'num_workers': 0,
24
+ 'ppl': False,
25
+ 'seed': 42,
26
+ 'seq_length': 1024,
27
+ 'use_starcoder': True},
28
+ 'data_dir': PosixPath('litgpt/data/arxiv_test_qwen2_tokenized'),
29
+ 'devices': 'auto',
30
+ 'eval': {'evaluate_example': 'first',
31
+ 'final_validation': False,
32
+ 'initial_validation': False,
33
+ 'interval': 9999,
34
+ 'max_iters': 100,
35
+ 'max_new_tokens': None},
36
+ 'initial_checkpoint_dir': PosixPath('checkpoints/Qwen/Qwen2-7B'),
37
+ 'log': {'group': None, 'project': None, 'run': None},
38
+ 'logger_name': 'tensorboard',
39
+ 'model_config': {'attention_logit_softcapping': None,
40
+ 'attention_scores_scalar': None,
41
+ 'attn_bias': True,
42
+ 'bias': False,
43
+ 'block_size': 131072,
44
+ 'final_logit_softcapping': None,
45
+ 'gelu_approximate': 'none',
46
+ 'head_size': 128,
47
+ 'hf_config': {'name': 'Qwen2-7B', 'org': 'Qwen'},
48
+ 'intermediate_size': 18944,
49
+ 'lm_head_bias': False,
50
+ 'mlp_class_name': 'LLaMAMLP',
51
+ 'moe_intermediate_size': None,
52
+ 'n_embd': 3584,
53
+ 'n_expert': 0,
54
+ 'n_expert_per_token': 0,
55
+ 'n_head': 28,
56
+ 'n_layer': 28,
57
+ 'n_query_groups': 4,
58
+ 'name': 'Qwen2-7B',
59
+ 'norm_1': True,
60
+ 'norm_2': True,
61
+ 'norm_class_name': 'RMSNorm',
62
+ 'norm_eps': 1e-06,
63
+ 'norm_qk': False,
64
+ 'norm_qk_type': 'default',
65
+ 'padded_vocab_size': 152064,
66
+ 'padding_multiple': 512,
67
+ 'parallel_residual': False,
68
+ 'post_attention_norm': False,
69
+ 'post_mlp_norm': False,
70
+ 'rope_adjustments': None,
71
+ 'rope_base': 1000000,
72
+ 'rope_condense_ratio': 1,
73
+ 'rope_indices': None,
74
+ 'rope_local_base_freq': None,
75
+ 'rotary_percentage': 1.0,
76
+ 'scale_embeddings': False,
77
+ 'shared_attention_norm': False,
78
+ 'sliding_window_indices': None,
79
+ 'sliding_window_size': None,
80
+ 'vocab_size': 151643},
81
+ 'model_name': 'Qwen2-7B',
82
+ 'num_nodes': 1,
83
+ 'optimizer': "{'class_path': 'torch.optim.AdamW', 'init_args': {'lr': 8e-05, "
84
+ "'weight_decay': 0.1, 'betas': [0.9, 0.95]}}",
85
+ 'out_dir': PosixPath('out/pretrain/qwen2_7b_question_focus_lr_plus'),
86
+ 'precision': 'bf16-true',
87
+ 'resume': False,
88
+ 'seed': 42,
89
+ 'tokenizer_dir': PosixPath('checkpoints/Qwen/Qwen2-7B'),
90
+ 'train': {'epochs': None,
91
+ 'global_batch_size': 512,
92
+ 'log_interval': 1,
93
+ 'lr_warmup_fraction': None,
94
+ 'lr_warmup_steps': 0,
95
+ 'max_norm': 1.0,
96
+ 'max_seq_length': 1024,
97
+ 'max_steps': None,
98
+ 'max_tokens': 52428800,
99
+ 'micro_batch_size': 4,
100
+ 'min_lr': 8e-05,
101
+ 'save_interval': 10,
102
+ 'tie_embeddings': None}}
103
+ Time to instantiate model: 0.06 seconds.
104
+ Total parameters: 7,615,616,512
105
+ [ok] checkpoints/Qwen/Qwen2-7B/lit_model.pth 已是纯权重
106
+ /mnt/data/litgpt/litgpt/pretrain.py:495: UserWarning: A newer version of litdata is available (0.2.58). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
107
+ train_dataloader = data.train_dataloader()
108
+ Verifying settings ...
109
+ Measured TFLOPs: 419979.36
110
+ Epoch 2 | iter 32 step 1 | loss train: 1.784, val: n/a | iter time: 1130.43 ms (step) remaining time: 0:49:35
111
+ Epoch 4 | iter 64 step 2 | loss train: 3.834, val: n/a | iter time: 990.00 ms (step) remaining time: 0:48:32
112
+ Epoch 6 | iter 96 step 3 | loss train: 2.444, val: n/a | iter time: 987.93 ms (step) remaining time: 0:47:53
113
+ Epoch 7 | iter 128 step 4 | loss train: 1.904, val: n/a | iter time: 992.12 ms (step) remaining time: 0:47:20
114
+ Epoch 9 | iter 160 step 5 | loss train: 1.595, val: n/a | iter time: 991.60 ms (step) remaining time: 0:46:49
115
+ Epoch 11 | iter 192 step 6 | loss train: 1.370, val: n/a | iter time: 991.67 ms (step) remaining time: 0:46:19
116
+ Epoch 12 | iter 224 step 7 | loss train: 1.167, val: n/a | iter time: 989.51 ms (step) remaining time: 0:45:48
117
+ Epoch 14 | iter 256 step 8 | loss train: 0.951, val: n/a | iter time: 992.01 ms (step) remaining time: 0:45:18
118
+ Epoch 16 | iter 288 step 9 | loss train: 0.720, val: n/a | iter time: 993.83 ms (step) remaining time: 0:44:48
119
+ Epoch 17 | iter 320 step 10 | loss train: 0.543, val: n/a | iter time: 992.87 ms (step) remaining time: 0:44:19
120
+ Saving checkpoint to 'out/pretrain/qwen2_7b_question_focus_lr_plus/step-00000010/lit_model.pth'
121
+ Epoch 19 | iter 352 step 11 | loss train: 0.338, val: n/a | iter time: 986.02 ms (step) remaining time: 1:20:01
122
+ Epoch 21 | iter 384 step 12 | loss train: 0.221, val: n/a | iter time: 988.10 ms (step) remaining time: 1:16:07
123
+ Epoch 22 | iter 416 step 13 | loss train: 0.146, val: n/a | iter time: 991.41 ms (step) remaining time: 1:12:45
124
+ Epoch 24 | iter 448 step 14 | loss train: 0.107, val: n/a | iter time: 992.48 ms (step) remaining time: 1:09:48
125
+ Epoch 26 | iter 480 step 15 | loss train: 0.100, val: n/a | iter time: 991.17 ms (step) remaining time: 1:07:10
126
+ Epoch 27 | iter 512 step 16 | loss train: 0.068, val: n/a | iter time: 992.15 ms (step) remaining time: 1:04:49
127
+ Epoch 29 | iter 544 step 17 | loss train: 0.048, val: n/a | iter time: 991.18 ms (step) remaining time: 1:02:40
128
+ Epoch 31 | iter 576 step 18 | loss train: 0.036, val: n/a | iter time: 988.18 ms (step) remaining time: 1:00:46
129
+ Epoch 32 | iter 608 step 19 | loss train: 0.027, val: n/a | iter time: 989.82 ms (step) remaining time: 0:58:58
130
+ Epoch 34 | iter 640 step 20 | loss train: 0.020, val: n/a | iter time: 989.76 ms (step) remaining time: 0:57:18
131
+ Saving checkpoint to 'out/pretrain/qwen2_7b_question_focus_lr_plus/step-00000020/lit_model.pth'
132
+ Epoch 36 | iter 672 step 21 | loss train: 0.013, val: n/a | iter time: 988.12 ms (step) remaining time: 0:59:14
133
+ Epoch 38 | iter 704 step 22 | loss train: 0.008, val: n/a | iter time: 988.99 ms (step) remaining time: 0:57:33
134
+ Epoch 39 | iter 736 step 23 | loss train: 0.007, val: n/a | iter time: 989.45 ms (step) remaining time: 0:56:00
135
+ Epoch 41 | iter 768 step 24 | loss train: 0.006, val: n/a | iter time: 992.48 ms (step) remaining time: 0:54:31
136
+ Epoch 43 | iter 800 step 25 | loss train: 0.006, val: n/a | iter time: 989.03 ms (step) remaining time: 0:53:07
137
+ Epoch 44 | iter 832 step 26 | loss train: 0.005, val: n/a | iter time: 990.26 ms (step) remaining time: 0:51:47
138
+ Epoch 46 | iter 864 step 27 | loss train: 0.005, val: n/a | iter time: 990.10 ms (step) remaining time: 0:50:32
139
+ Epoch 48 | iter 896 step 28 | loss train: 0.005, val: n/a | iter time: 990.45 ms (step) remaining time: 0:49:19
140
+ Epoch 49 | iter 928 step 29 | loss train: 0.005, val: n/a | iter time: 989.48 ms (step) remaining time: 0:48:09
141
+ Epoch 51 | iter 960 step 30 | loss train: 0.004, val: n/a | iter time: 989.24 ms (step) remaining time: 0:47:03
142
+ Saving checkpoint to 'out/pretrain/qwen2_7b_question_focus_lr_plus/step-00000030/lit_model.pth'
143
+ Epoch 53 | iter 992 step 31 | loss train: 0.005, val: n/a | iter time: 986.57 ms (step) remaining time: 0:48:01
144
+ Epoch 54 | iter 1024 step 32 | loss train: 0.004, val: n/a | iter time: 990.88 ms (step) remaining time: 0:46:53
145
+ Epoch 56 | iter 1056 step 33 | loss train: 0.004, val: n/a | iter time: 990.30 ms (step) remaining time: 0:45:48
146
+ Epoch 58 | iter 1088 step 34 | loss train: 0.005, val: n/a | iter time: 990.78 ms (step) remaining time: 0:44:44
147
+ Epoch 59 | iter 1120 step 35 | loss train: 0.004, val: n/a | iter time: 988.25 ms (step) remaining time: 0:43:43
148
+ Epoch 61 | iter 1152 step 36 | loss train: 0.003, val: n/a | iter time: 994.09 ms (step) remaining time: 0:42:43
149
+ Epoch 63 | iter 1184 step 37 | loss train: 0.003, val: n/a | iter time: 990.38 ms (step) remaining time: 0:41:45
150
+ Epoch 64 | iter 1216 step 38 | loss train: 0.003, val: n/a | iter time: 991.09 ms (step) remaining time: 0:40:48
151
+ Epoch 66 | iter 1248 step 39 | loss train: 0.003, val: n/a | iter time: 990.35 ms (step) remaining time: 0:39:53
152
+ Epoch 68 | iter 1280 step 40 | loss train: 0.003, val: n/a | iter time: 990.50 ms (step) remaining time: 0:38:59
153
+ Saving checkpoint to 'out/pretrain/qwen2_7b_question_focus_lr_plus/step-00000040/lit_model.pth'
154
+ Epoch 70 | iter 1312 step 41 | loss train: 0.002, val: n/a | iter time: 985.04 ms (step) remaining time: 0:39:28
155
+ Epoch 71 | iter 1344 step 42 | loss train: 0.002, val: n/a | iter time: 988.08 ms (step) remaining time: 0:38:33
156
+ Epoch 73 | iter 1376 step 43 | loss train: 0.002, val: n/a | iter time: 989.22 ms (step) remaining time: 0:37:39
157
+ Epoch 75 | iter 1408 step 44 | loss train: 0.002, val: n/a | iter time: 988.61 ms (step) remaining time: 0:36:47
158
+ Epoch 76 | iter 1440 step 45 | loss train: 0.002, val: n/a | iter time: 988.63 ms (step) remaining time: 0:35:56
159
+ Epoch 78 | iter 1472 step 46 | loss train: 0.002, val: n/a | iter time: 991.07 ms (step) remaining time: 0:35:05
160
+ Epoch 80 | iter 1504 step 47 | loss train: 0.002, val: n/a | iter time: 993.71 ms (step) remaining time: 0:34:16
161
+ Epoch 81 | iter 1536 step 48 | loss train: 0.002, val: n/a | iter time: 994.17 ms (step) remaining time: 0:33:27
162
+ Epoch 83 | iter 1568 step 49 | loss train: 0.002, val: n/a | iter time: 992.13 ms (step) remaining time: 0:32:39
163
+ Epoch 85 | iter 1600 step 50 | loss train: 0.001, val: n/a | iter time: 989.11 ms (step) remaining time: 0:31:52
164
+ Saving checkpoint to 'out/pretrain/qwen2_7b_question_focus_lr_plus/step-00000050/lit_model.pth'
165
+ Epoch 86 | iter 1632 step 51 | loss train: 0.001, val: n/a | iter time: 987.12 ms (step) remaining time: 0:31:57
166
+ Epoch 88 | iter 1664 step 52 | loss train: 0.001, val: n/a | iter time: 988.33 ms (step) remaining time: 0:31:09
167
+ Epoch 90 | iter 1696 step 53 | loss train: 0.001, val: n/a | iter time: 988.65 ms (step) remaining time: 0:30:21
168
+ Epoch 91 | iter 1728 step 54 | loss train: 0.001, val: n/a | iter time: 990.42 ms (step) remaining time: 0:29:35
169
+ Epoch 93 | iter 1760 step 55 | loss train: 0.001, val: n/a | iter time: 991.74 ms (step) remaining time: 0:28:49
170
+ Epoch 95 | iter 1792 step 56 | loss train: 0.001, val: n/a | iter time: 992.71 ms (step) remaining time: 0:28:03
171
+ Epoch 96 | iter 1824 step 57 | loss train: 0.001, val: n/a | iter time: 990.15 ms (step) remaining time: 0:27:18
172
+ Epoch 98 | iter 1856 step 58 | loss train: 0.001, val: n/a | iter time: 988.80 ms (step) remaining time: 0:26:34
173
+ Epoch 100 | iter 1888 step 59 | loss train: 0.001, val: n/a | iter time: 992.54 ms (step) remaining time: 0:25:50
174
+ Epoch 102 | iter 1920 step 60 | loss train: 0.001, val: n/a | iter time: 987.75 ms (step) remaining time: 0:25:07
175
+ Saving checkpoint to 'out/pretrain/qwen2_7b_question_focus_lr_plus/step-00000060/lit_model.pth'
176
+ Epoch 103 | iter 1952 step 61 | loss train: 0.001, val: n/a | iter time: 990.43 ms (step) remaining time: 0:25:00
177
+ Epoch 105 | iter 1984 step 62 | loss train: 0.001, val: n/a | iter time: 987.48 ms (step) remaining time: 0:24:16
178
+ Epoch 107 | iter 2016 step 63 | loss train: 0.001, val: n/a | iter time: 989.32 ms (step) remaining time: 0:23:32
179
+ Epoch 108 | iter 2048 step 64 | loss train: 0.001, val: n/a | iter time: 988.99 ms (step) remaining time: 0:22:49
180
+ Epoch 110 | iter 2080 step 65 | loss train: 0.001, val: n/a | iter time: 994.41 ms (step) remaining time: 0:22:07
181
+ Epoch 112 | iter 2112 step 66 | loss train: 0.001, val: n/a | iter time: 992.36 ms (step) remaining time: 0:21:24
182
+ Epoch 113 | iter 2144 step 67 | loss train: 0.001, val: n/a | iter time: 990.94 ms (step) remaining time: 0:20:43
183
+ Epoch 115 | iter 2176 step 68 | loss train: 0.001, val: n/a | iter time: 990.64 ms (step) remaining time: 0:20:01
184
+ Epoch 117 | iter 2208 step 69 | loss train: 0.001, val: n/a | iter time: 993.15 ms (step) remaining time: 0:19:20
185
+ Epoch 118 | iter 2240 step 70 | loss train: 0.001, val: n/a | iter time: 992.78 ms (step) remaining time: 0:18:39
186
+ Saving checkpoint to 'out/pretrain/qwen2_7b_question_focus_lr_plus/step-00000070/lit_model.pth'
187
+ Epoch 120 | iter 2272 step 71 | loss train: 0.001, val: n/a | iter time: 987.65 ms (step) remaining time: 0:18:22
188
+ Epoch 122 | iter 2304 step 72 | loss train: 0.001, val: n/a | iter time: 989.30 ms (step) remaining time: 0:17:41
189
+ Epoch 123 | iter 2336 step 73 | loss train: 0.001, val: n/a | iter time: 988.89 ms (step) remaining time: 0:17:00
190
+ Epoch 125 | iter 2368 step 74 | loss train: 0.001, val: n/a | iter time: 995.48 ms (step) remaining time: 0:16:19
191
+ Epoch 127 | iter 2400 step 75 | loss train: 0.001, val: n/a | iter time: 992.03 ms (step) remaining time: 0:15:39
192
+ Epoch 128 | iter 2432 step 76 | loss train: 0.001, val: n/a | iter time: 990.07 ms (step) remaining time: 0:14:59
193
+ Epoch 130 | iter 2464 step 77 | loss train: 0.001, val: n/a | iter time: 988.70 ms (step) remaining time: 0:14:19
194
+ Epoch 132 | iter 2496 step 78 | loss train: 0.001, val: n/a | iter time: 994.75 ms (step) remaining time: 0:13:39
195
+ Epoch 134 | iter 2528 step 79 | loss train: 0.001, val: n/a | iter time: 991.65 ms (step) remaining time: 0:13:00
196
+ Epoch 135 | iter 2560 step 80 | loss train: 0.001, val: n/a | iter time: 994.05 ms (step) remaining time: 0:12:21
197
+ Saving checkpoint to 'out/pretrain/qwen2_7b_question_focus_lr_plus/step-00000080/lit_model.pth'
198
+ Epoch 137 | iter 2592 step 81 | loss train: 0.001, val: n/a | iter time: 989.12 ms (step) remaining time: 0:11:55
199
+ Epoch 139 | iter 2624 step 82 | loss train: 0.001, val: n/a | iter time: 992.55 ms (step) remaining time: 0:11:15
200
+ Epoch 140 | iter 2656 step 83 | loss train: 0.001, val: n/a | iter time: 993.71 ms (step) remaining time: 0:10:36
201
+ Epoch 142 | iter 2688 step 84 | loss train: 0.001, val: n/a | iter time: 987.33 ms (step) remaining time: 0:09:57
202
+ Epoch 144 | iter 2720 step 85 | loss train: 0.001, val: n/a | iter time: 992.72 ms (step) remaining time: 0:09:18
203
+ Epoch 145 | iter 2752 step 86 | loss train: 0.001, val: n/a | iter time: 992.75 ms (step) remaining time: 0:08:40
204
+ Epoch 147 | iter 2784 step 87 | loss train: 0.001, val: n/a | iter time: 990.86 ms (step) remaining time: 0:08:02
205
+ Epoch 149 | iter 2816 step 88 | loss train: 0.001, val: n/a | iter time: 992.41 ms (step) remaining time: 0:07:23
206
+ Epoch 150 | iter 2848 step 89 | loss train: 0.001, val: n/a | iter time: 996.79 ms (step) remaining time: 0:06:46
207
+ Epoch 152 | iter 2880 step 90 | loss train: 0.001, val: n/a | iter time: 987.57 ms (step) remaining time: 0:06:08
208
+ Saving checkpoint to 'out/pretrain/qwen2_7b_question_focus_lr_plus/step-00000090/lit_model.pth'
209
+ Epoch 154 | iter 2912 step 91 | loss train: 0.001, val: n/a | iter time: 988.50 ms (step) remaining time: 0:05:36
210
+ Epoch 155 | iter 2944 step 92 | loss train: 0.001, val: n/a | iter time: 989.06 ms (step) remaining time: 0:04:57
211
+ Epoch 157 | iter 2976 step 93 | loss train: 0.001, val: n/a | iter time: 994.86 ms (step) remaining time: 0:04:20
212
+ Epoch 159 | iter 3008 step 94 | loss train: 0.001, val: n/a | iter time: 990.57 ms (step) remaining time: 0:03:42
213
+ Epoch 160 | iter 3040 step 95 | loss train: 0.001, val: n/a | iter time: 993.62 ms (step) remaining time: 0:03:05
214
+ Epoch 162 | iter 3072 step 96 | loss train: 0.001, val: n/a | iter time: 990.85 ms (step) remaining time: 0:02:27
215
+ Epoch 164 | iter 3104 step 97 | loss train: 0.001, val: n/a | iter time: 992.47 ms (step) remaining time: 0:01:50
216
+ Epoch 166 | iter 3136 step 98 | loss train: 0.001, val: n/a | iter time: 989.93 ms (step) remaining time: 0:01:13
217
+ Epoch 167 | iter 3168 step 99 | loss train: 0.001, val: n/a | iter time: 990.92 ms (step) remaining time: 0:00:36
218
+ Epoch 169 | iter 3200 step 100 | loss train: 0.001, val: n/a | iter time: 992.26 ms (step) remaining time: 0:00:00
219
+ Saving checkpoint to 'out/pretrain/qwen2_7b_question_focus_lr_plus/step-00000100/lit_model.pth'
220
+ Saving checkpoint to 'out/pretrain/qwen2_7b_question_focus_lr_plus/final/lit_model.pth'
221
+ ----------------------------------------
222
+ | Performance
223
+ | - Total tokens : 52,428,800
224
+ | - Training Time : 4129.90 s
225
+ | - Tok/sec : 6.05 tok/s
226
+ | ----------------------------------------
227
+ | Memory Usage
228
+ | - Memory Used : 56.04 GB
229
+ ----------------------------------------
out/pretrain/tinyllama/teelogs/2407.txt ADDED
@@ -0,0 +1,321 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Initializing distributed: GLOBAL_RANK: 0, MEMBER: 1/4
2
+ Initializing distributed: GLOBAL_RANK: 2, MEMBER: 3/4
3
+ [rank: 2] Seed set to 42
4
+ Initializing distributed: GLOBAL_RANK: 1, MEMBER: 2/4
5
+ Initializing distributed: GLOBAL_RANK: 3, MEMBER: 4/4
6
+ [rank: 3] Seed set to 42
7
+ ----------------------------------------------------------------------------------------------------
8
+ distributed_backend=nccl
9
+ All distributed processes registered. Starting with 4 processes
10
+ ----------------------------------------------------------------------------------------------------
11
+
12
+ [rank: 1] Seed set to 42
13
+ [rank: 0] Seed set to 42
14
+ All GPUs are fully connected via NVLink.
15
+ {'data': {'batch_size': 1,
16
+ 'data_path': PosixPath('litgpt/data/arxiv'),
17
+ 'num_workers': 8,
18
+ 'seed': 42,
19
+ 'seq_length': 2048,
20
+ 'use_starcoder': True},
21
+ 'devices': 'auto',
22
+ 'eval': {'evaluate_example': 'first',
23
+ 'final_validation': True,
24
+ 'initial_validation': True,
25
+ 'interval': 50,
26
+ 'max_iters': 100,
27
+ 'max_new_tokens': None},
28
+ 'initial_checkpoint_dir': PosixPath('checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T'),
29
+ 'log': {'group': None, 'project': None, 'run': None},
30
+ 'logger_name': 'tensorboard',
31
+ 'model_config': {'attention_logit_softcapping': None,
32
+ 'attention_scores_scalar': None,
33
+ 'attn_bias': False,
34
+ 'bias': False,
35
+ 'block_size': 2048,
36
+ 'final_logit_softcapping': None,
37
+ 'gelu_approximate': 'none',
38
+ 'head_size': 64,
39
+ 'hf_config': {'name': 'TinyLlama-1.1B-intermediate-step-1431k-3T',
40
+ 'org': 'TinyLlama'},
41
+ 'intermediate_size': 5632,
42
+ 'lm_head_bias': False,
43
+ 'mlp_class_name': 'LLaMAMLP',
44
+ 'moe_intermediate_size': None,
45
+ 'n_embd': 2048,
46
+ 'n_expert': 0,
47
+ 'n_expert_per_token': 0,
48
+ 'n_head': 32,
49
+ 'n_layer': 22,
50
+ 'n_query_groups': 4,
51
+ 'name': 'tiny-llama-1.1b',
52
+ 'norm_1': True,
53
+ 'norm_2': True,
54
+ 'norm_class_name': 'RMSNorm',
55
+ 'norm_eps': 1e-05,
56
+ 'norm_qk': False,
57
+ 'norm_qk_type': 'default',
58
+ 'padded_vocab_size': 32000,
59
+ 'padding_multiple': 64,
60
+ 'parallel_residual': False,
61
+ 'post_attention_norm': False,
62
+ 'post_mlp_norm': False,
63
+ 'rope_adjustments': None,
64
+ 'rope_base': 10000,
65
+ 'rope_condense_ratio': 1,
66
+ 'rope_indices': None,
67
+ 'rope_local_base_freq': None,
68
+ 'rotary_percentage': 1.0,
69
+ 'scale_embeddings': False,
70
+ 'shared_attention_norm': False,
71
+ 'sliding_window_indices': None,
72
+ 'sliding_window_size': None,
73
+ 'vocab_size': 32000},
74
+ 'model_name': 'tiny-llama-1.1b',
75
+ 'num_nodes': 1,
76
+ 'optimizer': "{'class_path': 'torch.optim.AdamW', 'init_args': {'lr': 0.0004, "
77
+ "'weight_decay': 0.1, 'betas': [0.9, 0.95]}}",
78
+ 'out_dir': PosixPath('out/pretrain/tiny-llama-cl-2407'),
79
+ 'precision': 'bf16-mixed',
80
+ 'resume': False,
81
+ 'seed': 42,
82
+ 'tokenizer_dir': PosixPath('checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T'),
83
+ 'train': {'epochs': None,
84
+ 'global_batch_size': 512,
85
+ 'log_interval': 1,
86
+ 'lr_warmup_fraction': None,
87
+ 'lr_warmup_steps': 20,
88
+ 'max_norm': 1.0,
89
+ 'max_seq_length': 2048,
90
+ 'max_steps': None,
91
+ 'max_tokens': 209715200,
92
+ 'micro_batch_size': 4,
93
+ 'min_lr': 4e-05,
94
+ 'save_interval': 100,
95
+ 'tie_embeddings': None}}
96
+ Time to instantiate model: 0.02 seconds.
97
+ Total parameters: 1,100,048,384
98
+ Validating ...
99
+ Measured TFLOPs: 239.66
100
+ Epoch 1 | iter 32 step 1 | loss train: 1.482, val: 1.516 | iter time: 539.29 ms (step) remaining time: 0:39:18
101
+ Epoch 1 | iter 64 step 2 | loss train: 1.503, val: 1.516 | iter time: 357.93 ms (step) remaining time: 0:37:16
102
+ Epoch 1 | iter 96 step 3 | loss train: 1.500, val: 1.516 | iter time: 357.13 ms (step) remaining time: 0:36:30
103
+ Epoch 1 | iter 128 step 4 | loss train: 1.515, val: 1.516 | iter time: 357.97 ms (step) remaining time: 0:36:03
104
+ Epoch 1 | iter 160 step 5 | loss train: 1.564, val: 1.516 | iter time: 359.36 ms (step) remaining time: 0:35:43
105
+ Epoch 1 | iter 192 step 6 | loss train: 1.819, val: 1.516 | iter time: 359.46 ms (step) remaining time: 0:35:27
106
+ Epoch 1 | iter 224 step 7 | loss train: 1.808, val: 1.516 | iter time: 357.37 ms (step) remaining time: 0:35:12
107
+ Epoch 1 | iter 256 step 8 | loss train: 1.692, val: 1.516 | iter time: 358.38 ms (step) remaining time: 0:34:58
108
+ Epoch 1 | iter 288 step 9 | loss train: 1.620, val: 1.516 | iter time: 359.95 ms (step) remaining time: 0:34:45
109
+ Epoch 1 | iter 320 step 10 | loss train: 1.598, val: 1.516 | iter time: 359.83 ms (step) remaining time: 0:34:33
110
+ Epoch 1 | iter 352 step 11 | loss train: 1.594, val: 1.516 | iter time: 359.85 ms (step) remaining time: 0:34:21
111
+ Epoch 1 | iter 384 step 12 | loss train: 1.536, val: 1.516 | iter time: 359.22 ms (step) remaining time: 0:34:09
112
+ Epoch 1 | iter 416 step 13 | loss train: 1.553, val: 1.516 | iter time: 359.10 ms (step) remaining time: 0:33:57
113
+ Epoch 1 | iter 448 step 14 | loss train: 1.647, val: 1.516 | iter time: 359.76 ms (step) remaining time: 0:33:46
114
+ Epoch 1 | iter 480 step 15 | loss train: 1.564, val: 1.516 | iter time: 357.45 ms (step) remaining time: 0:33:34
115
+ Epoch 1 | iter 512 step 16 | loss train: 1.564, val: 1.516 | iter time: 358.71 ms (step) remaining time: 0:33:23
116
+ Epoch 1 | iter 544 step 17 | loss train: 1.575, val: 1.516 | iter time: 359.15 ms (step) remaining time: 0:33:12
117
+ Epoch 1 | iter 576 step 18 | loss train: 1.584, val: 1.516 | iter time: 358.31 ms (step) remaining time: 0:33:00
118
+ Epoch 1 | iter 608 step 19 | loss train: 1.653, val: 1.516 | iter time: 359.31 ms (step) remaining time: 0:32:49
119
+ Epoch 1 | iter 640 step 20 | loss train: 1.619, val: 1.516 | iter time: 359.90 ms (step) remaining time: 0:32:38
120
+ Epoch 1 | iter 672 step 21 | loss train: 1.645, val: 1.516 | iter time: 361.40 ms (step) remaining time: 0:32:27
121
+ Epoch 1 | iter 704 step 22 | loss train: 1.527, val: 1.516 | iter time: 359.42 ms (step) remaining time: 0:32:16
122
+ Epoch 1 | iter 736 step 23 | loss train: 1.632, val: 1.516 | iter time: 359.27 ms (step) remaining time: 0:32:05
123
+ Epoch 1 | iter 768 step 24 | loss train: 1.552, val: 1.516 | iter time: 359.93 ms (step) remaining time: 0:31:54
124
+ Epoch 1 | iter 800 step 25 | loss train: 1.586, val: 1.516 | iter time: 360.68 ms (step) remaining time: 0:31:43
125
+ Epoch 1 | iter 832 step 26 | loss train: 1.574, val: 1.516 | iter time: 359.89 ms (step) remaining time: 0:31:32
126
+ Epoch 1 | iter 864 step 27 | loss train: 1.554, val: 1.516 | iter time: 358.49 ms (step) remaining time: 0:31:21
127
+ Epoch 1 | iter 896 step 28 | loss train: 1.662, val: 1.516 | iter time: 360.49 ms (step) remaining time: 0:31:10
128
+ Epoch 1 | iter 928 step 29 | loss train: 1.539, val: 1.516 | iter time: 357.63 ms (step) remaining time: 0:30:59
129
+ Epoch 1 | iter 960 step 30 | loss train: 1.608, val: 1.516 | iter time: 359.51 ms (step) remaining time: 0:30:48
130
+ Epoch 1 | iter 992 step 31 | loss train: 1.547, val: 1.516 | iter time: 360.57 ms (step) remaining time: 0:30:37
131
+ Epoch 1 | iter 1024 step 32 | loss train: 1.619, val: 1.516 | iter time: 359.55 ms (step) remaining time: 0:30:26
132
+ Epoch 1 | iter 1056 step 33 | loss train: 1.577, val: 1.516 | iter time: 361.56 ms (step) remaining time: 0:30:15
133
+ Epoch 1 | iter 1088 step 34 | loss train: 1.599, val: 1.516 | iter time: 358.77 ms (step) remaining time: 0:30:05
134
+ Epoch 1 | iter 1120 step 35 | loss train: 1.550, val: 1.516 | iter time: 360.67 ms (step) remaining time: 0:29:54
135
+ Epoch 1 | iter 1152 step 36 | loss train: 1.533, val: 1.516 | iter time: 360.06 ms (step) remaining time: 0:29:43
136
+ Epoch 1 | iter 1184 step 37 | loss train: 1.513, val: 1.516 | iter time: 359.84 ms (step) remaining time: 0:29:33
137
+ Epoch 1 | iter 1216 step 38 | loss train: 1.606, val: 1.516 | iter time: 360.69 ms (step) remaining time: 0:29:22
138
+ Epoch 1 | iter 1248 step 39 | loss train: 1.558, val: 1.516 | iter time: 359.09 ms (step) remaining time: 0:29:12
139
+ Epoch 1 | iter 1280 step 40 | loss train: 1.489, val: 1.516 | iter time: 361.79 ms (step) remaining time: 0:29:01
140
+ Epoch 1 | iter 1312 step 41 | loss train: 1.533, val: 1.516 | iter time: 360.68 ms (step) remaining time: 0:28:50
141
+ Epoch 1 | iter 1344 step 42 | loss train: 1.495, val: 1.516 | iter time: 361.09 ms (step) remaining time: 0:28:39
142
+ Epoch 1 | iter 1376 step 43 | loss train: 1.547, val: 1.516 | iter time: 360.76 ms (step) remaining time: 0:28:28
143
+ Epoch 1 | iter 1408 step 44 | loss train: 1.558, val: 1.516 | iter time: 359.65 ms (step) remaining time: 0:28:17
144
+ Epoch 1 | iter 1440 step 45 | loss train: 1.586, val: 1.516 | iter time: 359.14 ms (step) remaining time: 0:28:06
145
+ Epoch 1 | iter 1472 step 46 | loss train: 1.533, val: 1.516 | iter time: 360.92 ms (step) remaining time: 0:27:55
146
+ Epoch 1 | iter 1504 step 47 | loss train: 1.532, val: 1.516 | iter time: 358.97 ms (step) remaining time: 0:27:44
147
+ Epoch 1 | iter 1536 step 48 | loss train: 1.502, val: 1.516 | iter time: 359.92 ms (step) remaining time: 0:27:33
148
+ Epoch 1 | iter 1568 step 49 | loss train: 1.510, val: 1.516 | iter time: 358.66 ms (step) remaining time: 0:27:23
149
+ Epoch 1 | iter 1600 step 50 | loss train: 1.517, val: 1.516 | iter time: 359.42 ms (step) remaining time: 0:27:12
150
+ Validating ...
151
+ iter 1600: val loss 1.5616, val time: 6628.66 ms
152
+ Epoch 1 | iter 1632 step 51 | loss train: 1.507, val: 1.562 | iter time: 358.98 ms (step) remaining time: 0:27:20
153
+ Epoch 1 | iter 1664 step 52 | loss train: 1.527, val: 1.562 | iter time: 359.82 ms (step) remaining time: 0:27:09
154
+ Epoch 1 | iter 1696 step 53 | loss train: 1.489, val: 1.562 | iter time: 361.73 ms (step) remaining time: 0:26:57
155
+ Epoch 1 | iter 1728 step 54 | loss train: 1.456, val: 1.562 | iter time: 359.39 ms (step) remaining time: 0:26:46
156
+ Epoch 1 | iter 1760 step 55 | loss train: 1.488, val: 1.562 | iter time: 359.70 ms (step) remaining time: 0:26:35
157
+ Epoch 1 | iter 1792 step 56 | loss train: 1.533, val: 1.562 | iter time: 359.97 ms (step) remaining time: 0:26:23
158
+ Epoch 1 | iter 1824 step 57 | loss train: 1.526, val: 1.562 | iter time: 359.95 ms (step) remaining time: 0:26:12
159
+ Epoch 1 | iter 1856 step 58 | loss train: 1.454, val: 1.562 | iter time: 359.66 ms (step) remaining time: 0:26:01
160
+ Epoch 1 | iter 1888 step 59 | loss train: 1.540, val: 1.562 | iter time: 358.72 ms (step) remaining time: 0:25:49
161
+ Epoch 1 | iter 1920 step 60 | loss train: 1.457, val: 1.562 | iter time: 358.39 ms (step) remaining time: 0:25:38
162
+ Epoch 1 | iter 1952 step 61 | loss train: 1.477, val: 1.562 | iter time: 362.21 ms (step) remaining time: 0:25:27
163
+ Epoch 1 | iter 1984 step 62 | loss train: 1.504, val: 1.562 | iter time: 360.92 ms (step) remaining time: 0:25:16
164
+ Epoch 1 | iter 2016 step 63 | loss train: 1.550, val: 1.562 | iter time: 360.32 ms (step) remaining time: 0:25:04
165
+ Epoch 1 | iter 2048 step 64 | loss train: 1.455, val: 1.562 | iter time: 358.84 ms (step) remaining time: 0:24:53
166
+ Epoch 1 | iter 2080 step 65 | loss train: 1.513, val: 1.562 | iter time: 360.02 ms (step) remaining time: 0:24:42
167
+ Epoch 1 | iter 2112 step 66 | loss train: 1.529, val: 1.562 | iter time: 360.58 ms (step) remaining time: 0:24:31
168
+ Epoch 1 | iter 2144 step 67 | loss train: 1.479, val: 1.562 | iter time: 360.08 ms (step) remaining time: 0:24:20
169
+ Epoch 1 | iter 2176 step 68 | loss train: 1.486, val: 1.562 | iter time: 359.46 ms (step) remaining time: 0:24:08
170
+ Epoch 1 | iter 2208 step 69 | loss train: 1.503, val: 1.562 | iter time: 360.66 ms (step) remaining time: 0:23:57
171
+ Epoch 1 | iter 2240 step 70 | loss train: 1.510, val: 1.562 | iter time: 361.29 ms (step) remaining time: 0:23:46
172
+ Epoch 1 | iter 2272 step 71 | loss train: 1.531, val: 1.562 | iter time: 359.82 ms (step) remaining time: 0:23:35
173
+ Epoch 1 | iter 2304 step 72 | loss train: 1.534, val: 1.562 | iter time: 359.55 ms (step) remaining time: 0:23:24
174
+ Epoch 1 | iter 2336 step 73 | loss train: 1.541, val: 1.562 | iter time: 360.04 ms (step) remaining time: 0:23:13
175
+ Epoch 1 | iter 2368 step 74 | loss train: 1.502, val: 1.562 | iter time: 360.40 ms (step) remaining time: 0:23:02
176
+ Epoch 1 | iter 2400 step 75 | loss train: 1.452, val: 1.562 | iter time: 358.72 ms (step) remaining time: 0:22:50
177
+ Epoch 1 | iter 2432 step 76 | loss train: 1.507, val: 1.562 | iter time: 360.41 ms (step) remaining time: 0:22:39
178
+ Epoch 1 | iter 2464 step 77 | loss train: 1.470, val: 1.562 | iter time: 358.43 ms (step) remaining time: 0:22:28
179
+ Epoch 1 | iter 2496 step 78 | loss train: 1.470, val: 1.562 | iter time: 359.45 ms (step) remaining time: 0:22:17
180
+ Epoch 1 | iter 2528 step 79 | loss train: 1.500, val: 1.562 | iter time: 360.39 ms (step) remaining time: 0:22:06
181
+ Epoch 1 | iter 2560 step 80 | loss train: 1.505, val: 1.562 | iter time: 360.25 ms (step) remaining time: 0:21:55
182
+ Epoch 1 | iter 2592 step 81 | loss train: 1.536, val: 1.562 | iter time: 360.15 ms (step) remaining time: 0:21:44
183
+ Epoch 1 | iter 2624 step 82 | loss train: 1.445, val: 1.562 | iter time: 360.00 ms (step) remaining time: 0:21:33
184
+ Epoch 1 | iter 2656 step 83 | loss train: 1.493, val: 1.562 | iter time: 360.27 ms (step) remaining time: 0:21:22
185
+ Epoch 1 | iter 2688 step 84 | loss train: 1.455, val: 1.562 | iter time: 360.84 ms (step) remaining time: 0:21:11
186
+ Epoch 1 | iter 2720 step 85 | loss train: 1.536, val: 1.562 | iter time: 360.31 ms (step) remaining time: 0:21:00
187
+ Epoch 1 | iter 2752 step 86 | loss train: 1.533, val: 1.562 | iter time: 359.74 ms (step) remaining time: 0:20:49
188
+ Epoch 1 | iter 2784 step 87 | loss train: 1.469, val: 1.562 | iter time: 359.46 ms (step) remaining time: 0:20:38
189
+ Epoch 1 | iter 2816 step 88 | loss train: 1.495, val: 1.562 | iter time: 359.86 ms (step) remaining time: 0:20:26
190
+ Epoch 1 | iter 2848 step 89 | loss train: 1.477, val: 1.562 | iter time: 360.08 ms (step) remaining time: 0:20:15
191
+ Epoch 1 | iter 2880 step 90 | loss train: 1.643, val: 1.562 | iter time: 362.43 ms (step) remaining time: 0:20:05
192
+ Epoch 1 | iter 2912 step 91 | loss train: 1.529, val: 1.562 | iter time: 360.04 ms (step) remaining time: 0:19:54
193
+ Epoch 1 | iter 2944 step 92 | loss train: 1.460, val: 1.562 | iter time: 360.48 ms (step) remaining time: 0:19:43
194
+ Epoch 1 | iter 2976 step 93 | loss train: 1.487, val: 1.562 | iter time: 359.14 ms (step) remaining time: 0:19:32
195
+ Epoch 1 | iter 3008 step 94 | loss train: 1.463, val: 1.562 | iter time: 357.89 ms (step) remaining time: 0:19:20
196
+ Epoch 1 | iter 3040 step 95 | loss train: 1.511, val: 1.562 | iter time: 359.13 ms (step) remaining time: 0:19:09
197
+ Epoch 1 | iter 3072 step 96 | loss train: 1.442, val: 1.562 | iter time: 358.94 ms (step) remaining time: 0:18:58
198
+ Epoch 1 | iter 3104 step 97 | loss train: 1.454, val: 1.562 | iter time: 358.38 ms (step) remaining time: 0:18:47
199
+ Epoch 1 | iter 3136 step 98 | loss train: 1.474, val: 1.562 | iter time: 360.59 ms (step) remaining time: 0:18:36
200
+ Epoch 1 | iter 3168 step 99 | loss train: 1.445, val: 1.562 | iter time: 357.61 ms (step) remaining time: 0:18:25
201
+ Epoch 1 | iter 3200 step 100 | loss train: 1.512, val: 1.562 | iter time: 362.55 ms (step) remaining time: 0:18:14
202
+ Validating ...
203
+ iter 3200: val loss 1.5233, val time: 6640.06 ms
204
+ Saving checkpoint to 'out/pretrain/tiny-llama-cl-2407/step-00000100/lit_model.pth'
205
+ Epoch 1 | iter 3232 step 101 | loss train: 1.448, val: 1.523 | iter time: 356.20 ms (step) remaining time: 0:18:27
206
+ Epoch 1 | iter 3264 step 102 | loss train: 1.466, val: 1.523 | iter time: 356.62 ms (step) remaining time: 0:18:15
207
+ Epoch 1 | iter 3296 step 103 | loss train: 1.437, val: 1.523 | iter time: 359.01 ms (step) remaining time: 0:18:04
208
+ Epoch 1 | iter 3328 step 104 | loss train: 1.412, val: 1.523 | iter time: 360.12 ms (step) remaining time: 0:17:52
209
+ Epoch 1 | iter 3360 step 105 | loss train: 1.415, val: 1.523 | iter time: 358.81 ms (step) remaining time: 0:17:41
210
+ Epoch 1 | iter 3392 step 106 | loss train: 1.438, val: 1.523 | iter time: 360.69 ms (step) remaining time: 0:17:29
211
+ Epoch 1 | iter 3424 step 107 | loss train: 1.398, val: 1.523 | iter time: 360.57 ms (step) remaining time: 0:17:18
212
+ Epoch 1 | iter 3456 step 108 | loss train: 1.464, val: 1.523 | iter time: 359.88 ms (step) remaining time: 0:17:06
213
+ Epoch 1 | iter 3488 step 109 | loss train: 1.482, val: 1.523 | iter time: 360.54 ms (step) remaining time: 0:16:55
214
+ Epoch 1 | iter 3520 step 110 | loss train: 1.419, val: 1.523 | iter time: 359.31 ms (step) remaining time: 0:16:44
215
+ Epoch 1 | iter 3552 step 111 | loss train: 1.473, val: 1.523 | iter time: 359.97 ms (step) remaining time: 0:16:32
216
+ Epoch 1 | iter 3584 step 112 | loss train: 1.395, val: 1.523 | iter time: 360.46 ms (step) remaining time: 0:16:21
217
+ Epoch 1 | iter 3616 step 113 | loss train: 1.497, val: 1.523 | iter time: 358.72 ms (step) remaining time: 0:16:09
218
+ Epoch 1 | iter 3648 step 114 | loss train: 1.438, val: 1.523 | iter time: 360.58 ms (step) remaining time: 0:15:58
219
+ Epoch 1 | iter 3680 step 115 | loss train: 1.437, val: 1.523 | iter time: 359.75 ms (step) remaining time: 0:15:47
220
+ Epoch 1 | iter 3712 step 116 | loss train: 1.425, val: 1.523 | iter time: 359.15 ms (step) remaining time: 0:15:35
221
+ Epoch 1 | iter 3744 step 117 | loss train: 1.457, val: 1.523 | iter time: 360.72 ms (step) remaining time: 0:15:24
222
+ Epoch 1 | iter 3776 step 118 | loss train: 1.492, val: 1.523 | iter time: 359.21 ms (step) remaining time: 0:15:13
223
+ Epoch 1 | iter 3808 step 119 | loss train: 1.471, val: 1.523 | iter time: 358.41 ms (step) remaining time: 0:15:01
224
+ Epoch 1 | iter 3840 step 120 | loss train: 1.430, val: 1.523 | iter time: 359.80 ms (step) remaining time: 0:14:50
225
+ Epoch 1 | iter 3872 step 121 | loss train: 1.463, val: 1.523 | iter time: 359.17 ms (step) remaining time: 0:14:39
226
+ Epoch 1 | iter 3904 step 122 | loss train: 1.557, val: 1.523 | iter time: 358.69 ms (step) remaining time: 0:14:28
227
+ Epoch 1 | iter 3936 step 123 | loss train: 1.398, val: 1.523 | iter time: 357.78 ms (step) remaining time: 0:14:16
228
+ Epoch 1 | iter 3968 step 124 | loss train: 1.395, val: 1.523 | iter time: 359.69 ms (step) remaining time: 0:14:05
229
+ Epoch 1 | iter 4000 step 125 | loss train: 1.423, val: 1.523 | iter time: 361.51 ms (step) remaining time: 0:13:54
230
+ Epoch 1 | iter 4032 step 126 | loss train: 1.393, val: 1.523 | iter time: 359.88 ms (step) remaining time: 0:13:42
231
+ Epoch 1 | iter 4064 step 127 | loss train: 1.474, val: 1.523 | iter time: 359.89 ms (step) remaining time: 0:13:31
232
+ Epoch 1 | iter 4096 step 128 | loss train: 1.410, val: 1.523 | iter time: 360.42 ms (step) remaining time: 0:13:20
233
+ Epoch 1 | iter 4128 step 129 | loss train: 1.413, val: 1.523 | iter time: 358.71 ms (step) remaining time: 0:13:09
234
+ Epoch 1 | iter 4160 step 130 | loss train: 1.391, val: 1.523 | iter time: 360.61 ms (step) remaining time: 0:12:57
235
+ Epoch 1 | iter 4192 step 131 | loss train: 1.536, val: 1.523 | iter time: 359.01 ms (step) remaining time: 0:12:46
236
+ Epoch 1 | iter 4224 step 132 | loss train: 1.413, val: 1.523 | iter time: 358.95 ms (step) remaining time: 0:12:35
237
+ Epoch 1 | iter 4256 step 133 | loss train: 1.446, val: 1.523 | iter time: 359.39 ms (step) remaining time: 0:12:24
238
+ Epoch 1 | iter 4288 step 134 | loss train: 1.382, val: 1.523 | iter time: 360.98 ms (step) remaining time: 0:12:12
239
+ Epoch 1 | iter 4320 step 135 | loss train: 1.471, val: 1.523 | iter time: 359.84 ms (step) remaining time: 0:12:01
240
+ Epoch 1 | iter 4352 step 136 | loss train: 1.497, val: 1.523 | iter time: 359.45 ms (step) remaining time: 0:11:50
241
+ Epoch 1 | iter 4384 step 137 | loss train: 1.422, val: 1.523 | iter time: 359.18 ms (step) remaining time: 0:11:39
242
+ Epoch 1 | iter 4416 step 138 | loss train: 1.464, val: 1.523 | iter time: 358.30 ms (step) remaining time: 0:11:28
243
+ Epoch 1 | iter 4448 step 139 | loss train: 1.430, val: 1.523 | iter time: 358.24 ms (step) remaining time: 0:11:16
244
+ Epoch 1 | iter 4480 step 140 | loss train: 1.441, val: 1.523 | iter time: 361.83 ms (step) remaining time: 0:11:05
245
+ Epoch 1 | iter 4512 step 141 | loss train: 1.416, val: 1.523 | iter time: 359.28 ms (step) remaining time: 0:10:54
246
+ Epoch 1 | iter 4544 step 142 | loss train: 1.491, val: 1.523 | iter time: 360.77 ms (step) remaining time: 0:10:43
247
+ Epoch 1 | iter 4576 step 143 | loss train: 1.399, val: 1.523 | iter time: 360.71 ms (step) remaining time: 0:10:32
248
+ Epoch 1 | iter 4608 step 144 | loss train: 1.367, val: 1.523 | iter time: 359.67 ms (step) remaining time: 0:10:21
249
+ Epoch 1 | iter 4640 step 145 | loss train: 1.403, val: 1.523 | iter time: 359.76 ms (step) remaining time: 0:10:10
250
+ Epoch 1 | iter 4672 step 146 | loss train: 1.433, val: 1.523 | iter time: 360.78 ms (step) remaining time: 0:09:58
251
+ Epoch 1 | iter 4704 step 147 | loss train: 1.380, val: 1.523 | iter time: 358.94 ms (step) remaining time: 0:09:47
252
+ Epoch 1 | iter 4736 step 148 | loss train: 1.496, val: 1.523 | iter time: 358.98 ms (step) remaining time: 0:09:36
253
+ Epoch 1 | iter 4768 step 149 | loss train: 1.386, val: 1.523 | iter time: 359.20 ms (step) remaining time: 0:09:25
254
+ Epoch 1 | iter 4800 step 150 | loss train: 1.393, val: 1.523 | iter time: 358.29 ms (step) remaining time: 0:09:14
255
+ Validating ...
256
+ iter 4800: val loss 1.4899, val time: 6636.92 ms
257
+ Epoch 1 | iter 4832 step 151 | loss train: 1.341, val: 1.490 | iter time: 360.17 ms (step) remaining time: 0:09:05
258
+ Epoch 1 | iter 4864 step 152 | loss train: 1.394, val: 1.490 | iter time: 357.93 ms (step) remaining time: 0:08:53
259
+ Epoch 1 | iter 4896 step 153 | loss train: 1.427, val: 1.490 | iter time: 360.60 ms (step) remaining time: 0:08:42
260
+ Epoch 1 | iter 4928 step 154 | loss train: 1.407, val: 1.490 | iter time: 359.68 ms (step) remaining time: 0:08:31
261
+ Epoch 1 | iter 4960 step 155 | loss train: 1.346, val: 1.490 | iter time: 358.64 ms (step) remaining time: 0:08:20
262
+ Epoch 1 | iter 4992 step 156 | loss train: 1.481, val: 1.490 | iter time: 359.33 ms (step) remaining time: 0:08:09
263
+ Epoch 1 | iter 5024 step 157 | loss train: 1.417, val: 1.490 | iter time: 360.26 ms (step) remaining time: 0:07:57
264
+ Epoch 1 | iter 5056 step 158 | loss train: 1.428, val: 1.490 | iter time: 359.38 ms (step) remaining time: 0:07:46
265
+ Epoch 1 | iter 5088 step 159 | loss train: 1.464, val: 1.490 | iter time: 360.43 ms (step) remaining time: 0:07:35
266
+ Epoch 1 | iter 5120 step 160 | loss train: 1.417, val: 1.490 | iter time: 359.95 ms (step) remaining time: 0:07:24
267
+ Epoch 1 | iter 5152 step 161 | loss train: 1.494, val: 1.490 | iter time: 359.53 ms (step) remaining time: 0:07:13
268
+ Epoch 1 | iter 5184 step 162 | loss train: 1.436, val: 1.490 | iter time: 360.53 ms (step) remaining time: 0:07:02
269
+ Epoch 1 | iter 5216 step 163 | loss train: 1.371, val: 1.490 | iter time: 360.98 ms (step) remaining time: 0:06:50
270
+ Epoch 1 | iter 5248 step 164 | loss train: 1.407, val: 1.490 | iter time: 357.79 ms (step) remaining time: 0:06:39
271
+ Epoch 1 | iter 5280 step 165 | loss train: 1.421, val: 1.490 | iter time: 360.59 ms (step) remaining time: 0:06:28
272
+ Epoch 1 | iter 5312 step 166 | loss train: 1.395, val: 1.490 | iter time: 357.97 ms (step) remaining time: 0:06:17
273
+ Epoch 1 | iter 5344 step 167 | loss train: 1.363, val: 1.490 | iter time: 359.06 ms (step) remaining time: 0:06:06
274
+ Epoch 1 | iter 5376 step 168 | loss train: 1.416, val: 1.490 | iter time: 360.67 ms (step) remaining time: 0:05:55
275
+ Epoch 1 | iter 5408 step 169 | loss train: 1.430, val: 1.490 | iter time: 360.25 ms (step) remaining time: 0:05:44
276
+ Epoch 1 | iter 5440 step 170 | loss train: 1.386, val: 1.490 | iter time: 360.98 ms (step) remaining time: 0:05:32
277
+ Epoch 1 | iter 5472 step 171 | loss train: 1.478, val: 1.490 | iter time: 360.58 ms (step) remaining time: 0:05:21
278
+ Epoch 1 | iter 5504 step 172 | loss train: 1.396, val: 1.490 | iter time: 360.11 ms (step) remaining time: 0:05:10
279
+ Epoch 1 | iter 5536 step 173 | loss train: 1.390, val: 1.490 | iter time: 359.66 ms (step) remaining time: 0:04:59
280
+ Epoch 1 | iter 5568 step 174 | loss train: 1.377, val: 1.490 | iter time: 359.39 ms (step) remaining time: 0:04:48
281
+ Epoch 1 | iter 5600 step 175 | loss train: 1.370, val: 1.490 | iter time: 360.91 ms (step) remaining time: 0:04:37
282
+ Epoch 1 | iter 5632 step 176 | loss train: 1.407, val: 1.490 | iter time: 360.76 ms (step) remaining time: 0:04:26
283
+ Epoch 1 | iter 5664 step 177 | loss train: 1.474, val: 1.490 | iter time: 358.30 ms (step) remaining time: 0:04:15
284
+ Epoch 1 | iter 5696 step 178 | loss train: 1.480, val: 1.490 | iter time: 359.12 ms (step) remaining time: 0:04:03
285
+ Epoch 1 | iter 5728 step 179 | loss train: 1.418, val: 1.490 | iter time: 361.16 ms (step) remaining time: 0:03:52
286
+ Epoch 1 | iter 5760 step 180 | loss train: 1.447, val: 1.490 | iter time: 359.35 ms (step) remaining time: 0:03:41
287
+ Epoch 1 | iter 5792 step 181 | loss train: 1.380, val: 1.490 | iter time: 359.73 ms (step) remaining time: 0:03:30
288
+ Epoch 1 | iter 5824 step 182 | loss train: 1.433, val: 1.490 | iter time: 359.42 ms (step) remaining time: 0:03:19
289
+ Epoch 1 | iter 5856 step 183 | loss train: 1.377, val: 1.490 | iter time: 359.69 ms (step) remaining time: 0:03:08
290
+ Epoch 1 | iter 5888 step 184 | loss train: 1.391, val: 1.490 | iter time: 358.68 ms (step) remaining time: 0:02:57
291
+ Epoch 1 | iter 5920 step 185 | loss train: 1.303, val: 1.490 | iter time: 360.25 ms (step) remaining time: 0:02:46
292
+ Epoch 1 | iter 5952 step 186 | loss train: 1.414, val: 1.490 | iter time: 357.99 ms (step) remaining time: 0:02:35
293
+ Epoch 1 | iter 5984 step 187 | loss train: 1.396, val: 1.490 | iter time: 360.09 ms (step) remaining time: 0:02:23
294
+ Epoch 1 | iter 6016 step 188 | loss train: 1.407, val: 1.490 | iter time: 360.49 ms (step) remaining time: 0:02:12
295
+ Epoch 1 | iter 6048 step 189 | loss train: 1.336, val: 1.490 | iter time: 359.10 ms (step) remaining time: 0:02:01
296
+ Epoch 1 | iter 6080 step 190 | loss train: 1.436, val: 1.490 | iter time: 359.23 ms (step) remaining time: 0:01:50
297
+ Epoch 1 | iter 6112 step 191 | loss train: 1.408, val: 1.490 | iter time: 359.31 ms (step) remaining time: 0:01:39
298
+ Epoch 1 | iter 6144 step 192 | loss train: 1.466, val: 1.490 | iter time: 359.06 ms (step) remaining time: 0:01:28
299
+ Epoch 1 | iter 6176 step 193 | loss train: 1.413, val: 1.490 | iter time: 359.53 ms (step) remaining time: 0:01:17
300
+ Epoch 1 | iter 6208 step 194 | loss train: 1.356, val: 1.490 | iter time: 361.61 ms (step) remaining time: 0:01:06
301
+ Epoch 1 | iter 6240 step 195 | loss train: 1.480, val: 1.490 | iter time: 361.66 ms (step) remaining time: 0:00:55
302
+ Epoch 1 | iter 6272 step 196 | loss train: 1.454, val: 1.490 | iter time: 361.34 ms (step) remaining time: 0:00:44
303
+ Epoch 1 | iter 6304 step 197 | loss train: 1.361, val: 1.490 | iter time: 358.70 ms (step) remaining time: 0:00:33
304
+ Epoch 1 | iter 6336 step 198 | loss train: 1.430, val: 1.490 | iter time: 360.10 ms (step) remaining time: 0:00:22
305
+ Epoch 1 | iter 6368 step 199 | loss train: 1.464, val: 1.490 | iter time: 359.90 ms (step) remaining time: 0:00:11
306
+ Epoch 1 | iter 6400 step 200 | loss train: 1.453, val: 1.490 | iter time: 359.55 ms (step) remaining time: 0:00:00
307
+ Validating ...
308
+ iter 6400: val loss 1.4696, val time: 6635.54 ms
309
+ Saving checkpoint to 'out/pretrain/tiny-llama-cl-2407/step-00000200/lit_model.pth'
310
+ Validating ...
311
+ Final evaluation | val loss: 1.470 | val ppl: 4.348
312
+ Saving checkpoint to 'out/pretrain/tiny-llama-cl-2407/final/lit_model.pth'
313
+ ----------------------------------------
314
+ | Performance
315
+ | - Total tokens : 209,715,200
316
+ | - Training Time : 2269.73 s
317
+ | - Tok/sec : 214.47 tok/s
318
+ | ----------------------------------------
319
+ | Memory Usage
320
+ | - Memory Used : 26.32 GB
321
+ ----------------------------------------
out/pretrain/tinyllama/teelogs/2407_lr4e-5.txt ADDED
@@ -0,0 +1,334 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Initializing distributed: GLOBAL_RANK: 0, MEMBER: 1/4
2
+ Initializing distributed: GLOBAL_RANK: 2, MEMBER: 3/4
3
+ Initializing distributed: GLOBAL_RANK: 1, MEMBER: 2/4
4
+ [rank: 2] Seed set to 42
5
+ [rank: 1] Seed set to 42
6
+ Initializing distributed: GLOBAL_RANK: 3, MEMBER: 4/4
7
+ [rank: 3] Seed set to 42
8
+ ----------------------------------------------------------------------------------------------------
9
+ distributed_backend=nccl
10
+ All distributed processes registered. Starting with 4 processes
11
+ ----------------------------------------------------------------------------------------------------
12
+
13
+ [rank: 0] Seed set to 42
14
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
15
+ train_dataloader = data.train_dataloader()
16
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
17
+ train_dataloader = data.train_dataloader()
18
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
19
+ train_dataloader = data.train_dataloader()
20
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
21
+ train_dataloader = data.train_dataloader()
22
+ All GPUs are fully connected via NVLink.
23
+ {'data': {'batch_size': 1,
24
+ 'data_path': PosixPath('litgpt/data/arxiv'),
25
+ 'num_workers': 0,
26
+ 'seed': 42,
27
+ 'seq_length': 2048,
28
+ 'use_starcoder': True},
29
+ 'data_dir': PosixPath('litgpt/data/arxiv/2407'),
30
+ 'devices': 'auto',
31
+ 'eval': {'evaluate_example': 'first',
32
+ 'final_validation': True,
33
+ 'initial_validation': True,
34
+ 'interval': 50,
35
+ 'max_iters': 200,
36
+ 'max_new_tokens': None},
37
+ 'initial_checkpoint_dir': PosixPath('checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T'),
38
+ 'log': {'group': None, 'project': None, 'run': None},
39
+ 'logger_name': 'tensorboard',
40
+ 'model_config': {'attention_logit_softcapping': None,
41
+ 'attention_scores_scalar': None,
42
+ 'attn_bias': False,
43
+ 'bias': False,
44
+ 'block_size': 2048,
45
+ 'final_logit_softcapping': None,
46
+ 'gelu_approximate': 'none',
47
+ 'head_size': 64,
48
+ 'hf_config': {'name': 'TinyLlama-1.1B-intermediate-step-1431k-3T',
49
+ 'org': 'TinyLlama'},
50
+ 'intermediate_size': 5632,
51
+ 'lm_head_bias': False,
52
+ 'mlp_class_name': 'LLaMAMLP',
53
+ 'moe_intermediate_size': None,
54
+ 'n_embd': 2048,
55
+ 'n_expert': 0,
56
+ 'n_expert_per_token': 0,
57
+ 'n_head': 32,
58
+ 'n_layer': 22,
59
+ 'n_query_groups': 4,
60
+ 'name': 'tiny-llama-1.1b',
61
+ 'norm_1': True,
62
+ 'norm_2': True,
63
+ 'norm_class_name': 'RMSNorm',
64
+ 'norm_eps': 1e-05,
65
+ 'norm_qk': False,
66
+ 'norm_qk_type': 'default',
67
+ 'padded_vocab_size': 32000,
68
+ 'padding_multiple': 64,
69
+ 'parallel_residual': False,
70
+ 'post_attention_norm': False,
71
+ 'post_mlp_norm': False,
72
+ 'rope_adjustments': None,
73
+ 'rope_base': 10000,
74
+ 'rope_condense_ratio': 1,
75
+ 'rope_indices': None,
76
+ 'rope_local_base_freq': None,
77
+ 'rotary_percentage': 1.0,
78
+ 'scale_embeddings': False,
79
+ 'shared_attention_norm': False,
80
+ 'sliding_window_indices': None,
81
+ 'sliding_window_size': None,
82
+ 'vocab_size': 32000},
83
+ 'model_name': 'tiny-llama-1.1b',
84
+ 'num_nodes': 1,
85
+ 'optimizer': "{'class_path': 'torch.optim.AdamW', 'init_args': {'lr': 4e-05, "
86
+ "'weight_decay': 0.1, 'betas': [0.9, 0.95]}}",
87
+ 'out_dir': PosixPath('out/pretrain/tinyllama/2407_lr4e-5'),
88
+ 'ppl': False,
89
+ 'precision': 'bf16-mixed',
90
+ 'resume': False,
91
+ 'seed': 42,
92
+ 'tokenizer_dir': PosixPath('checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T'),
93
+ 'train': {'epochs': None,
94
+ 'global_batch_size': 512,
95
+ 'log_interval': 1,
96
+ 'lr_warmup_fraction': None,
97
+ 'lr_warmup_steps': 20,
98
+ 'max_norm': 1.0,
99
+ 'max_seq_length': 2048,
100
+ 'max_steps': None,
101
+ 'max_tokens': 211812352,
102
+ 'micro_batch_size': 4,
103
+ 'min_lr': 4e-05,
104
+ 'save_interval': 100,
105
+ 'tie_embeddings': None}}
106
+ Time to instantiate model: 0.04 seconds.
107
+ Total parameters: 1,100,048,384
108
+ [ok] checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T/lit_model.pth 已是纯权重
109
+ Validating ...
110
+ Measured TFLOPs: 239.66
111
+ Epoch 1 | iter 32 step 1 | loss train: 1.462, val: 1.508 | iter time: 537.76 ms (step) remaining time: 0:39:00
112
+ Epoch 1 | iter 64 step 2 | loss train: 1.481, val: 1.508 | iter time: 356.70 ms (step) remaining time: 0:37:19
113
+ Epoch 1 | iter 96 step 3 | loss train: 1.474, val: 1.508 | iter time: 358.11 ms (step) remaining time: 0:36:46
114
+ Epoch 1 | iter 128 step 4 | loss train: 1.453, val: 1.508 | iter time: 355.68 ms (step) remaining time: 0:36:20
115
+ Epoch 1 | iter 160 step 5 | loss train: 1.454, val: 1.508 | iter time: 358.16 ms (step) remaining time: 0:36:04
116
+ Epoch 1 | iter 192 step 6 | loss train: 1.477, val: 1.508 | iter time: 355.87 ms (step) remaining time: 0:35:48
117
+ Epoch 1 | iter 224 step 7 | loss train: 1.460, val: 1.508 | iter time: 357.58 ms (step) remaining time: 0:35:33
118
+ Epoch 1 | iter 256 step 8 | loss train: 1.446, val: 1.508 | iter time: 359.29 ms (step) remaining time: 0:35:19
119
+ Epoch 1 | iter 288 step 9 | loss train: 1.419, val: 1.508 | iter time: 357.58 ms (step) remaining time: 0:35:06
120
+ Epoch 1 | iter 320 step 10 | loss train: 1.465, val: 1.508 | iter time: 359.52 ms (step) remaining time: 0:34:54
121
+ Epoch 1 | iter 352 step 11 | loss train: 1.426, val: 1.508 | iter time: 359.00 ms (step) remaining time: 0:34:41
122
+ Epoch 1 | iter 384 step 12 | loss train: 1.438, val: 1.508 | iter time: 357.75 ms (step) remaining time: 0:34:29
123
+ Epoch 1 | iter 416 step 13 | loss train: 1.439, val: 1.508 | iter time: 358.76 ms (step) remaining time: 0:34:18
124
+ Epoch 1 | iter 448 step 14 | loss train: 1.486, val: 1.508 | iter time: 358.92 ms (step) remaining time: 0:34:06
125
+ Epoch 1 | iter 480 step 15 | loss train: 1.448, val: 1.508 | iter time: 359.39 ms (step) remaining time: 0:33:55
126
+ Epoch 1 | iter 512 step 16 | loss train: 1.503, val: 1.508 | iter time: 357.43 ms (step) remaining time: 0:33:44
127
+ Epoch 1 | iter 544 step 17 | loss train: 1.446, val: 1.508 | iter time: 358.08 ms (step) remaining time: 0:33:32
128
+ Epoch 1 | iter 576 step 18 | loss train: 1.427, val: 1.508 | iter time: 359.87 ms (step) remaining time: 0:33:21
129
+ Epoch 1 | iter 608 step 19 | loss train: 1.503, val: 1.508 | iter time: 360.56 ms (step) remaining time: 0:33:10
130
+ Epoch 1 | iter 640 step 20 | loss train: 1.482, val: 1.508 | iter time: 358.93 ms (step) remaining time: 0:32:59
131
+ Epoch 1 | iter 672 step 21 | loss train: 1.503, val: 1.508 | iter time: 357.52 ms (step) remaining time: 0:32:47
132
+ Epoch 1 | iter 704 step 22 | loss train: 1.463, val: 1.508 | iter time: 358.45 ms (step) remaining time: 0:32:36
133
+ Epoch 1 | iter 736 step 23 | loss train: 1.421, val: 1.508 | iter time: 360.40 ms (step) remaining time: 0:32:25
134
+ Epoch 1 | iter 768 step 24 | loss train: 1.413, val: 1.508 | iter time: 361.00 ms (step) remaining time: 0:32:14
135
+ Epoch 1 | iter 800 step 25 | loss train: 1.403, val: 1.508 | iter time: 357.88 ms (step) remaining time: 0:32:03
136
+ Epoch 1 | iter 832 step 26 | loss train: 1.445, val: 1.508 | iter time: 358.75 ms (step) remaining time: 0:31:52
137
+ Epoch 1 | iter 864 step 27 | loss train: 1.435, val: 1.508 | iter time: 358.12 ms (step) remaining time: 0:31:41
138
+ Epoch 1 | iter 896 step 28 | loss train: 1.460, val: 1.508 | iter time: 358.23 ms (step) remaining time: 0:31:30
139
+ Epoch 1 | iter 928 step 29 | loss train: 1.451, val: 1.508 | iter time: 361.24 ms (step) remaining time: 0:31:19
140
+ Epoch 1 | iter 960 step 30 | loss train: 1.380, val: 1.508 | iter time: 359.07 ms (step) remaining time: 0:31:08
141
+ Epoch 1 | iter 992 step 31 | loss train: 1.504, val: 1.508 | iter time: 358.70 ms (step) remaining time: 0:30:57
142
+ Epoch 1 | iter 1024 step 32 | loss train: 1.461, val: 1.508 | iter time: 359.33 ms (step) remaining time: 0:30:46
143
+ Epoch 1 | iter 1056 step 33 | loss train: 1.500, val: 1.508 | iter time: 357.92 ms (step) remaining time: 0:30:35
144
+ Epoch 1 | iter 1088 step 34 | loss train: 1.432, val: 1.508 | iter time: 358.42 ms (step) remaining time: 0:30:24
145
+ Epoch 1 | iter 1120 step 35 | loss train: 1.414, val: 1.508 | iter time: 360.10 ms (step) remaining time: 0:30:14
146
+ Epoch 1 | iter 1152 step 36 | loss train: 1.477, val: 1.508 | iter time: 358.07 ms (step) remaining time: 0:30:03
147
+ Epoch 1 | iter 1184 step 37 | loss train: 1.475, val: 1.508 | iter time: 360.08 ms (step) remaining time: 0:29:52
148
+ Epoch 1 | iter 1216 step 38 | loss train: 1.434, val: 1.508 | iter time: 359.17 ms (step) remaining time: 0:29:41
149
+ Epoch 1 | iter 1248 step 39 | loss train: 1.452, val: 1.508 | iter time: 358.45 ms (step) remaining time: 0:29:30
150
+ Epoch 1 | iter 1280 step 40 | loss train: 1.398, val: 1.508 | iter time: 358.07 ms (step) remaining time: 0:29:19
151
+ Epoch 1 | iter 1312 step 41 | loss train: 1.371, val: 1.508 | iter time: 358.66 ms (step) remaining time: 0:29:08
152
+ Epoch 1 | iter 1344 step 42 | loss train: 1.421, val: 1.508 | iter time: 359.26 ms (step) remaining time: 0:28:57
153
+ Epoch 1 | iter 1376 step 43 | loss train: 1.410, val: 1.508 | iter time: 358.69 ms (step) remaining time: 0:28:46
154
+ Epoch 1 | iter 1408 step 44 | loss train: 1.445, val: 1.508 | iter time: 358.95 ms (step) remaining time: 0:28:35
155
+ Epoch 1 | iter 1440 step 45 | loss train: 1.412, val: 1.508 | iter time: 357.91 ms (step) remaining time: 0:28:24
156
+ Epoch 1 | iter 1472 step 46 | loss train: 1.496, val: 1.508 | iter time: 357.62 ms (step) remaining time: 0:28:14
157
+ Epoch 1 | iter 1504 step 47 | loss train: 1.396, val: 1.508 | iter time: 360.43 ms (step) remaining time: 0:28:03
158
+ Epoch 1 | iter 1536 step 48 | loss train: 1.528, val: 1.508 | iter time: 357.49 ms (step) remaining time: 0:27:52
159
+ Epoch 1 | iter 1568 step 49 | loss train: 1.403, val: 1.508 | iter time: 360.38 ms (step) remaining time: 0:27:42
160
+ Epoch 1 | iter 1600 step 50 | loss train: 1.436, val: 1.508 | iter time: 359.54 ms (step) remaining time: 0:27:31
161
+ Validating ...
162
+ iter 1600: val loss 1.4361, val time: 21889.40 ms
163
+ Epoch 1 | iter 1632 step 51 | loss train: 1.399, val: 1.436 | iter time: 359.17 ms (step) remaining time: 0:28:25
164
+ Epoch 1 | iter 1664 step 52 | loss train: 1.365, val: 1.436 | iter time: 357.95 ms (step) remaining time: 0:28:12
165
+ Epoch 1 | iter 1696 step 53 | loss train: 1.386, val: 1.436 | iter time: 358.82 ms (step) remaining time: 0:27:59
166
+ Epoch 1 | iter 1728 step 54 | loss train: 1.465, val: 1.436 | iter time: 358.77 ms (step) remaining time: 0:27:47
167
+ Epoch 1 | iter 1760 step 55 | loss train: 1.426, val: 1.436 | iter time: 743.32 ms (step) remaining time: 0:27:36
168
+ Epoch 1 | iter 1792 step 56 | loss train: 1.446, val: 1.436 | iter time: 359.90 ms (step) remaining time: 0:27:23
169
+ Epoch 1 | iter 1824 step 57 | loss train: 1.410, val: 1.436 | iter time: 358.13 ms (step) remaining time: 0:27:11
170
+ Epoch 1 | iter 1856 step 58 | loss train: 1.328, val: 1.436 | iter time: 358.15 ms (step) remaining time: 0:26:59
171
+ Epoch 1 | iter 1888 step 59 | loss train: 1.427, val: 1.436 | iter time: 361.03 ms (step) remaining time: 0:26:47
172
+ Epoch 1 | iter 1920 step 60 | loss train: 1.401, val: 1.436 | iter time: 358.23 ms (step) remaining time: 0:26:34
173
+ Epoch 1 | iter 1952 step 61 | loss train: 1.462, val: 1.436 | iter time: 360.43 ms (step) remaining time: 0:26:22
174
+ Epoch 1 | iter 1984 step 62 | loss train: 1.468, val: 1.436 | iter time: 359.00 ms (step) remaining time: 0:26:10
175
+ Epoch 1 | iter 2016 step 63 | loss train: 1.465, val: 1.436 | iter time: 359.68 ms (step) remaining time: 0:25:58
176
+ Epoch 1 | iter 2048 step 64 | loss train: 1.435, val: 1.436 | iter time: 358.28 ms (step) remaining time: 0:25:46
177
+ Epoch 1 | iter 2080 step 65 | loss train: 1.422, val: 1.436 | iter time: 358.90 ms (step) remaining time: 0:25:34
178
+ Epoch 1 | iter 2112 step 66 | loss train: 1.490, val: 1.436 | iter time: 360.03 ms (step) remaining time: 0:25:22
179
+ Epoch 1 | iter 2144 step 67 | loss train: 1.377, val: 1.436 | iter time: 358.40 ms (step) remaining time: 0:25:10
180
+ Epoch 1 | iter 2176 step 68 | loss train: 1.425, val: 1.436 | iter time: 359.52 ms (step) remaining time: 0:24:59
181
+ Epoch 1 | iter 2208 step 69 | loss train: 1.444, val: 1.436 | iter time: 359.57 ms (step) remaining time: 0:24:47
182
+ Epoch 1 | iter 2240 step 70 | loss train: 1.378, val: 1.436 | iter time: 360.12 ms (step) remaining time: 0:24:35
183
+ Epoch 1 | iter 2272 step 71 | loss train: 1.439, val: 1.436 | iter time: 358.92 ms (step) remaining time: 0:24:23
184
+ Epoch 1 | iter 2304 step 72 | loss train: 1.468, val: 1.436 | iter time: 359.64 ms (step) remaining time: 0:24:11
185
+ Epoch 1 | iter 2336 step 73 | loss train: 1.421, val: 1.436 | iter time: 358.99 ms (step) remaining time: 0:24:00
186
+ Epoch 1 | iter 2368 step 74 | loss train: 1.446, val: 1.436 | iter time: 359.12 ms (step) remaining time: 0:23:48
187
+ Epoch 1 | iter 2400 step 75 | loss train: 1.320, val: 1.436 | iter time: 359.87 ms (step) remaining time: 0:23:36
188
+ Epoch 1 | iter 2432 step 76 | loss train: 1.437, val: 1.436 | iter time: 358.82 ms (step) remaining time: 0:23:25
189
+ Epoch 1 | iter 2464 step 77 | loss train: 1.389, val: 1.436 | iter time: 360.37 ms (step) remaining time: 0:23:13
190
+ Epoch 1 | iter 2496 step 78 | loss train: 1.423, val: 1.436 | iter time: 357.35 ms (step) remaining time: 0:23:01
191
+ Epoch 1 | iter 2528 step 79 | loss train: 1.434, val: 1.436 | iter time: 358.72 ms (step) remaining time: 0:22:50
192
+ Epoch 1 | iter 2560 step 80 | loss train: 1.422, val: 1.436 | iter time: 358.92 ms (step) remaining time: 0:22:38
193
+ Epoch 1 | iter 2592 step 81 | loss train: 1.413, val: 1.436 | iter time: 359.58 ms (step) remaining time: 0:22:26
194
+ Epoch 1 | iter 2624 step 82 | loss train: 1.347, val: 1.436 | iter time: 360.71 ms (step) remaining time: 0:22:15
195
+ Epoch 1 | iter 2656 step 83 | loss train: 1.413, val: 1.436 | iter time: 360.00 ms (step) remaining time: 0:22:03
196
+ Epoch 1 | iter 2688 step 84 | loss train: 1.431, val: 1.436 | iter time: 360.67 ms (step) remaining time: 0:21:52
197
+ Epoch 1 | iter 2720 step 85 | loss train: 1.463, val: 1.436 | iter time: 359.26 ms (step) remaining time: 0:21:41
198
+ Epoch 1 | iter 2752 step 86 | loss train: 1.453, val: 1.436 | iter time: 358.85 ms (step) remaining time: 0:21:29
199
+ Epoch 1 | iter 2784 step 87 | loss train: 1.369, val: 1.436 | iter time: 358.83 ms (step) remaining time: 0:21:18
200
+ Epoch 1 | iter 2816 step 88 | loss train: 1.450, val: 1.436 | iter time: 360.72 ms (step) remaining time: 0:21:06
201
+ Epoch 1 | iter 2848 step 89 | loss train: 1.467, val: 1.436 | iter time: 360.84 ms (step) remaining time: 0:20:55
202
+ Epoch 1 | iter 2880 step 90 | loss train: 1.429, val: 1.436 | iter time: 360.97 ms (step) remaining time: 0:20:43
203
+ Epoch 1 | iter 2912 step 91 | loss train: 1.427, val: 1.436 | iter time: 360.29 ms (step) remaining time: 0:20:32
204
+ Epoch 1 | iter 2944 step 92 | loss train: 1.341, val: 1.436 | iter time: 359.94 ms (step) remaining time: 0:20:21
205
+ Epoch 1 | iter 2976 step 93 | loss train: 1.379, val: 1.436 | iter time: 358.34 ms (step) remaining time: 0:20:09
206
+ Epoch 1 | iter 3008 step 94 | loss train: 1.442, val: 1.436 | iter time: 359.57 ms (step) remaining time: 0:19:58
207
+ Epoch 1 | iter 3040 step 95 | loss train: 1.353, val: 1.436 | iter time: 359.47 ms (step) remaining time: 0:19:47
208
+ Epoch 1 | iter 3072 step 96 | loss train: 1.358, val: 1.436 | iter time: 359.14 ms (step) remaining time: 0:19:35
209
+ Epoch 1 | iter 3104 step 97 | loss train: 1.383, val: 1.436 | iter time: 359.20 ms (step) remaining time: 0:19:24
210
+ Epoch 1 | iter 3136 step 98 | loss train: 1.347, val: 1.436 | iter time: 358.01 ms (step) remaining time: 0:19:13
211
+ Epoch 1 | iter 3168 step 99 | loss train: 1.429, val: 1.436 | iter time: 355.51 ms (step) remaining time: 0:19:01
212
+ Epoch 1 | iter 3200 step 100 | loss train: 1.425, val: 1.436 | iter time: 358.42 ms (step) remaining time: 0:18:50
213
+ Validating ...
214
+ iter 3200: val loss 1.3808, val time: 21890.43 ms
215
+ Saving checkpoint to 'out/pretrain/tinyllama/2407_lr4e-5/step-00000100/lit_model.pth'
216
+ Epoch 1 | iter 3232 step 101 | loss train: 1.519, val: 1.381 | iter time: 356.30 ms (step) remaining time: 0:19:17
217
+ Epoch 1 | iter 3264 step 102 | loss train: 1.437, val: 1.381 | iter time: 354.98 ms (step) remaining time: 0:19:05
218
+ Epoch 1 | iter 3296 step 103 | loss train: 1.498, val: 1.381 | iter time: 356.48 ms (step) remaining time: 0:18:53
219
+ Epoch 1 | iter 3328 step 104 | loss train: 1.403, val: 1.381 | iter time: 358.68 ms (step) remaining time: 0:18:41
220
+ Epoch 1 | iter 3360 step 105 | loss train: 1.367, val: 1.381 | iter time: 357.85 ms (step) remaining time: 0:18:29
221
+ Epoch 1 | iter 3392 step 106 | loss train: 1.405, val: 1.381 | iter time: 359.50 ms (step) remaining time: 0:18:17
222
+ Epoch 1 | iter 3424 step 107 | loss train: 1.477, val: 1.381 | iter time: 358.88 ms (step) remaining time: 0:18:05
223
+ Epoch 1 | iter 3456 step 108 | loss train: 1.476, val: 1.381 | iter time: 359.27 ms (step) remaining time: 0:17:53
224
+ Epoch 1 | iter 3488 step 109 | loss train: 1.497, val: 1.381 | iter time: 357.47 ms (step) remaining time: 0:17:41
225
+ Epoch 1 | iter 3520 step 110 | loss train: 1.401, val: 1.381 | iter time: 359.16 ms (step) remaining time: 0:17:29
226
+ Epoch 1 | iter 3552 step 111 | loss train: 1.444, val: 1.381 | iter time: 358.70 ms (step) remaining time: 0:17:17
227
+ Epoch 1 | iter 3584 step 112 | loss train: 1.447, val: 1.381 | iter time: 358.16 ms (step) remaining time: 0:17:05
228
+ Epoch 1 | iter 3616 step 113 | loss train: 1.399, val: 1.381 | iter time: 358.70 ms (step) remaining time: 0:16:54
229
+ Epoch 1 | iter 3648 step 114 | loss train: 1.456, val: 1.381 | iter time: 359.15 ms (step) remaining time: 0:16:42
230
+ Epoch 1 | iter 3680 step 115 | loss train: 1.437, val: 1.381 | iter time: 358.86 ms (step) remaining time: 0:16:30
231
+ Epoch 1 | iter 3712 step 116 | loss train: 1.467, val: 1.381 | iter time: 358.82 ms (step) remaining time: 0:16:18
232
+ Epoch 1 | iter 3744 step 117 | loss train: 1.431, val: 1.381 | iter time: 358.59 ms (step) remaining time: 0:16:06
233
+ Epoch 1 | iter 3776 step 118 | loss train: 1.429, val: 1.381 | iter time: 360.06 ms (step) remaining time: 0:15:55
234
+ Epoch 1 | iter 3808 step 119 | loss train: 1.382, val: 1.381 | iter time: 359.83 ms (step) remaining time: 0:15:43
235
+ Epoch 1 | iter 3840 step 120 | loss train: 1.425, val: 1.381 | iter time: 356.77 ms (step) remaining time: 0:15:31
236
+ Epoch 1 | iter 3872 step 121 | loss train: 1.457, val: 1.381 | iter time: 359.41 ms (step) remaining time: 0:15:20
237
+ Epoch 1 | iter 3904 step 122 | loss train: 1.426, val: 1.381 | iter time: 358.54 ms (step) remaining time: 0:15:08
238
+ Epoch 1 | iter 3936 step 123 | loss train: 1.414, val: 1.381 | iter time: 359.50 ms (step) remaining time: 0:14:56
239
+ Epoch 1 | iter 3968 step 124 | loss train: 1.405, val: 1.381 | iter time: 361.57 ms (step) remaining time: 0:14:45
240
+ Epoch 1 | iter 4000 step 125 | loss train: 1.412, val: 1.381 | iter time: 361.58 ms (step) remaining time: 0:14:33
241
+ Epoch 1 | iter 4032 step 126 | loss train: 1.390, val: 1.381 | iter time: 360.64 ms (step) remaining time: 0:14:22
242
+ Epoch 1 | iter 4064 step 127 | loss train: 1.404, val: 1.381 | iter time: 360.52 ms (step) remaining time: 0:14:10
243
+ Epoch 1 | iter 4096 step 128 | loss train: 1.421, val: 1.381 | iter time: 362.00 ms (step) remaining time: 0:13:58
244
+ Epoch 1 | iter 4128 step 129 | loss train: 1.398, val: 1.381 | iter time: 359.93 ms (step) remaining time: 0:13:47
245
+ Epoch 1 | iter 4160 step 130 | loss train: 1.489, val: 1.381 | iter time: 360.07 ms (step) remaining time: 0:13:35
246
+ Epoch 1 | iter 4192 step 131 | loss train: 1.397, val: 1.381 | iter time: 359.08 ms (step) remaining time: 0:13:24
247
+ Epoch 1 | iter 4224 step 132 | loss train: 1.450, val: 1.381 | iter time: 361.62 ms (step) remaining time: 0:13:12
248
+ Epoch 1 | iter 4256 step 133 | loss train: 1.457, val: 1.381 | iter time: 359.58 ms (step) remaining time: 0:13:01
249
+ Epoch 1 | iter 4288 step 134 | loss train: 1.382, val: 1.381 | iter time: 359.90 ms (step) remaining time: 0:12:49
250
+ Epoch 1 | iter 4320 step 135 | loss train: 1.433, val: 1.381 | iter time: 360.97 ms (step) remaining time: 0:12:38
251
+ Epoch 1 | iter 4352 step 136 | loss train: 1.434, val: 1.381 | iter time: 359.51 ms (step) remaining time: 0:12:26
252
+ Epoch 1 | iter 4384 step 137 | loss train: 1.520, val: 1.381 | iter time: 360.19 ms (step) remaining time: 0:12:15
253
+ Epoch 1 | iter 4416 step 138 | loss train: 1.460, val: 1.381 | iter time: 358.76 ms (step) remaining time: 0:12:03
254
+ Epoch 1 | iter 4448 step 139 | loss train: 1.441, val: 1.381 | iter time: 359.84 ms (step) remaining time: 0:11:52
255
+ Epoch 1 | iter 4480 step 140 | loss train: 1.412, val: 1.381 | iter time: 359.28 ms (step) remaining time: 0:11:40
256
+ Epoch 1 | iter 4512 step 141 | loss train: 1.409, val: 1.381 | iter time: 359.51 ms (step) remaining time: 0:11:29
257
+ Epoch 1 | iter 4544 step 142 | loss train: 1.470, val: 1.381 | iter time: 360.79 ms (step) remaining time: 0:11:17
258
+ Epoch 1 | iter 4576 step 143 | loss train: 1.502, val: 1.381 | iter time: 358.40 ms (step) remaining time: 0:11:06
259
+ Epoch 1 | iter 4608 step 144 | loss train: 1.418, val: 1.381 | iter time: 359.06 ms (step) remaining time: 0:10:54
260
+ Epoch 1 | iter 4640 step 145 | loss train: 1.382, val: 1.381 | iter time: 358.36 ms (step) remaining time: 0:10:43
261
+ Epoch 1 | iter 4672 step 146 | loss train: 1.505, val: 1.381 | iter time: 358.19 ms (step) remaining time: 0:10:31
262
+ Epoch 1 | iter 4704 step 147 | loss train: 1.473, val: 1.381 | iter time: 359.93 ms (step) remaining time: 0:10:20
263
+ Epoch 1 | iter 4736 step 148 | loss train: 1.456, val: 1.381 | iter time: 358.15 ms (step) remaining time: 0:10:08
264
+ Epoch 1 | iter 4768 step 149 | loss train: 1.362, val: 1.381 | iter time: 357.96 ms (step) remaining time: 0:09:57
265
+ Epoch 1 | iter 4800 step 150 | loss train: 1.491, val: 1.381 | iter time: 357.98 ms (step) remaining time: 0:09:45
266
+ Validating ...
267
+ iter 4800: val loss 1.3862, val time: 21884.89 ms
268
+ Epoch 1 | iter 4832 step 151 | loss train: 1.477, val: 1.386 | iter time: 361.41 ms (step) remaining time: 0:09:42
269
+ Epoch 1 | iter 4864 step 152 | loss train: 1.443, val: 1.386 | iter time: 359.29 ms (step) remaining time: 0:09:30
270
+ Epoch 1 | iter 4896 step 153 | loss train: 1.419, val: 1.386 | iter time: 360.99 ms (step) remaining time: 0:09:18
271
+ Epoch 1 | iter 4928 step 154 | loss train: 1.400, val: 1.386 | iter time: 359.78 ms (step) remaining time: 0:09:07
272
+ Epoch 1 | iter 4960 step 155 | loss train: 1.409, val: 1.386 | iter time: 358.89 ms (step) remaining time: 0:08:55
273
+ Epoch 1 | iter 4992 step 156 | loss train: 1.495, val: 1.386 | iter time: 359.49 ms (step) remaining time: 0:08:44
274
+ Epoch 1 | iter 5024 step 157 | loss train: 1.392, val: 1.386 | iter time: 360.13 ms (step) remaining time: 0:08:32
275
+ Epoch 1 | iter 5056 step 158 | loss train: 1.416, val: 1.386 | iter time: 360.78 ms (step) remaining time: 0:08:21
276
+ Epoch 1 | iter 5088 step 159 | loss train: 1.372, val: 1.386 | iter time: 359.35 ms (step) remaining time: 0:08:09
277
+ Epoch 1 | iter 5120 step 160 | loss train: 1.405, val: 1.386 | iter time: 360.12 ms (step) remaining time: 0:07:58
278
+ Epoch 1 | iter 5152 step 161 | loss train: 1.435, val: 1.386 | iter time: 357.74 ms (step) remaining time: 0:07:46
279
+ Epoch 1 | iter 5184 step 162 | loss train: 1.424, val: 1.386 | iter time: 360.51 ms (step) remaining time: 0:07:34
280
+ Epoch 1 | iter 5216 step 163 | loss train: 1.463, val: 1.386 | iter time: 357.19 ms (step) remaining time: 0:07:23
281
+ Epoch 1 | iter 5248 step 164 | loss train: 1.399, val: 1.386 | iter time: 359.01 ms (step) remaining time: 0:07:12
282
+ Epoch 1 | iter 5280 step 165 | loss train: 1.414, val: 1.386 | iter time: 359.56 ms (step) remaining time: 0:07:00
283
+ Epoch 1 | iter 5312 step 166 | loss train: 1.443, val: 1.386 | iter time: 358.55 ms (step) remaining time: 0:06:49
284
+ Epoch 1 | iter 5344 step 167 | loss train: 1.411, val: 1.386 | iter time: 359.22 ms (step) remaining time: 0:06:37
285
+ Epoch 1 | iter 5376 step 168 | loss train: 1.386, val: 1.386 | iter time: 359.22 ms (step) remaining time: 0:06:26
286
+ Epoch 1 | iter 5408 step 169 | loss train: 1.408, val: 1.386 | iter time: 359.34 ms (step) remaining time: 0:06:14
287
+ Epoch 1 | iter 5440 step 170 | loss train: 1.394, val: 1.386 | iter time: 359.45 ms (step) remaining time: 0:06:03
288
+ Epoch 1 | iter 5472 step 171 | loss train: 1.445, val: 1.386 | iter time: 360.94 ms (step) remaining time: 0:05:51
289
+ Epoch 1 | iter 5504 step 172 | loss train: 1.413, val: 1.386 | iter time: 364.07 ms (step) remaining time: 0:05:40
290
+ Epoch 1 | iter 5536 step 173 | loss train: 1.430, val: 1.386 | iter time: 360.39 ms (step) remaining time: 0:05:28
291
+ Epoch 1 | iter 5568 step 174 | loss train: 1.449, val: 1.386 | iter time: 358.68 ms (step) remaining time: 0:05:17
292
+ Epoch 1 | iter 5600 step 175 | loss train: 1.434, val: 1.386 | iter time: 358.88 ms (step) remaining time: 0:05:06
293
+ Epoch 1 | iter 5632 step 176 | loss train: 1.372, val: 1.386 | iter time: 358.92 ms (step) remaining time: 0:04:54
294
+ Epoch 1 | iter 5664 step 177 | loss train: 1.382, val: 1.386 | iter time: 358.70 ms (step) remaining time: 0:04:43
295
+ Epoch 1 | iter 5696 step 178 | loss train: 1.395, val: 1.386 | iter time: 359.08 ms (step) remaining time: 0:04:31
296
+ Epoch 1 | iter 5728 step 179 | loss train: 1.432, val: 1.386 | iter time: 360.66 ms (step) remaining time: 0:04:20
297
+ Epoch 1 | iter 5760 step 180 | loss train: 1.405, val: 1.386 | iter time: 357.47 ms (step) remaining time: 0:04:09
298
+ Epoch 1 | iter 5792 step 181 | loss train: 1.406, val: 1.386 | iter time: 359.14 ms (step) remaining time: 0:03:57
299
+ Epoch 1 | iter 5824 step 182 | loss train: 1.435, val: 1.386 | iter time: 360.04 ms (step) remaining time: 0:03:46
300
+ Epoch 1 | iter 5856 step 183 | loss train: 1.459, val: 1.386 | iter time: 359.13 ms (step) remaining time: 0:03:35
301
+ Epoch 1 | iter 5888 step 184 | loss train: 1.389, val: 1.386 | iter time: 358.51 ms (step) remaining time: 0:03:23
302
+ Epoch 1 | iter 5920 step 185 | loss train: 1.427, val: 1.386 | iter time: 359.62 ms (step) remaining time: 0:03:12
303
+ Epoch 1 | iter 5952 step 186 | loss train: 1.407, val: 1.386 | iter time: 358.62 ms (step) remaining time: 0:03:00
304
+ Epoch 1 | iter 5984 step 187 | loss train: 1.366, val: 1.386 | iter time: 359.50 ms (step) remaining time: 0:02:49
305
+ Epoch 1 | iter 6016 step 188 | loss train: 1.361, val: 1.386 | iter time: 358.08 ms (step) remaining time: 0:02:38
306
+ Epoch 1 | iter 6048 step 189 | loss train: 1.419, val: 1.386 | iter time: 359.68 ms (step) remaining time: 0:02:26
307
+ Epoch 1 | iter 6080 step 190 | loss train: 1.437, val: 1.386 | iter time: 359.31 ms (step) remaining time: 0:02:15
308
+ Epoch 1 | iter 6112 step 191 | loss train: 1.411, val: 1.386 | iter time: 359.74 ms (step) remaining time: 0:02:04
309
+ Epoch 1 | iter 6144 step 192 | loss train: 1.377, val: 1.386 | iter time: 358.37 ms (step) remaining time: 0:01:52
310
+ Epoch 1 | iter 6176 step 193 | loss train: 1.469, val: 1.386 | iter time: 359.70 ms (step) remaining time: 0:01:41
311
+ Epoch 1 | iter 6208 step 194 | loss train: 1.386, val: 1.386 | iter time: 588.60 ms (step) remaining time: 0:01:30
312
+ Epoch 1 | iter 6240 step 195 | loss train: 1.365, val: 1.386 | iter time: 358.18 ms (step) remaining time: 0:01:19
313
+ Epoch 1 | iter 6272 step 196 | loss train: 1.379, val: 1.386 | iter time: 358.97 ms (step) remaining time: 0:01:07
314
+ Epoch 1 | iter 6304 step 197 | loss train: 1.389, val: 1.386 | iter time: 358.96 ms (step) remaining time: 0:00:56
315
+ Epoch 1 | iter 6336 step 198 | loss train: 1.476, val: 1.386 | iter time: 359.36 ms (step) remaining time: 0:00:45
316
+ Epoch 1 | iter 6368 step 199 | loss train: 1.538, val: 1.386 | iter time: 359.78 ms (step) remaining time: 0:00:33
317
+ Epoch 1 | iter 6400 step 200 | loss train: 1.453, val: 1.386 | iter time: 356.84 ms (step) remaining time: 0:00:22
318
+ Validating ...
319
+ iter 6400: val loss 1.3940, val time: 21890.79 ms
320
+ Saving checkpoint to 'out/pretrain/tinyllama/2407_lr4e-5/step-00000200/lit_model.pth'
321
+ Epoch 1 | iter 6432 step 201 | loss train: 1.508, val: 1.394 | iter time: 356.12 ms (step) remaining time: 0:00:11
322
+ Epoch 2 | iter 6464 step 202 | loss train: 1.352, val: 1.394 | iter time: 357.03 ms (step) remaining time: 0:00:00
323
+ Validating ...
324
+ Final evaluation | val loss: 1.393 | val ppl: 4.026
325
+ Saving checkpoint to 'out/pretrain/tinyllama/2407_lr4e-5/final/lit_model.pth'
326
+ ----------------------------------------
327
+ | Performance
328
+ | - Total tokens : 211,812,352
329
+ | - Training Time : 2377.70 s
330
+ | - Tok/sec : 109.67 tok/s
331
+ | ----------------------------------------
332
+ | Memory Usage
333
+ | - Memory Used : 26.32 GB
334
+ ----------------------------------------
out/pretrain/tinyllama/teelogs/2408.txt ADDED
@@ -0,0 +1,288 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Initializing distributed: GLOBAL_RANK: 0, MEMBER: 1/4
2
+ Initializing distributed: GLOBAL_RANK: 3, MEMBER: 4/4
3
+ [rank: 3] Seed set to 42
4
+ Initializing distributed: GLOBAL_RANK: 1, MEMBER: 2/4
5
+ [rank: 1] Seed set to 42
6
+ Initializing distributed: GLOBAL_RANK: 2, MEMBER: 3/4
7
+ [rank: 2] Seed set to 42
8
+ ----------------------------------------------------------------------------------------------------
9
+ distributed_backend=nccl
10
+ All distributed processes registered. Starting with 4 processes
11
+ ----------------------------------------------------------------------------------------------------
12
+
13
+ [rank: 0] Seed set to 42
14
+ All GPUs are fully connected via NVLink.
15
+ {'data': {'batch_size': 1,
16
+ 'data_path': PosixPath('litgpt/data/arxiv'),
17
+ 'num_workers': 8,
18
+ 'seed': 42,
19
+ 'seq_length': 2048,
20
+ 'use_starcoder': True},
21
+ 'data_dir': PosixPath('litgpt/data/arxiv/2408'),
22
+ 'devices': 'auto',
23
+ 'eval': {'evaluate_example': 'first',
24
+ 'final_validation': True,
25
+ 'initial_validation': True,
26
+ 'interval': 50,
27
+ 'max_iters': 100,
28
+ 'max_new_tokens': None},
29
+ 'initial_checkpoint_dir': PosixPath('out/pretrain/2407/final'),
30
+ 'log': {'group': None, 'project': None, 'run': None},
31
+ 'logger_name': 'tensorboard',
32
+ 'model_config': {'attention_logit_softcapping': None,
33
+ 'attention_scores_scalar': None,
34
+ 'attn_bias': False,
35
+ 'bias': False,
36
+ 'block_size': 2048,
37
+ 'final_logit_softcapping': None,
38
+ 'gelu_approximate': 'none',
39
+ 'head_size': 64,
40
+ 'hf_config': {'name': 'TinyLlama-1.1B-intermediate-step-1431k-3T',
41
+ 'org': 'TinyLlama'},
42
+ 'intermediate_size': 5632,
43
+ 'lm_head_bias': False,
44
+ 'mlp_class_name': 'LLaMAMLP',
45
+ 'moe_intermediate_size': None,
46
+ 'n_embd': 2048,
47
+ 'n_expert': 0,
48
+ 'n_expert_per_token': 0,
49
+ 'n_head': 32,
50
+ 'n_layer': 22,
51
+ 'n_query_groups': 4,
52
+ 'name': 'tiny-llama-1.1b',
53
+ 'norm_1': True,
54
+ 'norm_2': True,
55
+ 'norm_class_name': 'RMSNorm',
56
+ 'norm_eps': 1e-05,
57
+ 'norm_qk': False,
58
+ 'norm_qk_type': 'default',
59
+ 'padded_vocab_size': 32000,
60
+ 'padding_multiple': 64,
61
+ 'parallel_residual': False,
62
+ 'post_attention_norm': False,
63
+ 'post_mlp_norm': False,
64
+ 'rope_adjustments': None,
65
+ 'rope_base': 10000,
66
+ 'rope_condense_ratio': 1,
67
+ 'rope_indices': None,
68
+ 'rope_local_base_freq': None,
69
+ 'rotary_percentage': 1.0,
70
+ 'scale_embeddings': False,
71
+ 'shared_attention_norm': False,
72
+ 'sliding_window_indices': None,
73
+ 'sliding_window_size': None,
74
+ 'vocab_size': 32000},
75
+ 'model_name': 'tiny-llama-1.1b',
76
+ 'num_nodes': 1,
77
+ 'optimizer': "{'class_path': 'torch.optim.AdamW', 'init_args': {'lr': 0.0004, "
78
+ "'weight_decay': 0.1, 'betas': [0.9, 0.95]}}",
79
+ 'out_dir': PosixPath('out/pretrain/2408'),
80
+ 'precision': 'bf16-mixed',
81
+ 'resume': False,
82
+ 'seed': 42,
83
+ 'tokenizer_dir': PosixPath('checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T'),
84
+ 'train': {'epochs': None,
85
+ 'global_batch_size': 512,
86
+ 'log_interval': 1,
87
+ 'lr_warmup_fraction': None,
88
+ 'lr_warmup_steps': 20,
89
+ 'max_norm': 1.0,
90
+ 'max_seq_length': 2048,
91
+ 'max_steps': None,
92
+ 'max_tokens': 176160768,
93
+ 'micro_batch_size': 4,
94
+ 'min_lr': 4e-05,
95
+ 'save_interval': 100,
96
+ 'tie_embeddings': None}}
97
+ Time to instantiate model: 0.02 seconds.
98
+ Total parameters: 1,100,048,384
99
+ [ok] out/pretrain/2407/final/lit_model.pth 已是纯权重
100
+ Validating ...
101
+ Measured TFLOPs: 239.66
102
+ Epoch 1 | iter 32 step 1 | loss train: 1.454, val: 1.313 | iter time: 545.49 ms (step) remaining time: 0:33:04
103
+ Epoch 1 | iter 64 step 2 | loss train: 1.463, val: 1.313 | iter time: 355.62 ms (step) remaining time: 0:31:19
104
+ Epoch 1 | iter 96 step 3 | loss train: 1.483, val: 1.313 | iter time: 355.27 ms (step) remaining time: 0:30:37
105
+ Epoch 1 | iter 128 step 4 | loss train: 1.514, val: 1.313 | iter time: 359.19 ms (step) remaining time: 0:30:11
106
+ Epoch 1 | iter 160 step 5 | loss train: 1.452, val: 1.313 | iter time: 357.38 ms (step) remaining time: 0:29:52
107
+ Epoch 1 | iter 192 step 6 | loss train: 1.396, val: 1.313 | iter time: 358.99 ms (step) remaining time: 0:29:36
108
+ Epoch 1 | iter 224 step 7 | loss train: 1.378, val: 1.313 | iter time: 357.63 ms (step) remaining time: 0:29:21
109
+ Epoch 1 | iter 256 step 8 | loss train: 1.391, val: 1.313 | iter time: 360.11 ms (step) remaining time: 0:29:08
110
+ Epoch 1 | iter 288 step 9 | loss train: 1.457, val: 1.313 | iter time: 358.34 ms (step) remaining time: 0:28:55
111
+ Epoch 1 | iter 320 step 10 | loss train: 1.460, val: 1.313 | iter time: 358.68 ms (step) remaining time: 0:28:43
112
+ Epoch 1 | iter 352 step 11 | loss train: 1.493, val: 1.313 | iter time: 358.48 ms (step) remaining time: 0:28:32
113
+ Epoch 1 | iter 384 step 12 | loss train: 1.494, val: 1.313 | iter time: 358.84 ms (step) remaining time: 0:28:20
114
+ Epoch 1 | iter 416 step 13 | loss train: 1.476, val: 1.313 | iter time: 360.24 ms (step) remaining time: 0:28:09
115
+ Epoch 1 | iter 448 step 14 | loss train: 1.484, val: 1.313 | iter time: 359.47 ms (step) remaining time: 0:27:57
116
+ Epoch 1 | iter 480 step 15 | loss train: 1.495, val: 1.313 | iter time: 357.80 ms (step) remaining time: 0:27:46
117
+ Epoch 1 | iter 512 step 16 | loss train: 1.462, val: 1.313 | iter time: 358.26 ms (step) remaining time: 0:27:35
118
+ Epoch 1 | iter 544 step 17 | loss train: 1.520, val: 1.313 | iter time: 358.70 ms (step) remaining time: 0:27:24
119
+ Epoch 1 | iter 576 step 18 | loss train: 1.462, val: 1.313 | iter time: 359.21 ms (step) remaining time: 0:27:13
120
+ Epoch 1 | iter 608 step 19 | loss train: 1.503, val: 1.313 | iter time: 360.84 ms (step) remaining time: 0:27:02
121
+ Epoch 1 | iter 640 step 20 | loss train: 1.516, val: 1.313 | iter time: 358.59 ms (step) remaining time: 0:26:51
122
+ Epoch 1 | iter 672 step 21 | loss train: 1.550, val: 1.313 | iter time: 358.53 ms (step) remaining time: 0:26:40
123
+ Epoch 1 | iter 704 step 22 | loss train: 1.521, val: 1.313 | iter time: 358.48 ms (step) remaining time: 0:26:29
124
+ Epoch 1 | iter 736 step 23 | loss train: 1.507, val: 1.313 | iter time: 361.00 ms (step) remaining time: 0:26:18
125
+ Epoch 1 | iter 768 step 24 | loss train: 1.496, val: 1.313 | iter time: 359.17 ms (step) remaining time: 0:26:07
126
+ Epoch 1 | iter 800 step 25 | loss train: 1.516, val: 1.313 | iter time: 358.59 ms (step) remaining time: 0:25:56
127
+ Epoch 1 | iter 832 step 26 | loss train: 1.479, val: 1.313 | iter time: 358.94 ms (step) remaining time: 0:25:45
128
+ Epoch 1 | iter 864 step 27 | loss train: 1.520, val: 1.313 | iter time: 359.95 ms (step) remaining time: 0:25:34
129
+ Epoch 1 | iter 896 step 28 | loss train: 1.584, val: 1.313 | iter time: 358.97 ms (step) remaining time: 0:25:23
130
+ Epoch 1 | iter 928 step 29 | loss train: 1.453, val: 1.313 | iter time: 358.05 ms (step) remaining time: 0:25:12
131
+ Epoch 1 | iter 960 step 30 | loss train: 1.601, val: 1.313 | iter time: 360.72 ms (step) remaining time: 0:25:01
132
+ Epoch 1 | iter 992 step 31 | loss train: 1.561, val: 1.313 | iter time: 361.10 ms (step) remaining time: 0:24:50
133
+ Epoch 1 | iter 1024 step 32 | loss train: 1.432, val: 1.313 | iter time: 358.49 ms (step) remaining time: 0:24:39
134
+ Epoch 1 | iter 1056 step 33 | loss train: 1.495, val: 1.313 | iter time: 360.98 ms (step) remaining time: 0:24:28
135
+ Epoch 1 | iter 1088 step 34 | loss train: 1.560, val: 1.313 | iter time: 358.22 ms (step) remaining time: 0:24:18
136
+ Epoch 1 | iter 1120 step 35 | loss train: 1.490, val: 1.313 | iter time: 358.52 ms (step) remaining time: 0:24:07
137
+ Epoch 1 | iter 1152 step 36 | loss train: 1.493, val: 1.313 | iter time: 360.04 ms (step) remaining time: 0:23:56
138
+ Epoch 1 | iter 1184 step 37 | loss train: 1.412, val: 1.313 | iter time: 357.77 ms (step) remaining time: 0:23:45
139
+ Epoch 1 | iter 1216 step 38 | loss train: 1.451, val: 1.313 | iter time: 360.18 ms (step) remaining time: 0:23:34
140
+ Epoch 1 | iter 1248 step 39 | loss train: 1.396, val: 1.313 | iter time: 361.18 ms (step) remaining time: 0:23:24
141
+ Epoch 1 | iter 1280 step 40 | loss train: 1.498, val: 1.313 | iter time: 360.85 ms (step) remaining time: 0:23:13
142
+ Epoch 1 | iter 1312 step 41 | loss train: 1.472, val: 1.313 | iter time: 360.48 ms (step) remaining time: 0:23:02
143
+ Epoch 1 | iter 1344 step 42 | loss train: 1.481, val: 1.313 | iter time: 359.84 ms (step) remaining time: 0:22:52
144
+ Epoch 1 | iter 1376 step 43 | loss train: 1.498, val: 1.313 | iter time: 360.22 ms (step) remaining time: 0:22:41
145
+ Epoch 1 | iter 1408 step 44 | loss train: 1.439, val: 1.313 | iter time: 359.80 ms (step) remaining time: 0:22:30
146
+ Epoch 1 | iter 1440 step 45 | loss train: 1.463, val: 1.313 | iter time: 359.85 ms (step) remaining time: 0:22:19
147
+ Epoch 1 | iter 1472 step 46 | loss train: 1.489, val: 1.313 | iter time: 360.05 ms (step) remaining time: 0:22:08
148
+ Epoch 1 | iter 1504 step 47 | loss train: 1.545, val: 1.313 | iter time: 359.81 ms (step) remaining time: 0:21:57
149
+ Epoch 1 | iter 1536 step 48 | loss train: 1.525, val: 1.313 | iter time: 359.03 ms (step) remaining time: 0:21:46
150
+ Epoch 1 | iter 1568 step 49 | loss train: 1.371, val: 1.313 | iter time: 358.91 ms (step) remaining time: 0:21:35
151
+ Epoch 1 | iter 1600 step 50 | loss train: 1.438, val: 1.313 | iter time: 360.97 ms (step) remaining time: 0:21:24
152
+ Validating ...
153
+ iter 1600: val loss 1.3687, val time: 5910.61 ms
154
+ Epoch 1 | iter 1632 step 51 | loss train: 1.563, val: 1.369 | iter time: 358.07 ms (step) remaining time: 0:21:27
155
+ Epoch 1 | iter 1664 step 52 | loss train: 1.511, val: 1.369 | iter time: 359.70 ms (step) remaining time: 0:21:15
156
+ Epoch 1 | iter 1696 step 53 | loss train: 1.454, val: 1.369 | iter time: 359.04 ms (step) remaining time: 0:21:04
157
+ Epoch 1 | iter 1728 step 54 | loss train: 1.441, val: 1.369 | iter time: 361.51 ms (step) remaining time: 0:20:53
158
+ Epoch 1 | iter 1760 step 55 | loss train: 1.495, val: 1.369 | iter time: 358.30 ms (step) remaining time: 0:20:42
159
+ Epoch 1 | iter 1792 step 56 | loss train: 1.546, val: 1.369 | iter time: 360.41 ms (step) remaining time: 0:20:30
160
+ Epoch 1 | iter 1824 step 57 | loss train: 1.446, val: 1.369 | iter time: 360.86 ms (step) remaining time: 0:20:19
161
+ Epoch 1 | iter 1856 step 58 | loss train: 1.440, val: 1.369 | iter time: 361.60 ms (step) remaining time: 0:20:08
162
+ Epoch 1 | iter 1888 step 59 | loss train: 1.455, val: 1.369 | iter time: 360.70 ms (step) remaining time: 0:19:57
163
+ Epoch 1 | iter 1920 step 60 | loss train: 1.493, val: 1.369 | iter time: 360.65 ms (step) remaining time: 0:19:45
164
+ Epoch 1 | iter 1952 step 61 | loss train: 1.385, val: 1.369 | iter time: 360.22 ms (step) remaining time: 0:19:34
165
+ Epoch 1 | iter 1984 step 62 | loss train: 1.443, val: 1.369 | iter time: 360.32 ms (step) remaining time: 0:19:23
166
+ Epoch 1 | iter 2016 step 63 | loss train: 1.404, val: 1.369 | iter time: 359.78 ms (step) remaining time: 0:19:12
167
+ Epoch 1 | iter 2048 step 64 | loss train: 1.460, val: 1.369 | iter time: 360.64 ms (step) remaining time: 0:19:01
168
+ Epoch 1 | iter 2080 step 65 | loss train: 1.474, val: 1.369 | iter time: 359.02 ms (step) remaining time: 0:18:50
169
+ Epoch 1 | iter 2112 step 66 | loss train: 1.486, val: 1.369 | iter time: 360.39 ms (step) remaining time: 0:18:39
170
+ Epoch 1 | iter 2144 step 67 | loss train: 1.451, val: 1.369 | iter time: 358.92 ms (step) remaining time: 0:18:27
171
+ Epoch 1 | iter 2176 step 68 | loss train: 1.423, val: 1.369 | iter time: 358.20 ms (step) remaining time: 0:18:16
172
+ Epoch 1 | iter 2208 step 69 | loss train: 1.385, val: 1.369 | iter time: 359.70 ms (step) remaining time: 0:18:05
173
+ Epoch 1 | iter 2240 step 70 | loss train: 1.494, val: 1.369 | iter time: 361.18 ms (step) remaining time: 0:17:54
174
+ Epoch 1 | iter 2272 step 71 | loss train: 1.427, val: 1.369 | iter time: 358.37 ms (step) remaining time: 0:17:43
175
+ Epoch 1 | iter 2304 step 72 | loss train: 1.461, val: 1.369 | iter time: 360.16 ms (step) remaining time: 0:17:32
176
+ Epoch 1 | iter 2336 step 73 | loss train: 1.462, val: 1.369 | iter time: 358.16 ms (step) remaining time: 0:17:21
177
+ Epoch 1 | iter 2368 step 74 | loss train: 1.428, val: 1.369 | iter time: 361.77 ms (step) remaining time: 0:17:10
178
+ Epoch 1 | iter 2400 step 75 | loss train: 1.473, val: 1.369 | iter time: 359.25 ms (step) remaining time: 0:16:59
179
+ Epoch 1 | iter 2432 step 76 | loss train: 1.452, val: 1.369 | iter time: 358.95 ms (step) remaining time: 0:16:48
180
+ Epoch 1 | iter 2464 step 77 | loss train: 1.437, val: 1.369 | iter time: 359.74 ms (step) remaining time: 0:16:36
181
+ Epoch 1 | iter 2496 step 78 | loss train: 1.412, val: 1.369 | iter time: 357.93 ms (step) remaining time: 0:16:25
182
+ Epoch 1 | iter 2528 step 79 | loss train: 1.410, val: 1.369 | iter time: 359.52 ms (step) remaining time: 0:16:14
183
+ Epoch 1 | iter 2560 step 80 | loss train: 1.495, val: 1.369 | iter time: 360.06 ms (step) remaining time: 0:16:03
184
+ Epoch 1 | iter 2592 step 81 | loss train: 1.445, val: 1.369 | iter time: 359.96 ms (step) remaining time: 0:15:52
185
+ Epoch 1 | iter 2624 step 82 | loss train: 1.395, val: 1.369 | iter time: 361.05 ms (step) remaining time: 0:15:41
186
+ Epoch 1 | iter 2656 step 83 | loss train: 1.394, val: 1.369 | iter time: 359.52 ms (step) remaining time: 0:15:30
187
+ Epoch 1 | iter 2688 step 84 | loss train: 1.393, val: 1.369 | iter time: 359.32 ms (step) remaining time: 0:15:19
188
+ Epoch 1 | iter 2720 step 85 | loss train: 1.453, val: 1.369 | iter time: 358.57 ms (step) remaining time: 0:15:08
189
+ Epoch 1 | iter 2752 step 86 | loss train: 1.357, val: 1.369 | iter time: 359.51 ms (step) remaining time: 0:14:57
190
+ Epoch 1 | iter 2784 step 87 | loss train: 1.437, val: 1.369 | iter time: 360.38 ms (step) remaining time: 0:14:46
191
+ Epoch 1 | iter 2816 step 88 | loss train: 1.412, val: 1.369 | iter time: 360.65 ms (step) remaining time: 0:14:35
192
+ Epoch 1 | iter 2848 step 89 | loss train: 1.428, val: 1.369 | iter time: 360.50 ms (step) remaining time: 0:14:24
193
+ Epoch 1 | iter 2880 step 90 | loss train: 1.461, val: 1.369 | iter time: 359.97 ms (step) remaining time: 0:14:13
194
+ Epoch 1 | iter 2912 step 91 | loss train: 1.411, val: 1.369 | iter time: 359.61 ms (step) remaining time: 0:14:02
195
+ Epoch 1 | iter 2944 step 92 | loss train: 1.393, val: 1.369 | iter time: 361.39 ms (step) remaining time: 0:13:51
196
+ Epoch 1 | iter 2976 step 93 | loss train: 1.489, val: 1.369 | iter time: 359.82 ms (step) remaining time: 0:13:40
197
+ Epoch 1 | iter 3008 step 94 | loss train: 1.439, val: 1.369 | iter time: 357.29 ms (step) remaining time: 0:13:29
198
+ Epoch 1 | iter 3040 step 95 | loss train: 1.472, val: 1.369 | iter time: 359.77 ms (step) remaining time: 0:13:18
199
+ Epoch 1 | iter 3072 step 96 | loss train: 1.467, val: 1.369 | iter time: 359.36 ms (step) remaining time: 0:13:07
200
+ Epoch 1 | iter 3104 step 97 | loss train: 1.438, val: 1.369 | iter time: 360.19 ms (step) remaining time: 0:12:56
201
+ Epoch 1 | iter 3136 step 98 | loss train: 1.414, val: 1.369 | iter time: 360.76 ms (step) remaining time: 0:12:45
202
+ Epoch 1 | iter 3168 step 99 | loss train: 1.436, val: 1.369 | iter time: 359.83 ms (step) remaining time: 0:12:34
203
+ Epoch 1 | iter 3200 step 100 | loss train: 1.443, val: 1.369 | iter time: 358.64 ms (step) remaining time: 0:12:23
204
+ Validating ...
205
+ iter 3200: val loss 1.3403, val time: 5880.97 ms
206
+ Saving checkpoint to 'out/pretrain/2408/step-00000100/lit_model.pth'
207
+ Epoch 1 | iter 3232 step 101 | loss train: 1.447, val: 1.340 | iter time: 354.20 ms (step) remaining time: 0:13:14
208
+ Epoch 1 | iter 3264 step 102 | loss train: 1.383, val: 1.340 | iter time: 356.96 ms (step) remaining time: 0:13:01
209
+ Epoch 1 | iter 3296 step 103 | loss train: 1.363, val: 1.340 | iter time: 355.70 ms (step) remaining time: 0:12:49
210
+ Epoch 1 | iter 3328 step 104 | loss train: 1.410, val: 1.340 | iter time: 357.99 ms (step) remaining time: 0:12:36
211
+ Epoch 1 | iter 3360 step 105 | loss train: 1.436, val: 1.340 | iter time: 359.20 ms (step) remaining time: 0:12:24
212
+ Epoch 1 | iter 3392 step 106 | loss train: 1.379, val: 1.340 | iter time: 360.20 ms (step) remaining time: 0:12:12
213
+ Epoch 1 | iter 3424 step 107 | loss train: 1.364, val: 1.340 | iter time: 357.29 ms (step) remaining time: 0:11:59
214
+ Epoch 1 | iter 3456 step 108 | loss train: 1.400, val: 1.340 | iter time: 358.10 ms (step) remaining time: 0:11:47
215
+ Epoch 1 | iter 3488 step 109 | loss train: 1.430, val: 1.340 | iter time: 358.16 ms (step) remaining time: 0:11:35
216
+ Epoch 1 | iter 3520 step 110 | loss train: 1.424, val: 1.340 | iter time: 358.76 ms (step) remaining time: 0:11:22
217
+ Epoch 1 | iter 3552 step 111 | loss train: 1.359, val: 1.340 | iter time: 361.14 ms (step) remaining time: 0:11:10
218
+ Epoch 1 | iter 3584 step 112 | loss train: 1.386, val: 1.340 | iter time: 360.25 ms (step) remaining time: 0:10:58
219
+ Epoch 1 | iter 3616 step 113 | loss train: 1.394, val: 1.340 | iter time: 359.48 ms (step) remaining time: 0:10:46
220
+ Epoch 1 | iter 3648 step 114 | loss train: 1.402, val: 1.340 | iter time: 361.39 ms (step) remaining time: 0:10:33
221
+ Epoch 1 | iter 3680 step 115 | loss train: 1.436, val: 1.340 | iter time: 362.00 ms (step) remaining time: 0:10:21
222
+ Epoch 1 | iter 3712 step 116 | loss train: 1.336, val: 1.340 | iter time: 359.77 ms (step) remaining time: 0:10:09
223
+ Epoch 1 | iter 3744 step 117 | loss train: 1.396, val: 1.340 | iter time: 362.14 ms (step) remaining time: 0:09:57
224
+ Epoch 1 | iter 3776 step 118 | loss train: 1.416, val: 1.340 | iter time: 359.51 ms (step) remaining time: 0:09:45
225
+ Epoch 1 | iter 3808 step 119 | loss train: 1.445, val: 1.340 | iter time: 359.81 ms (step) remaining time: 0:09:33
226
+ Epoch 1 | iter 3840 step 120 | loss train: 1.436, val: 1.340 | iter time: 360.43 ms (step) remaining time: 0:09:21
227
+ Epoch 1 | iter 3872 step 121 | loss train: 1.445, val: 1.340 | iter time: 359.62 ms (step) remaining time: 0:09:09
228
+ Epoch 1 | iter 3904 step 122 | loss train: 1.398, val: 1.340 | iter time: 360.20 ms (step) remaining time: 0:08:57
229
+ Epoch 1 | iter 3936 step 123 | loss train: 1.355, val: 1.340 | iter time: 360.86 ms (step) remaining time: 0:08:45
230
+ Epoch 1 | iter 3968 step 124 | loss train: 1.390, val: 1.340 | iter time: 358.78 ms (step) remaining time: 0:08:33
231
+ Epoch 1 | iter 4000 step 125 | loss train: 1.371, val: 1.340 | iter time: 362.27 ms (step) remaining time: 0:08:21
232
+ Epoch 1 | iter 4032 step 126 | loss train: 1.358, val: 1.340 | iter time: 357.21 ms (step) remaining time: 0:08:09
233
+ Epoch 1 | iter 4064 step 127 | loss train: 1.437, val: 1.340 | iter time: 359.92 ms (step) remaining time: 0:07:57
234
+ Epoch 1 | iter 4096 step 128 | loss train: 1.389, val: 1.340 | iter time: 360.57 ms (step) remaining time: 0:07:45
235
+ Epoch 1 | iter 4128 step 129 | loss train: 1.334, val: 1.340 | iter time: 360.53 ms (step) remaining time: 0:07:33
236
+ Epoch 1 | iter 4160 step 130 | loss train: 1.384, val: 1.340 | iter time: 358.85 ms (step) remaining time: 0:07:21
237
+ Epoch 1 | iter 4192 step 131 | loss train: 1.393, val: 1.340 | iter time: 360.78 ms (step) remaining time: 0:07:10
238
+ Epoch 1 | iter 4224 step 132 | loss train: 1.331, val: 1.340 | iter time: 360.15 ms (step) remaining time: 0:06:58
239
+ Epoch 1 | iter 4256 step 133 | loss train: 1.416, val: 1.340 | iter time: 361.20 ms (step) remaining time: 0:06:46
240
+ Epoch 1 | iter 4288 step 134 | loss train: 1.366, val: 1.340 | iter time: 360.69 ms (step) remaining time: 0:06:34
241
+ Epoch 1 | iter 4320 step 135 | loss train: 1.388, val: 1.340 | iter time: 360.74 ms (step) remaining time: 0:06:22
242
+ Epoch 1 | iter 4352 step 136 | loss train: 1.414, val: 1.340 | iter time: 360.34 ms (step) remaining time: 0:06:11
243
+ Epoch 1 | iter 4384 step 137 | loss train: 1.400, val: 1.340 | iter time: 360.01 ms (step) remaining time: 0:05:59
244
+ Epoch 1 | iter 4416 step 138 | loss train: 1.349, val: 1.340 | iter time: 359.45 ms (step) remaining time: 0:05:47
245
+ Epoch 1 | iter 4448 step 139 | loss train: 1.368, val: 1.340 | iter time: 361.23 ms (step) remaining time: 0:05:35
246
+ Epoch 1 | iter 4480 step 140 | loss train: 1.416, val: 1.340 | iter time: 359.25 ms (step) remaining time: 0:05:24
247
+ Epoch 1 | iter 4512 step 141 | loss train: 1.440, val: 1.340 | iter time: 360.27 ms (step) remaining time: 0:05:12
248
+ Epoch 1 | iter 4544 step 142 | loss train: 1.355, val: 1.340 | iter time: 360.02 ms (step) remaining time: 0:05:00
249
+ Epoch 1 | iter 4576 step 143 | loss train: 1.430, val: 1.340 | iter time: 360.39 ms (step) remaining time: 0:04:49
250
+ Epoch 1 | iter 4608 step 144 | loss train: 1.388, val: 1.340 | iter time: 357.89 ms (step) remaining time: 0:04:37
251
+ Epoch 1 | iter 4640 step 145 | loss train: 1.415, val: 1.340 | iter time: 361.73 ms (step) remaining time: 0:04:25
252
+ Epoch 1 | iter 4672 step 146 | loss train: 1.361, val: 1.340 | iter time: 361.89 ms (step) remaining time: 0:04:14
253
+ Epoch 1 | iter 4704 step 147 | loss train: 1.462, val: 1.340 | iter time: 360.22 ms (step) remaining time: 0:04:02
254
+ Epoch 1 | iter 4736 step 148 | loss train: 1.332, val: 1.340 | iter time: 358.17 ms (step) remaining time: 0:03:50
255
+ Epoch 1 | iter 4768 step 149 | loss train: 1.426, val: 1.340 | iter time: 359.78 ms (step) remaining time: 0:03:39
256
+ Epoch 1 | iter 4800 step 150 | loss train: 1.396, val: 1.340 | iter time: 358.87 ms (step) remaining time: 0:03:27
257
+ Validating ...
258
+ iter 4800: val loss 1.3124, val time: 5894.65 ms
259
+ Epoch 1 | iter 4832 step 151 | loss train: 1.331, val: 1.312 | iter time: 359.88 ms (step) remaining time: 0:03:16
260
+ Epoch 1 | iter 4864 step 152 | loss train: 1.366, val: 1.312 | iter time: 357.72 ms (step) remaining time: 0:03:05
261
+ Epoch 1 | iter 4896 step 153 | loss train: 1.350, val: 1.312 | iter time: 359.10 ms (step) remaining time: 0:02:53
262
+ Epoch 1 | iter 4928 step 154 | loss train: 1.319, val: 1.312 | iter time: 360.79 ms (step) remaining time: 0:02:41
263
+ Epoch 1 | iter 4960 step 155 | loss train: 1.384, val: 1.312 | iter time: 361.43 ms (step) remaining time: 0:02:30
264
+ Epoch 1 | iter 4992 step 156 | loss train: 1.324, val: 1.312 | iter time: 360.17 ms (step) remaining time: 0:02:18
265
+ Epoch 1 | iter 5024 step 157 | loss train: 1.419, val: 1.312 | iter time: 360.86 ms (step) remaining time: 0:02:06
266
+ Epoch 1 | iter 5056 step 158 | loss train: 1.291, val: 1.312 | iter time: 360.09 ms (step) remaining time: 0:01:55
267
+ Epoch 1 | iter 5088 step 159 | loss train: 1.428, val: 1.312 | iter time: 359.59 ms (step) remaining time: 0:01:43
268
+ Epoch 1 | iter 5120 step 160 | loss train: 1.356, val: 1.312 | iter time: 360.29 ms (step) remaining time: 0:01:32
269
+ Epoch 1 | iter 5152 step 161 | loss train: 1.402, val: 1.312 | iter time: 360.16 ms (step) remaining time: 0:01:20
270
+ Epoch 1 | iter 5184 step 162 | loss train: 1.347, val: 1.312 | iter time: 361.50 ms (step) remaining time: 0:01:09
271
+ Epoch 1 | iter 5216 step 163 | loss train: 1.360, val: 1.312 | iter time: 359.61 ms (step) remaining time: 0:00:57
272
+ Epoch 1 | iter 5248 step 164 | loss train: 1.400, val: 1.312 | iter time: 359.27 ms (step) remaining time: 0:00:46
273
+ Epoch 1 | iter 5280 step 165 | loss train: 1.421, val: 1.312 | iter time: 358.93 ms (step) remaining time: 0:00:34
274
+ Epoch 1 | iter 5312 step 166 | loss train: 1.320, val: 1.312 | iter time: 358.83 ms (step) remaining time: 0:00:23
275
+ Epoch 1 | iter 5344 step 167 | loss train: 1.381, val: 1.312 | iter time: 361.81 ms (step) remaining time: 0:00:11
276
+ Epoch 1 | iter 5376 step 168 | loss train: 1.324, val: 1.312 | iter time: 360.58 ms (step) remaining time: 0:00:00
277
+ Validating ...
278
+ Final evaluation | val loss: 1.308 | val ppl: 3.698
279
+ Saving checkpoint to 'out/pretrain/2408/final/lit_model.pth'
280
+ ----------------------------------------
281
+ | Performance
282
+ | - Total tokens : 176,160,768
283
+ | - Training Time : 2032.04 s
284
+ | - Tok/sec : 164.71 tok/s
285
+ | ----------------------------------------
286
+ | Memory Usage
287
+ | - Memory Used : 26.32 GB
288
+ ----------------------------------------
out/pretrain/tinyllama/teelogs/2408_full.txt ADDED
@@ -0,0 +1,291 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Initializing distributed: GLOBAL_RANK: 0, MEMBER: 1/4
2
+ Initializing distributed: GLOBAL_RANK: 3, MEMBER: 4/4
3
+ [rank: 3] Seed set to 42
4
+ Initializing distributed: GLOBAL_RANK: 2, MEMBER: 3/4
5
+ [rank: 2] Seed set to 42
6
+ Initializing distributed: GLOBAL_RANK: 1, MEMBER: 2/4
7
+ [rank: 1] Seed set to 42
8
+ ----------------------------------------------------------------------------------------------------
9
+ distributed_backend=nccl
10
+ All distributed processes registered. Starting with 4 processes
11
+ ----------------------------------------------------------------------------------------------------
12
+
13
+ [rank: 0] Seed set to 42
14
+ All GPUs are fully connected via NVLink.
15
+ {'data': {'batch_size': 1,
16
+ 'data_path': PosixPath('litgpt/data/arxiv'),
17
+ 'num_workers': 8,
18
+ 'seed': 42,
19
+ 'seq_length': 2048,
20
+ 'use_starcoder': True},
21
+ 'data_dir': PosixPath('litgpt/data/arxiv/2408'),
22
+ 'devices': 'auto',
23
+ 'eval': {'evaluate_example': 'first',
24
+ 'final_validation': True,
25
+ 'initial_validation': True,
26
+ 'interval': 50,
27
+ 'max_iters': 200,
28
+ 'max_new_tokens': None},
29
+ 'initial_checkpoint_dir': PosixPath('out/pretrain/2407_full/final'),
30
+ 'log': {'group': None, 'project': None, 'run': None},
31
+ 'logger_name': 'tensorboard',
32
+ 'model_config': {'attention_logit_softcapping': None,
33
+ 'attention_scores_scalar': None,
34
+ 'attn_bias': False,
35
+ 'bias': False,
36
+ 'block_size': 2048,
37
+ 'final_logit_softcapping': None,
38
+ 'gelu_approximate': 'none',
39
+ 'head_size': 64,
40
+ 'hf_config': {'name': 'TinyLlama-1.1B-intermediate-step-1431k-3T',
41
+ 'org': 'TinyLlama'},
42
+ 'intermediate_size': 5632,
43
+ 'lm_head_bias': False,
44
+ 'mlp_class_name': 'LLaMAMLP',
45
+ 'moe_intermediate_size': None,
46
+ 'n_embd': 2048,
47
+ 'n_expert': 0,
48
+ 'n_expert_per_token': 0,
49
+ 'n_head': 32,
50
+ 'n_layer': 22,
51
+ 'n_query_groups': 4,
52
+ 'name': 'tiny-llama-1.1b',
53
+ 'norm_1': True,
54
+ 'norm_2': True,
55
+ 'norm_class_name': 'RMSNorm',
56
+ 'norm_eps': 1e-05,
57
+ 'norm_qk': False,
58
+ 'norm_qk_type': 'default',
59
+ 'padded_vocab_size': 32000,
60
+ 'padding_multiple': 64,
61
+ 'parallel_residual': False,
62
+ 'post_attention_norm': False,
63
+ 'post_mlp_norm': False,
64
+ 'rope_adjustments': None,
65
+ 'rope_base': 10000,
66
+ 'rope_condense_ratio': 1,
67
+ 'rope_indices': None,
68
+ 'rope_local_base_freq': None,
69
+ 'rotary_percentage': 1.0,
70
+ 'scale_embeddings': False,
71
+ 'shared_attention_norm': False,
72
+ 'sliding_window_indices': None,
73
+ 'sliding_window_size': None,
74
+ 'vocab_size': 32000},
75
+ 'model_name': 'tiny-llama-1.1b',
76
+ 'num_nodes': 1,
77
+ 'optimizer': "{'class_path': 'torch.optim.AdamW', 'init_args': {'lr': 0.0004, "
78
+ "'weight_decay': 0.1, 'betas': [0.9, 0.95]}}",
79
+ 'out_dir': PosixPath('out/pretrain/2408_full'),
80
+ 'ppl': False,
81
+ 'precision': 'bf16-mixed',
82
+ 'resume': False,
83
+ 'seed': 42,
84
+ 'tokenizer_dir': PosixPath('checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T'),
85
+ 'train': {'epochs': None,
86
+ 'global_batch_size': 512,
87
+ 'log_interval': 1,
88
+ 'lr_warmup_fraction': None,
89
+ 'lr_warmup_steps': 20,
90
+ 'max_norm': 1.0,
91
+ 'max_seq_length': 2048,
92
+ 'max_steps': None,
93
+ 'max_tokens': 177209344,
94
+ 'micro_batch_size': 4,
95
+ 'min_lr': 4e-05,
96
+ 'save_interval': 100,
97
+ 'tie_embeddings': None}}
98
+ Time to instantiate model: 0.02 seconds.
99
+ Total parameters: 1,100,048,384
100
+ [fix] out/pretrain/2407_full/final/lit_model.pth 是整包 state,提取 model 权重并原子覆盖...
101
+ [fix] 已覆盖为纯权重: out/pretrain/2407_full/final/lit_model.pth
102
+ Validating ...
103
+ Measured TFLOPs: 239.66
104
+ Epoch 1 | iter 32 step 1 | loss train: 1.458, val: 1.344 | iter time: 556.79 ms (step) remaining time: 0:34:12
105
+ Epoch 1 | iter 64 step 2 | loss train: 1.465, val: 1.344 | iter time: 357.59 ms (step) remaining time: 0:32:01
106
+ Epoch 1 | iter 96 step 3 | loss train: 1.486, val: 1.344 | iter time: 358.10 ms (step) remaining time: 0:31:10
107
+ Epoch 1 | iter 128 step 4 | loss train: 1.515, val: 1.344 | iter time: 356.50 ms (step) remaining time: 0:30:39
108
+ Epoch 1 | iter 160 step 5 | loss train: 1.452, val: 1.344 | iter time: 358.44 ms (step) remaining time: 0:30:17
109
+ Epoch 1 | iter 192 step 6 | loss train: 1.397, val: 1.344 | iter time: 360.78 ms (step) remaining time: 0:30:00
110
+ Epoch 1 | iter 224 step 7 | loss train: 1.378, val: 1.344 | iter time: 359.16 ms (step) remaining time: 0:29:44
111
+ Epoch 1 | iter 256 step 8 | loss train: 1.393, val: 1.344 | iter time: 359.36 ms (step) remaining time: 0:29:30
112
+ Epoch 1 | iter 288 step 9 | loss train: 1.460, val: 1.344 | iter time: 361.40 ms (step) remaining time: 0:29:16
113
+ Epoch 1 | iter 320 step 10 | loss train: 1.464, val: 1.344 | iter time: 359.52 ms (step) remaining time: 0:29:03
114
+ Epoch 1 | iter 352 step 11 | loss train: 1.491, val: 1.344 | iter time: 359.94 ms (step) remaining time: 0:28:50
115
+ Epoch 1 | iter 384 step 12 | loss train: 1.493, val: 1.344 | iter time: 358.72 ms (step) remaining time: 0:28:38
116
+ Epoch 1 | iter 416 step 13 | loss train: 1.471, val: 1.344 | iter time: 359.25 ms (step) remaining time: 0:28:26
117
+ Epoch 1 | iter 448 step 14 | loss train: 1.476, val: 1.344 | iter time: 360.71 ms (step) remaining time: 0:28:14
118
+ Epoch 1 | iter 480 step 15 | loss train: 1.483, val: 1.344 | iter time: 358.84 ms (step) remaining time: 0:28:03
119
+ Epoch 1 | iter 512 step 16 | loss train: 1.444, val: 1.344 | iter time: 360.49 ms (step) remaining time: 0:27:51
120
+ Epoch 1 | iter 544 step 17 | loss train: 1.509, val: 1.344 | iter time: 359.45 ms (step) remaining time: 0:27:40
121
+ Epoch 1 | iter 576 step 18 | loss train: 1.444, val: 1.344 | iter time: 360.95 ms (step) remaining time: 0:27:28
122
+ Epoch 1 | iter 608 step 19 | loss train: 1.502, val: 1.344 | iter time: 361.80 ms (step) remaining time: 0:27:17
123
+ Epoch 1 | iter 640 step 20 | loss train: 1.585, val: 1.344 | iter time: 360.43 ms (step) remaining time: 0:27:05
124
+ Epoch 1 | iter 672 step 21 | loss train: 1.587, val: 1.344 | iter time: 358.02 ms (step) remaining time: 0:26:54
125
+ Epoch 1 | iter 704 step 22 | loss train: 1.556, val: 1.344 | iter time: 361.13 ms (step) remaining time: 0:26:43
126
+ Epoch 1 | iter 736 step 23 | loss train: 1.535, val: 1.344 | iter time: 362.11 ms (step) remaining time: 0:26:32
127
+ Epoch 1 | iter 768 step 24 | loss train: 1.511, val: 1.344 | iter time: 360.88 ms (step) remaining time: 0:26:21
128
+ Epoch 1 | iter 800 step 25 | loss train: 1.524, val: 1.344 | iter time: 359.76 ms (step) remaining time: 0:26:10
129
+ Epoch 1 | iter 832 step 26 | loss train: 1.486, val: 1.344 | iter time: 361.83 ms (step) remaining time: 0:25:58
130
+ Epoch 1 | iter 864 step 27 | loss train: 1.507, val: 1.344 | iter time: 360.51 ms (step) remaining time: 0:25:47
131
+ Epoch 1 | iter 896 step 28 | loss train: 1.537, val: 1.344 | iter time: 360.39 ms (step) remaining time: 0:25:36
132
+ Epoch 1 | iter 928 step 29 | loss train: 1.457, val: 1.344 | iter time: 360.06 ms (step) remaining time: 0:25:25
133
+ Epoch 1 | iter 960 step 30 | loss train: 1.603, val: 1.344 | iter time: 359.78 ms (step) remaining time: 0:25:14
134
+ Epoch 1 | iter 992 step 31 | loss train: 1.554, val: 1.344 | iter time: 360.81 ms (step) remaining time: 0:25:03
135
+ Epoch 1 | iter 1024 step 32 | loss train: 1.450, val: 1.344 | iter time: 359.00 ms (step) remaining time: 0:24:52
136
+ Epoch 1 | iter 1056 step 33 | loss train: 1.489, val: 1.344 | iter time: 360.18 ms (step) remaining time: 0:24:41
137
+ Epoch 1 | iter 1088 step 34 | loss train: 1.556, val: 1.344 | iter time: 360.34 ms (step) remaining time: 0:24:30
138
+ Epoch 1 | iter 1120 step 35 | loss train: 1.487, val: 1.344 | iter time: 361.29 ms (step) remaining time: 0:24:19
139
+ Epoch 1 | iter 1152 step 36 | loss train: 1.492, val: 1.344 | iter time: 359.15 ms (step) remaining time: 0:24:08
140
+ Epoch 1 | iter 1184 step 37 | loss train: 1.408, val: 1.344 | iter time: 360.39 ms (step) remaining time: 0:23:57
141
+ Epoch 1 | iter 1216 step 38 | loss train: 1.448, val: 1.344 | iter time: 360.25 ms (step) remaining time: 0:23:46
142
+ Epoch 1 | iter 1248 step 39 | loss train: 1.391, val: 1.344 | iter time: 360.56 ms (step) remaining time: 0:23:35
143
+ Epoch 1 | iter 1280 step 40 | loss train: 1.492, val: 1.344 | iter time: 360.26 ms (step) remaining time: 0:23:24
144
+ Epoch 1 | iter 1312 step 41 | loss train: 1.465, val: 1.344 | iter time: 359.92 ms (step) remaining time: 0:23:13
145
+ Epoch 1 | iter 1344 step 42 | loss train: 1.478, val: 1.344 | iter time: 361.10 ms (step) remaining time: 0:23:03
146
+ Epoch 1 | iter 1376 step 43 | loss train: 1.494, val: 1.344 | iter time: 360.51 ms (step) remaining time: 0:22:52
147
+ Epoch 1 | iter 1408 step 44 | loss train: 1.434, val: 1.344 | iter time: 359.39 ms (step) remaining time: 0:22:41
148
+ Epoch 1 | iter 1440 step 45 | loss train: 1.460, val: 1.344 | iter time: 360.42 ms (step) remaining time: 0:22:30
149
+ Epoch 1 | iter 1472 step 46 | loss train: 1.486, val: 1.344 | iter time: 361.43 ms (step) remaining time: 0:22:19
150
+ Epoch 1 | iter 1504 step 47 | loss train: 1.542, val: 1.344 | iter time: 360.10 ms (step) remaining time: 0:22:08
151
+ Epoch 1 | iter 1536 step 48 | loss train: 1.522, val: 1.344 | iter time: 360.44 ms (step) remaining time: 0:21:58
152
+ Epoch 1 | iter 1568 step 49 | loss train: 1.424, val: 1.344 | iter time: 360.86 ms (step) remaining time: 0:21:47
153
+ Epoch 1 | iter 1600 step 50 | loss train: 1.459, val: 1.344 | iter time: 360.89 ms (step) remaining time: 0:21:36
154
+ Validating ...
155
+ iter 1600: val loss 1.3552, val time: 22362.39 ms
156
+ Epoch 1 | iter 1632 step 51 | loss train: 1.595, val: 1.355 | iter time: 359.85 ms (step) remaining time: 0:22:18
157
+ Epoch 1 | iter 1664 step 52 | loss train: 1.544, val: 1.355 | iter time: 361.40 ms (step) remaining time: 0:22:05
158
+ Epoch 1 | iter 1696 step 53 | loss train: 1.481, val: 1.355 | iter time: 358.16 ms (step) remaining time: 0:21:53
159
+ Epoch 1 | iter 1728 step 54 | loss train: 1.469, val: 1.355 | iter time: 361.34 ms (step) remaining time: 0:21:41
160
+ Epoch 1 | iter 1760 step 55 | loss train: 1.547, val: 1.355 | iter time: 358.79 ms (step) remaining time: 0:21:28
161
+ Epoch 1 | iter 1792 step 56 | loss train: 1.819, val: 1.355 | iter time: 360.56 ms (step) remaining time: 0:21:16
162
+ Epoch 1 | iter 1824 step 57 | loss train: 1.541, val: 1.355 | iter time: 358.84 ms (step) remaining time: 0:21:04
163
+ Epoch 1 | iter 1856 step 58 | loss train: 1.501, val: 1.355 | iter time: 359.05 ms (step) remaining time: 0:20:52
164
+ Epoch 1 | iter 1888 step 59 | loss train: 1.502, val: 1.355 | iter time: 359.98 ms (step) remaining time: 0:20:40
165
+ Epoch 1 | iter 1920 step 60 | loss train: 1.523, val: 1.355 | iter time: 360.47 ms (step) remaining time: 0:20:28
166
+ Epoch 1 | iter 1952 step 61 | loss train: 1.408, val: 1.355 | iter time: 360.91 ms (step) remaining time: 0:20:16
167
+ Epoch 1 | iter 1984 step 62 | loss train: 3.327, val: 1.355 | iter time: 360.42 ms (step) remaining time: 0:20:04
168
+ Epoch 1 | iter 2016 step 63 | loss train: 1.724, val: 1.355 | iter time: 358.85 ms (step) remaining time: 0:19:52
169
+ Epoch 1 | iter 2048 step 64 | loss train: 1.597, val: 1.355 | iter time: 357.56 ms (step) remaining time: 0:19:40
170
+ Epoch 1 | iter 2080 step 65 | loss train: 1.699, val: 1.355 | iter time: 360.26 ms (step) remaining time: 0:19:28
171
+ Epoch 1 | iter 2112 step 66 | loss train: 1.603, val: 1.355 | iter time: 362.22 ms (step) remaining time: 0:19:16
172
+ Epoch 1 | iter 2144 step 67 | loss train: 1.530, val: 1.355 | iter time: 359.96 ms (step) remaining time: 0:19:05
173
+ Epoch 1 | iter 2176 step 68 | loss train: 1.480, val: 1.355 | iter time: 360.02 ms (step) remaining time: 0:18:53
174
+ Epoch 1 | iter 2208 step 69 | loss train: 1.426, val: 1.355 | iter time: 358.89 ms (step) remaining time: 0:18:41
175
+ Epoch 1 | iter 2240 step 70 | loss train: 1.540, val: 1.355 | iter time: 360.23 ms (step) remaining time: 0:18:29
176
+ Epoch 1 | iter 2272 step 71 | loss train: 1.459, val: 1.355 | iter time: 359.89 ms (step) remaining time: 0:18:18
177
+ Epoch 1 | iter 2304 step 72 | loss train: 1.489, val: 1.355 | iter time: 360.13 ms (step) remaining time: 0:18:06
178
+ Epoch 1 | iter 2336 step 73 | loss train: 1.490, val: 1.355 | iter time: 359.68 ms (step) remaining time: 0:17:54
179
+ Epoch 1 | iter 2368 step 74 | loss train: 1.455, val: 1.355 | iter time: 359.32 ms (step) remaining time: 0:17:43
180
+ Epoch 1 | iter 2400 step 75 | loss train: 1.498, val: 1.355 | iter time: 360.18 ms (step) remaining time: 0:17:31
181
+ Epoch 1 | iter 2432 step 76 | loss train: 1.472, val: 1.355 | iter time: 359.40 ms (step) remaining time: 0:17:20
182
+ Epoch 1 | iter 2464 step 77 | loss train: 1.456, val: 1.355 | iter time: 360.49 ms (step) remaining time: 0:17:08
183
+ Epoch 1 | iter 2496 step 78 | loss train: 1.461, val: 1.355 | iter time: 359.38 ms (step) remaining time: 0:16:56
184
+ Epoch 1 | iter 2528 step 79 | loss train: 1.450, val: 1.355 | iter time: 359.69 ms (step) remaining time: 0:16:45
185
+ Epoch 1 | iter 2560 step 80 | loss train: 1.558, val: 1.355 | iter time: 359.64 ms (step) remaining time: 0:16:33
186
+ Epoch 1 | iter 2592 step 81 | loss train: 1.482, val: 1.355 | iter time: 360.25 ms (step) remaining time: 0:16:22
187
+ Epoch 1 | iter 2624 step 82 | loss train: 1.444, val: 1.355 | iter time: 358.39 ms (step) remaining time: 0:16:10
188
+ Epoch 1 | iter 2656 step 83 | loss train: 1.424, val: 1.355 | iter time: 358.90 ms (step) remaining time: 0:15:59
189
+ Epoch 1 | iter 2688 step 84 | loss train: 1.422, val: 1.355 | iter time: 361.70 ms (step) remaining time: 0:15:47
190
+ Epoch 1 | iter 2720 step 85 | loss train: 1.477, val: 1.355 | iter time: 360.55 ms (step) remaining time: 0:15:36
191
+ Epoch 1 | iter 2752 step 86 | loss train: 1.381, val: 1.355 | iter time: 358.76 ms (step) remaining time: 0:15:25
192
+ Epoch 1 | iter 2784 step 87 | loss train: 1.458, val: 1.355 | iter time: 358.94 ms (step) remaining time: 0:15:13
193
+ Epoch 1 | iter 2816 step 88 | loss train: 1.430, val: 1.355 | iter time: 361.58 ms (step) remaining time: 0:15:02
194
+ Epoch 1 | iter 2848 step 89 | loss train: 1.447, val: 1.355 | iter time: 359.51 ms (step) remaining time: 0:14:50
195
+ Epoch 1 | iter 2880 step 90 | loss train: 1.477, val: 1.355 | iter time: 359.26 ms (step) remaining time: 0:14:39
196
+ Epoch 1 | iter 2912 step 91 | loss train: 1.429, val: 1.355 | iter time: 359.35 ms (step) remaining time: 0:14:28
197
+ Epoch 1 | iter 2944 step 92 | loss train: 1.407, val: 1.355 | iter time: 359.75 ms (step) remaining time: 0:14:16
198
+ Epoch 1 | iter 2976 step 93 | loss train: 1.504, val: 1.355 | iter time: 359.08 ms (step) remaining time: 0:14:05
199
+ Epoch 1 | iter 3008 step 94 | loss train: 1.450, val: 1.355 | iter time: 360.41 ms (step) remaining time: 0:13:54
200
+ Epoch 1 | iter 3040 step 95 | loss train: 1.485, val: 1.355 | iter time: 360.40 ms (step) remaining time: 0:13:42
201
+ Epoch 1 | iter 3072 step 96 | loss train: 1.481, val: 1.355 | iter time: 360.54 ms (step) remaining time: 0:13:31
202
+ Epoch 1 | iter 3104 step 97 | loss train: 1.449, val: 1.355 | iter time: 360.36 ms (step) remaining time: 0:13:20
203
+ Epoch 1 | iter 3136 step 98 | loss train: 1.428, val: 1.355 | iter time: 359.49 ms (step) remaining time: 0:13:08
204
+ Epoch 1 | iter 3168 step 99 | loss train: 1.448, val: 1.355 | iter time: 359.20 ms (step) remaining time: 0:12:57
205
+ Epoch 1 | iter 3200 step 100 | loss train: 1.454, val: 1.355 | iter time: 360.19 ms (step) remaining time: 0:12:46
206
+ Validating ...
207
+ iter 3200: val loss 1.2657, val time: 22383.18 ms
208
+ Saving checkpoint to 'out/pretrain/2408_full/step-00000100/lit_model.pth'
209
+ Epoch 1 | iter 3232 step 101 | loss train: 1.461, val: 1.266 | iter time: 356.17 ms (step) remaining time: 0:13:01
210
+ Epoch 1 | iter 3264 step 102 | loss train: 1.392, val: 1.266 | iter time: 357.68 ms (step) remaining time: 0:12:49
211
+ Epoch 1 | iter 3296 step 103 | loss train: 1.374, val: 1.266 | iter time: 360.23 ms (step) remaining time: 0:12:37
212
+ Epoch 1 | iter 3328 step 104 | loss train: 1.422, val: 1.266 | iter time: 359.53 ms (step) remaining time: 0:12:26
213
+ Epoch 1 | iter 3360 step 105 | loss train: 1.445, val: 1.266 | iter time: 357.92 ms (step) remaining time: 0:12:14
214
+ Epoch 1 | iter 3392 step 106 | loss train: 1.389, val: 1.266 | iter time: 360.29 ms (step) remaining time: 0:12:02
215
+ Epoch 1 | iter 3424 step 107 | loss train: 1.376, val: 1.266 | iter time: 359.41 ms (step) remaining time: 0:11:50
216
+ Epoch 1 | iter 3456 step 108 | loss train: 1.411, val: 1.266 | iter time: 360.28 ms (step) remaining time: 0:11:38
217
+ Epoch 1 | iter 3488 step 109 | loss train: 1.439, val: 1.266 | iter time: 359.12 ms (step) remaining time: 0:11:27
218
+ Epoch 1 | iter 3520 step 110 | loss train: 1.434, val: 1.266 | iter time: 360.70 ms (step) remaining time: 0:11:15
219
+ Epoch 1 | iter 3552 step 111 | loss train: 1.368, val: 1.266 | iter time: 358.52 ms (step) remaining time: 0:11:03
220
+ Epoch 1 | iter 3584 step 112 | loss train: 1.394, val: 1.266 | iter time: 360.97 ms (step) remaining time: 0:10:51
221
+ Epoch 1 | iter 3616 step 113 | loss train: 1.403, val: 1.266 | iter time: 359.62 ms (step) remaining time: 0:10:40
222
+ Epoch 1 | iter 3648 step 114 | loss train: 1.411, val: 1.266 | iter time: 358.79 ms (step) remaining time: 0:10:28
223
+ Epoch 1 | iter 3680 step 115 | loss train: 1.444, val: 1.266 | iter time: 360.07 ms (step) remaining time: 0:10:16
224
+ Epoch 1 | iter 3712 step 116 | loss train: 1.344, val: 1.266 | iter time: 359.49 ms (step) remaining time: 0:10:05
225
+ Epoch 1 | iter 3744 step 117 | loss train: 1.419, val: 1.266 | iter time: 359.36 ms (step) remaining time: 0:09:53
226
+ Epoch 1 | iter 3776 step 118 | loss train: 1.439, val: 1.266 | iter time: 362.05 ms (step) remaining time: 0:09:41
227
+ Epoch 1 | iter 3808 step 119 | loss train: 1.456, val: 1.266 | iter time: 360.09 ms (step) remaining time: 0:09:30
228
+ Epoch 1 | iter 3840 step 120 | loss train: 1.444, val: 1.266 | iter time: 360.96 ms (step) remaining time: 0:09:18
229
+ Epoch 1 | iter 3872 step 121 | loss train: 1.454, val: 1.266 | iter time: 359.45 ms (step) remaining time: 0:09:07
230
+ Epoch 1 | iter 3904 step 122 | loss train: 1.404, val: 1.266 | iter time: 359.08 ms (step) remaining time: 0:08:55
231
+ Epoch 1 | iter 3936 step 123 | loss train: 1.365, val: 1.266 | iter time: 358.56 ms (step) remaining time: 0:08:43
232
+ Epoch 1 | iter 3968 step 124 | loss train: 1.397, val: 1.266 | iter time: 358.56 ms (step) remaining time: 0:08:32
233
+ Epoch 1 | iter 4000 step 125 | loss train: 1.384, val: 1.266 | iter time: 360.95 ms (step) remaining time: 0:08:20
234
+ Epoch 1 | iter 4032 step 126 | loss train: 1.367, val: 1.266 | iter time: 359.92 ms (step) remaining time: 0:08:09
235
+ Epoch 1 | iter 4064 step 127 | loss train: 1.447, val: 1.266 | iter time: 359.99 ms (step) remaining time: 0:07:57
236
+ Epoch 1 | iter 4096 step 128 | loss train: 1.398, val: 1.266 | iter time: 358.36 ms (step) remaining time: 0:07:46
237
+ Epoch 1 | iter 4128 step 129 | loss train: 1.346, val: 1.266 | iter time: 361.45 ms (step) remaining time: 0:07:34
238
+ Epoch 1 | iter 4160 step 130 | loss train: 1.394, val: 1.266 | iter time: 357.60 ms (step) remaining time: 0:07:23
239
+ Epoch 1 | iter 4192 step 131 | loss train: 1.401, val: 1.266 | iter time: 360.07 ms (step) remaining time: 0:07:11
240
+ Epoch 1 | iter 4224 step 132 | loss train: 1.339, val: 1.266 | iter time: 361.45 ms (step) remaining time: 0:07:00
241
+ Epoch 1 | iter 4256 step 133 | loss train: 1.424, val: 1.266 | iter time: 361.24 ms (step) remaining time: 0:06:48
242
+ Epoch 1 | iter 4288 step 134 | loss train: 1.374, val: 1.266 | iter time: 361.40 ms (step) remaining time: 0:06:37
243
+ Epoch 1 | iter 4320 step 135 | loss train: 1.398, val: 1.266 | iter time: 358.66 ms (step) remaining time: 0:06:25
244
+ Epoch 1 | iter 4352 step 136 | loss train: 1.423, val: 1.266 | iter time: 359.68 ms (step) remaining time: 0:06:14
245
+ Epoch 1 | iter 4384 step 137 | loss train: 1.409, val: 1.266 | iter time: 358.44 ms (step) remaining time: 0:06:02
246
+ Epoch 1 | iter 4416 step 138 | loss train: 1.358, val: 1.266 | iter time: 360.90 ms (step) remaining time: 0:05:51
247
+ Epoch 1 | iter 4448 step 139 | loss train: 1.375, val: 1.266 | iter time: 361.52 ms (step) remaining time: 0:05:39
248
+ Epoch 1 | iter 4480 step 140 | loss train: 1.423, val: 1.266 | iter time: 360.35 ms (step) remaining time: 0:05:28
249
+ Epoch 1 | iter 4512 step 141 | loss train: 1.447, val: 1.266 | iter time: 359.10 ms (step) remaining time: 0:05:16
250
+ Epoch 1 | iter 4544 step 142 | loss train: 1.362, val: 1.266 | iter time: 360.46 ms (step) remaining time: 0:05:05
251
+ Epoch 1 | iter 4576 step 143 | loss train: 1.441, val: 1.266 | iter time: 360.89 ms (step) remaining time: 0:04:54
252
+ Epoch 1 | iter 4608 step 144 | loss train: 1.394, val: 1.266 | iter time: 360.13 ms (step) remaining time: 0:04:42
253
+ Epoch 1 | iter 4640 step 145 | loss train: 1.422, val: 1.266 | iter time: 359.09 ms (step) remaining time: 0:04:31
254
+ Epoch 1 | iter 4672 step 146 | loss train: 1.371, val: 1.266 | iter time: 358.57 ms (step) remaining time: 0:04:20
255
+ Epoch 1 | iter 4704 step 147 | loss train: 1.471, val: 1.266 | iter time: 361.35 ms (step) remaining time: 0:04:08
256
+ Epoch 1 | iter 4736 step 148 | loss train: 1.338, val: 1.266 | iter time: 360.11 ms (step) remaining time: 0:03:57
257
+ Epoch 1 | iter 4768 step 149 | loss train: 1.434, val: 1.266 | iter time: 360.77 ms (step) remaining time: 0:03:45
258
+ Epoch 1 | iter 4800 step 150 | loss train: 1.403, val: 1.266 | iter time: 360.44 ms (step) remaining time: 0:03:34
259
+ Validating ...
260
+ iter 4800: val loss 1.2096, val time: 22381.61 ms
261
+ Epoch 1 | iter 4832 step 151 | loss train: 1.339, val: 1.210 | iter time: 359.89 ms (step) remaining time: 0:03:25
262
+ Epoch 1 | iter 4864 step 152 | loss train: 1.374, val: 1.210 | iter time: 360.31 ms (step) remaining time: 0:03:14
263
+ Epoch 1 | iter 4896 step 153 | loss train: 1.354, val: 1.210 | iter time: 357.09 ms (step) remaining time: 0:03:02
264
+ Epoch 1 | iter 4928 step 154 | loss train: 1.325, val: 1.210 | iter time: 360.00 ms (step) remaining time: 0:02:51
265
+ Epoch 1 | iter 4960 step 155 | loss train: 1.394, val: 1.210 | iter time: 360.33 ms (step) remaining time: 0:02:39
266
+ Epoch 1 | iter 4992 step 156 | loss train: 1.333, val: 1.210 | iter time: 360.26 ms (step) remaining time: 0:02:28
267
+ Epoch 1 | iter 5024 step 157 | loss train: 1.425, val: 1.210 | iter time: 359.99 ms (step) remaining time: 0:02:17
268
+ Epoch 1 | iter 5056 step 158 | loss train: 1.298, val: 1.210 | iter time: 360.05 ms (step) remaining time: 0:02:05
269
+ Epoch 1 | iter 5088 step 159 | loss train: 1.434, val: 1.210 | iter time: 361.24 ms (step) remaining time: 0:01:54
270
+ Epoch 1 | iter 5120 step 160 | loss train: 1.362, val: 1.210 | iter time: 361.03 ms (step) remaining time: 0:01:42
271
+ Epoch 1 | iter 5152 step 161 | loss train: 1.410, val: 1.210 | iter time: 358.45 ms (step) remaining time: 0:01:31
272
+ Epoch 1 | iter 5184 step 162 | loss train: 1.355, val: 1.210 | iter time: 359.71 ms (step) remaining time: 0:01:19
273
+ Epoch 1 | iter 5216 step 163 | loss train: 1.368, val: 1.210 | iter time: 358.79 ms (step) remaining time: 0:01:08
274
+ Epoch 1 | iter 5248 step 164 | loss train: 1.409, val: 1.210 | iter time: 360.47 ms (step) remaining time: 0:00:56
275
+ Epoch 1 | iter 5280 step 165 | loss train: 1.429, val: 1.210 | iter time: 360.04 ms (step) remaining time: 0:00:45
276
+ Epoch 1 | iter 5312 step 166 | loss train: 1.328, val: 1.210 | iter time: 358.53 ms (step) remaining time: 0:00:34
277
+ Epoch 1 | iter 5344 step 167 | loss train: 1.389, val: 1.210 | iter time: 359.97 ms (step) remaining time: 0:00:22
278
+ Epoch 1 | iter 5376 step 168 | loss train: 1.330, val: 1.210 | iter time: 360.91 ms (step) remaining time: 0:00:11
279
+ Epoch 2 | iter 5408 step 169 | loss train: 1.329, val: 1.210 | iter time: 363.53 ms (step) remaining time: 0:00:00
280
+ Validating ...
281
+ Final evaluation | val loss: 1.188 | val ppl: 3.280
282
+ Saving checkpoint to 'out/pretrain/2408_full/final/lit_model.pth'
283
+ ----------------------------------------
284
+ | Performance
285
+ | - Total tokens : 177,209,344
286
+ | - Training Time : 1988.04 s
287
+ | - Tok/sec : 95.92 tok/s
288
+ | ----------------------------------------
289
+ | Memory Usage
290
+ | - Memory Used : 26.32 GB
291
+ ----------------------------------------
out/pretrain/tinyllama/teelogs/2408_lr4e-5.txt ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Initializing distributed: GLOBAL_RANK: 0, MEMBER: 1/4
2
+ Initializing distributed: GLOBAL_RANK: 3, MEMBER: 4/4
3
+ [rank: 3] Seed set to 42
4
+ Initializing distributed: GLOBAL_RANK: 2, MEMBER: 3/4
5
+ [rank: 2] Seed set to 42
6
+ Initializing distributed: GLOBAL_RANK: 1, MEMBER: 2/4
7
+ ----------------------------------------------------------------------------------------------------
8
+ distributed_backend=nccl
9
+ All distributed processes registered. Starting with 4 processes
10
+ ----------------------------------------------------------------------------------------------------
11
+
12
+ [rank: 1] Seed set to 42
13
+ [rank: 0] Seed set to 42
14
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
15
+ train_dataloader = data.train_dataloader()
16
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
17
+ train_dataloader = data.train_dataloader()
18
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
19
+ train_dataloader = data.train_dataloader()
20
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
21
+ train_dataloader = data.train_dataloader()
22
+ All GPUs are fully connected via NVLink.
23
+ {'data': {'batch_size': 1,
24
+ 'data_path': PosixPath('litgpt/data/arxiv'),
25
+ 'num_workers': 0,
26
+ 'seed': 42,
27
+ 'seq_length': 2048,
28
+ 'use_starcoder': True},
29
+ 'data_dir': PosixPath('litgpt/data/arxiv/2408'),
30
+ 'devices': 'auto',
31
+ 'eval': {'evaluate_example': 'first',
32
+ 'final_validation': True,
33
+ 'initial_validation': True,
34
+ 'interval': 50,
35
+ 'max_iters': 200,
36
+ 'max_new_tokens': None},
37
+ 'initial_checkpoint_dir': PosixPath('out/pretrain/tinyllama/2407_full/final'),
38
+ 'log': {'group': None, 'project': None, 'run': None},
39
+ 'logger_name': 'tensorboard',
40
+ 'model_config': {'attention_logit_softcapping': None,
41
+ 'attention_scores_scalar': None,
42
+ 'attn_bias': False,
43
+ 'bias': False,
44
+ 'block_size': 2048,
45
+ 'final_logit_softcapping': None,
46
+ 'gelu_approximate': 'none',
47
+ 'head_size': 64,
48
+ 'hf_config': {'name': 'TinyLlama-1.1B-intermediate-step-1431k-3T',
49
+ 'org': 'TinyLlama'},
50
+ 'intermediate_size': 5632,
51
+ 'lm_head_bias': False,
52
+ 'mlp_class_name': 'LLaMAMLP',
53
+ 'moe_intermediate_size': None,
54
+ 'n_embd': 2048,
55
+ 'n_expert': 0,
56
+ 'n_expert_per_token': 0,
57
+ 'n_head': 32,
58
+ 'n_layer': 22,
59
+ 'n_query_groups': 4,
60
+ 'name': 'tiny-llama-1.1b',
61
+ 'norm_1': True,
62
+ 'norm_2': True,
63
+ 'norm_class_name': 'RMSNorm',
64
+ 'norm_eps': 1e-05,
65
+ 'norm_qk': False,
66
+ 'norm_qk_type': 'default',
67
+ 'padded_vocab_size': 32000,
68
+ 'padding_multiple': 64,
69
+ 'parallel_residual': False,
70
+ 'post_attention_norm': False,
71
+ 'post_mlp_norm': False,
72
+ 'rope_adjustments': None,
73
+ 'rope_base': 10000,
74
+ 'rope_condense_ratio': 1,
75
+ 'rope_indices': None,
76
+ 'rope_local_base_freq': None,
77
+ 'rotary_percentage': 1.0,
78
+ 'scale_embeddings': False,
79
+ 'shared_attention_norm': False,
80
+ 'sliding_window_indices': None,
81
+ 'sliding_window_size': None,
82
+ 'vocab_size': 32000},
83
+ 'model_name': 'tiny-llama-1.1b',
84
+ 'num_nodes': 1,
85
+ 'optimizer': "{'class_path': 'torch.optim.AdamW', 'init_args': {'lr': 4e-05, "
86
+ "'weight_decay': 0.1, 'betas': [0.9, 0.95]}}",
87
+ 'out_dir': PosixPath('out/pretrain/tinyllama/2408_lr4e-5'),
88
+ 'ppl': False,
89
+ 'precision': 'bf16-mixed',
90
+ 'resume': False,
91
+ 'seed': 42,
92
+ 'tokenizer_dir': PosixPath('checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T'),
93
+ 'train': {'epochs': None,
94
+ 'global_batch_size': 512,
95
+ 'log_interval': 1,
96
+ 'lr_warmup_fraction': None,
97
+ 'lr_warmup_steps': 20,
98
+ 'max_norm': 1.0,
99
+ 'max_seq_length': 2048,
100
+ 'max_steps': None,
101
+ 'max_tokens': 177209344,
102
+ 'micro_batch_size': 4,
103
+ 'min_lr': 4e-05,
104
+ 'save_interval': 100,
105
+ 'tie_embeddings': None}}
106
+ Time to instantiate model: 0.02 seconds.
107
+ Total parameters: 1,100,048,384
108
+ [ok] out/pretrain/tinyllama/2407_full/final/lit_model.pth 已是纯权重
109
+ Validating ...
110
+ Measured TFLOPs: 239.66
111
+ Epoch 1 | iter 32 step 1 | loss train: 1.451, val: 1.357 | iter time: 564.25 ms (step) remaining time: 0:32:45
112
+ Epoch 1 | iter 64 step 2 | loss train: 1.412, val: 1.357 | iter time: 358.32 ms (step) remaining time: 0:31:16
113
+ Epoch 1 | iter 96 step 3 | loss train: 1.501, val: 1.357 | iter time: 358.58 ms (step) remaining time: 0:30:40
114
+ Epoch 1 | iter 128 step 4 | loss train: 1.405, val: 1.357 | iter time: 357.45 ms (step) remaining time: 0:30:17
115
+ Epoch 1 | iter 160 step 5 | loss train: 1.419, val: 1.357 | iter time: 359.22 ms (step) remaining time: 0:30:00
116
+ Epoch 1 | iter 192 step 6 | loss train: 1.398, val: 1.357 | iter time: 359.88 ms (step) remaining time: 0:29:45
117
+ Epoch 1 | iter 224 step 7 | loss train: 1.454, val: 1.357 | iter time: 358.66 ms (step) remaining time: 0:29:32
118
+ Epoch 1 | iter 256 step 8 | loss train: 1.449, val: 1.357 | iter time: 360.74 ms (step) remaining time: 0:29:19
119
+ Epoch 1 | iter 288 step 9 | loss train: 1.409, val: 1.357 | iter time: 359.51 ms (step) remaining time: 0:29:09
120
+ Epoch 1 | iter 320 step 10 | loss train: 1.417, val: 1.357 | iter time: 360.20 ms (step) remaining time: 0:28:58
121
+ Epoch 1 | iter 352 step 11 | loss train: 1.440, val: 1.357 | iter time: 359.61 ms (step) remaining time: 0:28:46
122
+ Epoch 1 | iter 384 step 12 | loss train: 1.415, val: 1.357 | iter time: 359.53 ms (step) remaining time: 0:28:34
123
+ Epoch 1 | iter 416 step 13 | loss train: 1.479, val: 1.357 | iter time: 358.18 ms (step) remaining time: 0:28:23
124
+ Epoch 1 | iter 448 step 14 | loss train: 1.429, val: 1.357 | iter time: 359.84 ms (step) remaining time: 0:28:11
125
+ Epoch 1 | iter 480 step 15 | loss train: 1.366, val: 1.357 | iter time: 359.35 ms (step) remaining time: 0:28:00
126
+ Epoch 1 | iter 512 step 16 | loss train: 1.497, val: 1.357 | iter time: 361.43 ms (step) remaining time: 0:27:48
127
+ Epoch 1 | iter 544 step 17 | loss train: 1.392, val: 1.357 | iter time: 360.36 ms (step) remaining time: 0:27:37
128
+ Epoch 1 | iter 576 step 18 | loss train: 1.447, val: 1.357 | iter time: 359.41 ms (step) remaining time: 0:27:26
129
+ Epoch 1 | iter 608 step 19 | loss train: 1.374, val: 1.357 | iter time: 459.52 ms (step) remaining time: 0:27:15
130
+ Epoch 1 | iter 640 step 20 | loss train: 1.462, val: 1.357 | iter time: 358.72 ms (step) remaining time: 0:27:04
131
+ Epoch 1 | iter 672 step 21 | loss train: 1.410, val: 1.357 | iter time: 359.14 ms (step) remaining time: 0:26:53
132
+ Epoch 1 | iter 704 step 22 | loss train: 1.441, val: 1.357 | iter time: 359.78 ms (step) remaining time: 0:26:42
133
+ Epoch 1 | iter 736 step 23 | loss train: 1.408, val: 1.357 | iter time: 361.34 ms (step) remaining time: 0:26:31
134
+ Epoch 1 | iter 768 step 24 | loss train: 1.472, val: 1.357 | iter time: 358.33 ms (step) remaining time: 0:26:20
135
+ Epoch 1 | iter 800 step 25 | loss train: 1.457, val: 1.357 | iter time: 359.30 ms (step) remaining time: 0:26:09
136
+ Epoch 1 | iter 832 step 26 | loss train: 1.491, val: 1.357 | iter time: 358.45 ms (step) remaining time: 0:25:58
137
+ Epoch 1 | iter 864 step 27 | loss train: 1.431, val: 1.357 | iter time: 361.55 ms (step) remaining time: 0:25:47
138
+ Epoch 1 | iter 896 step 28 | loss train: 1.428, val: 1.357 | iter time: 361.07 ms (step) remaining time: 0:25:36
139
+ Epoch 1 | iter 928 step 29 | loss train: 1.490, val: 1.357 | iter time: 359.53 ms (step) remaining time: 0:25:25
140
+ Epoch 1 | iter 960 step 30 | loss train: 1.387, val: 1.357 | iter time: 360.94 ms (step) remaining time: 0:25:14
141
+ Epoch 1 | iter 992 step 31 | loss train: 1.451, val: 1.357 | iter time: 359.43 ms (step) remaining time: 0:25:03
142
+ Epoch 1 | iter 1024 step 32 | loss train: 1.402, val: 1.357 | iter time: 359.41 ms (step) remaining time: 0:24:52
143
+ Epoch 1 | iter 1056 step 33 | loss train: 1.385, val: 1.357 | iter time: 359.88 ms (step) remaining time: 0:24:41
144
+ Epoch 1 | iter 1088 step 34 | loss train: 1.421, val: 1.357 | iter time: 359.45 ms (step) remaining time: 0:24:30
145
+ Epoch 1 | iter 1120 step 35 | loss train: 1.453, val: 1.357 | iter time: 360.28 ms (step) remaining time: 0:24:19
146
+ Epoch 1 | iter 1152 step 36 | loss train: 1.469, val: 1.357 | iter time: 359.00 ms (step) remaining time: 0:24:08
147
+ Epoch 1 | iter 1184 step 37 | loss train: 1.333, val: 1.357 | iter time: 358.67 ms (step) remaining time: 0:23:57
148
+ Epoch 1 | iter 1216 step 38 | loss train: 1.385, val: 1.357 | iter time: 359.26 ms (step) remaining time: 0:23:46
149
+ Epoch 1 | iter 1248 step 39 | loss train: 1.457, val: 1.357 | iter time: 360.87 ms (step) remaining time: 0:23:35
150
+ Epoch 1 | iter 1280 step 40 | loss train: 1.468, val: 1.357 | iter time: 359.63 ms (step) remaining time: 0:23:24
151
+ Epoch 1 | iter 1312 step 41 | loss train: 1.363, val: 1.357 | iter time: 360.46 ms (step) remaining time: 0:23:13
152
+ Epoch 1 | iter 1344 step 42 | loss train: 1.413, val: 1.357 | iter time: 360.37 ms (step) remaining time: 0:23:02
153
+ Epoch 1 | iter 1376 step 43 | loss train: 1.411, val: 1.357 | iter time: 360.40 ms (step) remaining time: 0:22:51
154
+ Epoch 1 | iter 1408 step 44 | loss train: 1.393, val: 1.357 | iter time: 360.33 ms (step) remaining time: 0:22:40
155
+ Epoch 1 | iter 1440 step 45 | loss train: 1.374, val: 1.357 | iter time: 359.64 ms (step) remaining time: 0:22:29
156
+ Epoch 1 | iter 1472 step 46 | loss train: 1.402, val: 1.357 | iter time: 360.88 ms (step) remaining time: 0:22:19
157
+ Epoch 1 | iter 1504 step 47 | loss train: 1.378, val: 1.357 | iter time: 358.82 ms (step) remaining time: 0:22:08
158
+ Epoch 1 | iter 1536 step 48 | loss train: 1.377, val: 1.357 | iter time: 358.09 ms (step) remaining time: 0:21:57
159
+ Epoch 1 | iter 1568 step 49 | loss train: 1.385, val: 1.357 | iter time: 358.65 ms (step) remaining time: 0:21:46
160
+ Epoch 1 | iter 1600 step 50 | loss train: 1.343, val: 1.357 | iter time: 359.79 ms (step) remaining time: 0:21:35
161
+ Validating ...
162
+ iter 1600: val loss 1.3176, val time: 21942.50 ms
163
+ Epoch 1 | iter 1632 step 51 | loss train: 1.459, val: 1.318 | iter time: 361.78 ms (step) remaining time: 0:22:15
164
+ Epoch 1 | iter 1664 step 52 | loss train: 1.454, val: 1.318 | iter time: 359.96 ms (step) remaining time: 0:22:03
165
+ Epoch 1 | iter 1696 step 53 | loss train: 1.396, val: 1.318 | iter time: 361.47 ms (step) remaining time: 0:21:50
166
+ Epoch 1 | iter 1728 step 54 | loss train: 1.476, val: 1.318 | iter time: 358.80 ms (step) remaining time: 0:21:38
167
+ Epoch 1 | iter 1760 step 55 | loss train: 1.390, val: 1.318 | iter time: 677.12 ms (step) remaining time: 0:21:26
168
+ Epoch 1 | iter 1792 step 56 | loss train: 1.415, val: 1.318 | iter time: 360.52 ms (step) remaining time: 0:21:14
169
+ Epoch 1 | iter 1824 step 57 | loss train: 1.442, val: 1.318 | iter time: 361.78 ms (step) remaining time: 0:21:02
170
+ Epoch 1 | iter 1856 step 58 | loss train: 1.456, val: 1.318 | iter time: 361.70 ms (step) remaining time: 0:20:50
171
+ Epoch 1 | iter 1888 step 59 | loss train: 1.411, val: 1.318 | iter time: 362.05 ms (step) remaining time: 0:20:38
172
+ Epoch 1 | iter 1920 step 60 | loss train: 1.436, val: 1.318 | iter time: 359.75 ms (step) remaining time: 0:20:26
173
+ Epoch 1 | iter 1952 step 61 | loss train: 1.412, val: 1.318 | iter time: 359.40 ms (step) remaining time: 0:20:14
174
+ Epoch 1 | iter 1984 step 62 | loss train: 1.395, val: 1.318 | iter time: 359.34 ms (step) remaining time: 0:20:02
175
+ Epoch 1 | iter 2016 step 63 | loss train: 1.419, val: 1.318 | iter time: 360.03 ms (step) remaining time: 0:19:50
176
+ Epoch 1 | iter 2048 step 64 | loss train: 1.410, val: 1.318 | iter time: 360.68 ms (step) remaining time: 0:19:39
177
+ Epoch 1 | iter 2080 step 65 | loss train: 1.336, val: 1.318 | iter time: 358.81 ms (step) remaining time: 0:19:27
178
+ Epoch 1 | iter 2112 step 66 | loss train: 1.338, val: 1.318 | iter time: 359.95 ms (step) remaining time: 0:19:15
179
+ Epoch 1 | iter 2144 step 67 | loss train: 1.389, val: 1.318 | iter time: 359.72 ms (step) remaining time: 0:19:03
180
+ Epoch 1 | iter 2176 step 68 | loss train: 1.424, val: 1.318 | iter time: 359.37 ms (step) remaining time: 0:18:52
181
+ Epoch 1 | iter 2208 step 69 | loss train: 1.476, val: 1.318 | iter time: 359.64 ms (step) remaining time: 0:18:40
182
+ Epoch 1 | iter 2240 step 70 | loss train: 1.349, val: 1.318 | iter time: 361.64 ms (step) remaining time: 0:18:28
183
+ Epoch 1 | iter 2272 step 71 | loss train: 1.398, val: 1.318 | iter time: 359.20 ms (step) remaining time: 0:18:17
184
+ Epoch 1 | iter 2304 step 72 | loss train: 1.403, val: 1.318 | iter time: 358.98 ms (step) remaining time: 0:18:05
185
+ Epoch 1 | iter 2336 step 73 | loss train: 1.407, val: 1.318 | iter time: 359.70 ms (step) remaining time: 0:17:53
186
+ Epoch 1 | iter 2368 step 74 | loss train: 1.438, val: 1.318 | iter time: 358.49 ms (step) remaining time: 0:17:42
187
+ Epoch 1 | iter 2400 step 75 | loss train: 1.436, val: 1.318 | iter time: 361.74 ms (step) remaining time: 0:17:30
188
+ Epoch 1 | iter 2432 step 76 | loss train: 1.371, val: 1.318 | iter time: 359.09 ms (step) remaining time: 0:17:19
189
+ Epoch 1 | iter 2464 step 77 | loss train: 1.419, val: 1.318 | iter time: 359.18 ms (step) remaining time: 0:17:07
190
+ Epoch 1 | iter 2496 step 78 | loss train: 1.439, val: 1.318 | iter time: 360.09 ms (step) remaining time: 0:16:55
191
+ Epoch 1 | iter 2528 step 79 | loss train: 1.384, val: 1.318 | iter time: 358.98 ms (step) remaining time: 0:16:44
192
+ Epoch 1 | iter 2560 step 80 | loss train: 1.366, val: 1.318 | iter time: 359.61 ms (step) remaining time: 0:16:33
193
+ Epoch 1 | iter 2592 step 81 | loss train: 1.482, val: 1.318 | iter time: 360.42 ms (step) remaining time: 0:16:21
194
+ Epoch 1 | iter 2624 step 82 | loss train: 1.332, val: 1.318 | iter time: 358.74 ms (step) remaining time: 0:16:10
195
+ Epoch 1 | iter 2656 step 83 | loss train: 1.492, val: 1.318 | iter time: 359.88 ms (step) remaining time: 0:15:58
196
+ Epoch 1 | iter 2688 step 84 | loss train: 1.445, val: 1.318 | iter time: 358.31 ms (step) remaining time: 0:15:47
197
+ Epoch 1 | iter 2720 step 85 | loss train: 1.398, val: 1.318 | iter time: 360.69 ms (step) remaining time: 0:15:35
198
+ Epoch 1 | iter 2752 step 86 | loss train: 1.450, val: 1.318 | iter time: 359.74 ms (step) remaining time: 0:15:24
199
+ Epoch 1 | iter 2784 step 87 | loss train: 1.416, val: 1.318 | iter time: 359.66 ms (step) remaining time: 0:15:13
200
+ Epoch 1 | iter 2816 step 88 | loss train: 1.476, val: 1.318 | iter time: 360.79 ms (step) remaining time: 0:15:01
201
+ Epoch 1 | iter 2848 step 89 | loss train: 1.371, val: 1.318 | iter time: 357.84 ms (step) remaining time: 0:14:50
202
+ Epoch 1 | iter 2880 step 90 | loss train: 1.470, val: 1.318 | iter time: 358.94 ms (step) remaining time: 0:14:39
203
+ Epoch 1 | iter 2912 step 91 | loss train: 1.328, val: 1.318 | iter time: 358.27 ms (step) remaining time: 0:14:27
204
+ Epoch 1 | iter 2944 step 92 | loss train: 1.442, val: 1.318 | iter time: 359.77 ms (step) remaining time: 0:14:16
205
+ Epoch 1 | iter 2976 step 93 | loss train: 1.475, val: 1.318 | iter time: 359.28 ms (step) remaining time: 0:14:05
206
+ Epoch 1 | iter 3008 step 94 | loss train: 1.383, val: 1.318 | iter time: 363.83 ms (step) remaining time: 0:13:53
207
+ Epoch 1 | iter 3040 step 95 | loss train: 1.422, val: 1.318 | iter time: 360.13 ms (step) remaining time: 0:13:42
208
+ Epoch 1 | iter 3072 step 96 | loss train: 1.459, val: 1.318 | iter time: 360.31 ms (step) remaining time: 0:13:31
209
+ Epoch 1 | iter 3104 step 97 | loss train: 1.492, val: 1.318 | iter time: 361.09 ms (step) remaining time: 0:13:19
210
+ Epoch 1 | iter 3136 step 98 | loss train: 1.476, val: 1.318 | iter time: 359.04 ms (step) remaining time: 0:13:08
211
+ Epoch 1 | iter 3168 step 99 | loss train: 1.344, val: 1.318 | iter time: 358.83 ms (step) remaining time: 0:12:57
212
+ Epoch 1 | iter 3200 step 100 | loss train: 1.407, val: 1.318 | iter time: 360.71 ms (step) remaining time: 0:12:46
213
+ Validating ...
214
+ iter 3200: val loss 1.2629, val time: 21940.88 ms
215
+ Saving checkpoint to 'out/pretrain/tinyllama/2408_lr4e-5/step-00000100/lit_model.pth'
216
+ Epoch 1 | iter 3232 step 101 | loss train: 1.465, val: 1.263 | iter time: 355.97 ms (step) remaining time: 0:13:00
217
+ Epoch 1 | iter 3264 step 102 | loss train: 1.429, val: 1.263 | iter time: 359.09 ms (step) remaining time: 0:12:48
218
+ Epoch 1 | iter 3296 step 103 | loss train: 1.436, val: 1.263 | iter time: 359.06 ms (step) remaining time: 0:12:36
219
+ Epoch 1 | iter 3328 step 104 | loss train: 1.424, val: 1.263 | iter time: 359.38 ms (step) remaining time: 0:12:24
220
+ Epoch 1 | iter 3360 step 105 | loss train: 1.425, val: 1.263 | iter time: 359.97 ms (step) remaining time: 0:12:13
221
+ Epoch 1 | iter 3392 step 106 | loss train: 1.426, val: 1.263 | iter time: 359.51 ms (step) remaining time: 0:12:01
222
+ Epoch 1 | iter 3424 step 107 | loss train: 1.420, val: 1.263 | iter time: 358.93 ms (step) remaining time: 0:11:49
223
+ Epoch 1 | iter 3456 step 108 | loss train: 1.394, val: 1.263 | iter time: 359.30 ms (step) remaining time: 0:11:37
224
+ Epoch 1 | iter 3488 step 109 | loss train: 1.424, val: 1.263 | iter time: 360.28 ms (step) remaining time: 0:11:25
225
+ Epoch 1 | iter 3520 step 110 | loss train: 1.407, val: 1.263 | iter time: 360.67 ms (step) remaining time: 0:11:14
226
+ Epoch 1 | iter 3552 step 111 | loss train: 1.416, val: 1.263 | iter time: 360.38 ms (step) remaining time: 0:11:02
227
+ Epoch 1 | iter 3584 step 112 | loss train: 1.397, val: 1.263 | iter time: 361.11 ms (step) remaining time: 0:10:50
228
+ Epoch 1 | iter 3616 step 113 | loss train: 1.408, val: 1.263 | iter time: 601.64 ms (step) remaining time: 0:10:39
229
+ Epoch 1 | iter 3648 step 114 | loss train: 1.481, val: 1.263 | iter time: 360.76 ms (step) remaining time: 0:10:27
230
+ Epoch 1 | iter 3680 step 115 | loss train: 1.409, val: 1.263 | iter time: 360.08 ms (step) remaining time: 0:10:15
231
+ Epoch 1 | iter 3712 step 116 | loss train: 1.380, val: 1.263 | iter time: 360.00 ms (step) remaining time: 0:10:04
232
+ Epoch 1 | iter 3744 step 117 | loss train: 1.389, val: 1.263 | iter time: 360.30 ms (step) remaining time: 0:09:52
233
+ Epoch 1 | iter 3776 step 118 | loss train: 1.352, val: 1.263 | iter time: 360.52 ms (step) remaining time: 0:09:40
234
+ Epoch 1 | iter 3808 step 119 | loss train: 1.371, val: 1.263 | iter time: 358.91 ms (step) remaining time: 0:09:29
235
+ Epoch 1 | iter 3840 step 120 | loss train: 1.421, val: 1.263 | iter time: 360.84 ms (step) remaining time: 0:09:17
236
+ Epoch 1 | iter 3872 step 121 | loss train: 1.337, val: 1.263 | iter time: 361.95 ms (step) remaining time: 0:09:06
237
+ Epoch 1 | iter 3904 step 122 | loss train: 1.489, val: 1.263 | iter time: 358.32 ms (step) remaining time: 0:08:54
238
+ Epoch 1 | iter 3936 step 123 | loss train: 1.377, val: 1.263 | iter time: 359.93 ms (step) remaining time: 0:08:42
239
+ Epoch 1 | iter 3968 step 124 | loss train: 1.439, val: 1.263 | iter time: 358.88 ms (step) remaining time: 0:08:31
240
+ Epoch 1 | iter 4000 step 125 | loss train: 1.362, val: 1.263 | iter time: 359.74 ms (step) remaining time: 0:08:20
241
+ Epoch 1 | iter 4032 step 126 | loss train: 1.406, val: 1.263 | iter time: 359.15 ms (step) remaining time: 0:08:08
242
+ Epoch 1 | iter 4064 step 127 | loss train: 1.366, val: 1.263 | iter time: 358.82 ms (step) remaining time: 0:07:57
243
+ Epoch 1 | iter 4096 step 128 | loss train: 1.380, val: 1.263 | iter time: 359.08 ms (step) remaining time: 0:07:45
244
+ Epoch 1 | iter 4128 step 129 | loss train: 1.384, val: 1.263 | iter time: 359.91 ms (step) remaining time: 0:07:34
245
+ Epoch 1 | iter 4160 step 130 | loss train: 1.419, val: 1.263 | iter time: 360.38 ms (step) remaining time: 0:07:22
246
+ Epoch 1 | iter 4192 step 131 | loss train: 1.391, val: 1.263 | iter time: 360.00 ms (step) remaining time: 0:07:11
247
+ Epoch 1 | iter 4224 step 132 | loss train: 1.495, val: 1.263 | iter time: 360.55 ms (step) remaining time: 0:06:59
248
+ Epoch 1 | iter 4256 step 133 | loss train: 1.401, val: 1.263 | iter time: 359.65 ms (step) remaining time: 0:06:48
249
+ Epoch 1 | iter 4288 step 134 | loss train: 1.386, val: 1.263 | iter time: 359.92 ms (step) remaining time: 0:06:36
250
+ Epoch 1 | iter 4320 step 135 | loss train: 1.427, val: 1.263 | iter time: 360.18 ms (step) remaining time: 0:06:25
251
+ Epoch 1 | iter 4352 step 136 | loss train: 1.408, val: 1.263 | iter time: 360.58 ms (step) remaining time: 0:06:13
252
+ Epoch 1 | iter 4384 step 137 | loss train: 1.412, val: 1.263 | iter time: 359.62 ms (step) remaining time: 0:06:02
253
+ Epoch 1 | iter 4416 step 138 | loss train: 1.416, val: 1.263 | iter time: 362.31 ms (step) remaining time: 0:05:50
254
+ Epoch 1 | iter 4448 step 139 | loss train: 1.433, val: 1.263 | iter time: 359.63 ms (step) remaining time: 0:05:39
255
+ Epoch 1 | iter 4480 step 140 | loss train: 1.450, val: 1.263 | iter time: 359.16 ms (step) remaining time: 0:05:28
256
+ Epoch 1 | iter 4512 step 141 | loss train: 1.384, val: 1.263 | iter time: 360.05 ms (step) remaining time: 0:05:16
257
+ Epoch 1 | iter 4544 step 142 | loss train: 1.348, val: 1.263 | iter time: 359.53 ms (step) remaining time: 0:05:05
258
+ Epoch 1 | iter 4576 step 143 | loss train: 1.396, val: 1.263 | iter time: 359.50 ms (step) remaining time: 0:04:53
259
+ Epoch 1 | iter 4608 step 144 | loss train: 1.374, val: 1.263 | iter time: 360.42 ms (step) remaining time: 0:04:42
260
+ Epoch 1 | iter 4640 step 145 | loss train: 1.386, val: 1.263 | iter time: 358.19 ms (step) remaining time: 0:04:31
261
+ Epoch 1 | iter 4672 step 146 | loss train: 1.401, val: 1.263 | iter time: 358.30 ms (step) remaining time: 0:04:19
262
+ Epoch 1 | iter 4704 step 147 | loss train: 1.368, val: 1.263 | iter time: 358.33 ms (step) remaining time: 0:04:08
263
+ Epoch 1 | iter 4736 step 148 | loss train: 1.373, val: 1.263 | iter time: 359.05 ms (step) remaining time: 0:03:57
264
+ Epoch 1 | iter 4768 step 149 | loss train: 1.351, val: 1.263 | iter time: 357.45 ms (step) remaining time: 0:03:45
265
+ Epoch 1 | iter 4800 step 150 | loss train: 1.422, val: 1.263 | iter time: 361.23 ms (step) remaining time: 0:03:34
266
+ Validating ...
267
+ iter 4800: val loss 1.2647, val time: 21943.29 ms
268
+ Epoch 1 | iter 4832 step 151 | loss train: 1.419, val: 1.265 | iter time: 360.61 ms (step) remaining time: 0:03:25
269
+ Epoch 1 | iter 4864 step 152 | loss train: 1.409, val: 1.265 | iter time: 359.59 ms (step) remaining time: 0:03:14
270
+ Epoch 1 | iter 4896 step 153 | loss train: 1.341, val: 1.265 | iter time: 361.08 ms (step) remaining time: 0:03:02
271
+ Epoch 1 | iter 4928 step 154 | loss train: 1.423, val: 1.265 | iter time: 359.58 ms (step) remaining time: 0:02:51
272
+ Epoch 1 | iter 4960 step 155 | loss train: 1.353, val: 1.265 | iter time: 361.92 ms (step) remaining time: 0:02:39
273
+ Epoch 1 | iter 4992 step 156 | loss train: 1.403, val: 1.265 | iter time: 361.02 ms (step) remaining time: 0:02:28
274
+ Epoch 1 | iter 5024 step 157 | loss train: 1.419, val: 1.265 | iter time: 360.62 ms (step) remaining time: 0:02:16
275
+ Epoch 1 | iter 5056 step 158 | loss train: 1.471, val: 1.265 | iter time: 358.83 ms (step) remaining time: 0:02:05
276
+ Epoch 1 | iter 5088 step 159 | loss train: 1.309, val: 1.265 | iter time: 359.32 ms (step) remaining time: 0:01:53
277
+ Epoch 1 | iter 5120 step 160 | loss train: 1.454, val: 1.265 | iter time: 359.49 ms (step) remaining time: 0:01:42
278
+ Epoch 1 | iter 5152 step 161 | loss train: 1.370, val: 1.265 | iter time: 360.16 ms (step) remaining time: 0:01:31
279
+ Epoch 1 | iter 5184 step 162 | loss train: 1.366, val: 1.265 | iter time: 359.12 ms (step) remaining time: 0:01:19
280
+ Epoch 1 | iter 5216 step 163 | loss train: 1.455, val: 1.265 | iter time: 361.51 ms (step) remaining time: 0:01:08
281
+ Epoch 1 | iter 5248 step 164 | loss train: 1.374, val: 1.265 | iter time: 360.18 ms (step) remaining time: 0:00:56
282
+ Epoch 1 | iter 5280 step 165 | loss train: 1.392, val: 1.265 | iter time: 360.68 ms (step) remaining time: 0:00:45
283
+ Epoch 1 | iter 5312 step 166 | loss train: 1.397, val: 1.265 | iter time: 360.54 ms (step) remaining time: 0:00:34
284
+ Epoch 1 | iter 5344 step 167 | loss train: 1.394, val: 1.265 | iter time: 361.64 ms (step) remaining time: 0:00:22
285
+ Epoch 1 | iter 5376 step 168 | loss train: 1.372, val: 1.265 | iter time: 359.24 ms (step) remaining time: 0:00:11
286
+ Epoch 2 | iter 5408 step 169 | loss train: 1.337, val: 1.265 | iter time: 358.30 ms (step) remaining time: 0:00:00
287
+ Validating ...
288
+ Final evaluation | val loss: 1.265 | val ppl: 3.545
289
+ Saving checkpoint to 'out/pretrain/tinyllama/2408_lr4e-5/final/lit_model.pth'
290
+ ----------------------------------------
291
+ | Performance
292
+ | - Total tokens : 177,209,344
293
+ | - Training Time : 1983.42 s
294
+ | - Tok/sec : 91.64 tok/s
295
+ | ----------------------------------------
296
+ | Memory Usage
297
+ | - Memory Used : 26.32 GB
298
+ ----------------------------------------
out/pretrain/tinyllama/teelogs/2409.txt ADDED
File without changes
out/pretrain/tinyllama/teelogs/2409_full.txt ADDED
@@ -0,0 +1,354 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Initializing distributed: GLOBAL_RANK: 0, MEMBER: 1/4
2
+ Initializing distributed: GLOBAL_RANK: 1, MEMBER: 2/4
3
+ [rank: 1] Seed set to 42
4
+ Initializing distributed: GLOBAL_RANK: 3, MEMBER: 4/4
5
+ Initializing distributed: GLOBAL_RANK: 2, MEMBER: 3/4
6
+ [rank: 2] Seed set to 42
7
+ [rank: 3] Seed set to 42
8
+ ----------------------------------------------------------------------------------------------------
9
+ distributed_backend=nccl
10
+ All distributed processes registered. Starting with 4 processes
11
+ ----------------------------------------------------------------------------------------------------
12
+
13
+ [rank: 0] Seed set to 42
14
+ All GPUs are fully connected via NVLink.
15
+ {'data': {'batch_size': 1,
16
+ 'data_path': PosixPath('litgpt/data/arxiv'),
17
+ 'num_workers': 8,
18
+ 'seed': 42,
19
+ 'seq_length': 2048,
20
+ 'use_starcoder': True},
21
+ 'data_dir': PosixPath('litgpt/data/arxiv/2409'),
22
+ 'devices': 'auto',
23
+ 'eval': {'evaluate_example': 'first',
24
+ 'final_validation': True,
25
+ 'initial_validation': True,
26
+ 'interval': 50,
27
+ 'max_iters': 200,
28
+ 'max_new_tokens': None},
29
+ 'initial_checkpoint_dir': PosixPath('out/pretrain/2408_full/final'),
30
+ 'log': {'group': None, 'project': None, 'run': None},
31
+ 'logger_name': 'tensorboard',
32
+ 'model_config': {'attention_logit_softcapping': None,
33
+ 'attention_scores_scalar': None,
34
+ 'attn_bias': False,
35
+ 'bias': False,
36
+ 'block_size': 2048,
37
+ 'final_logit_softcapping': None,
38
+ 'gelu_approximate': 'none',
39
+ 'head_size': 64,
40
+ 'hf_config': {'name': 'TinyLlama-1.1B-intermediate-step-1431k-3T',
41
+ 'org': 'TinyLlama'},
42
+ 'intermediate_size': 5632,
43
+ 'lm_head_bias': False,
44
+ 'mlp_class_name': 'LLaMAMLP',
45
+ 'moe_intermediate_size': None,
46
+ 'n_embd': 2048,
47
+ 'n_expert': 0,
48
+ 'n_expert_per_token': 0,
49
+ 'n_head': 32,
50
+ 'n_layer': 22,
51
+ 'n_query_groups': 4,
52
+ 'name': 'tiny-llama-1.1b',
53
+ 'norm_1': True,
54
+ 'norm_2': True,
55
+ 'norm_class_name': 'RMSNorm',
56
+ 'norm_eps': 1e-05,
57
+ 'norm_qk': False,
58
+ 'norm_qk_type': 'default',
59
+ 'padded_vocab_size': 32000,
60
+ 'padding_multiple': 64,
61
+ 'parallel_residual': False,
62
+ 'post_attention_norm': False,
63
+ 'post_mlp_norm': False,
64
+ 'rope_adjustments': None,
65
+ 'rope_base': 10000,
66
+ 'rope_condense_ratio': 1,
67
+ 'rope_indices': None,
68
+ 'rope_local_base_freq': None,
69
+ 'rotary_percentage': 1.0,
70
+ 'scale_embeddings': False,
71
+ 'shared_attention_norm': False,
72
+ 'sliding_window_indices': None,
73
+ 'sliding_window_size': None,
74
+ 'vocab_size': 32000},
75
+ 'model_name': 'tiny-llama-1.1b',
76
+ 'num_nodes': 1,
77
+ 'optimizer': "{'class_path': 'torch.optim.AdamW', 'init_args': {'lr': 0.0004, "
78
+ "'weight_decay': 0.1, 'betas': [0.9, 0.95]}}",
79
+ 'out_dir': PosixPath('out/pretrain/2409_full'),
80
+ 'ppl': False,
81
+ 'precision': 'bf16-mixed',
82
+ 'resume': False,
83
+ 'seed': 42,
84
+ 'tokenizer_dir': PosixPath('checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T'),
85
+ 'train': {'epochs': None,
86
+ 'global_batch_size': 512,
87
+ 'log_interval': 1,
88
+ 'lr_warmup_fraction': None,
89
+ 'lr_warmup_steps': 20,
90
+ 'max_norm': 1.0,
91
+ 'max_seq_length': 2048,
92
+ 'max_steps': None,
93
+ 'max_tokens': 240123904,
94
+ 'micro_batch_size': 4,
95
+ 'min_lr': 4e-05,
96
+ 'save_interval': 100,
97
+ 'tie_embeddings': None}}
98
+ Time to instantiate model: 0.04 seconds.
99
+ Total parameters: 1,100,048,384
100
+ [fix] out/pretrain/2408_full/final/lit_model.pth 是整包 state,提取 model 权重并原子覆盖...
101
+ [fix] 已覆盖为纯权重: out/pretrain/2408_full/final/lit_model.pth
102
+ Validating ...
103
+ Measured TFLOPs: 239.66
104
+ Epoch 1 | iter 32 step 1 | loss train: 1.498, val: 1.439 | iter time: 558.21 ms (step) remaining time: 0:45:30
105
+ Epoch 1 | iter 64 step 2 | loss train: 1.432, val: 1.439 | iter time: 355.65 ms (step) remaining time: 0:43:02
106
+ Epoch 1 | iter 96 step 3 | loss train: 1.408, val: 1.439 | iter time: 357.49 ms (step) remaining time: 0:42:06
107
+ Epoch 1 | iter 128 step 4 | loss train: 1.420, val: 1.439 | iter time: 359.73 ms (step) remaining time: 0:41:33
108
+ Epoch 1 | iter 160 step 5 | loss train: 1.421, val: 1.439 | iter time: 359.07 ms (step) remaining time: 0:41:10
109
+ Epoch 1 | iter 192 step 6 | loss train: 1.451, val: 1.439 | iter time: 360.19 ms (step) remaining time: 0:40:52
110
+ Epoch 1 | iter 224 step 7 | loss train: 1.450, val: 1.439 | iter time: 358.86 ms (step) remaining time: 0:40:35
111
+ Epoch 1 | iter 256 step 8 | loss train: 1.471, val: 1.439 | iter time: 358.43 ms (step) remaining time: 0:40:21
112
+ Epoch 1 | iter 288 step 9 | loss train: 1.468, val: 1.439 | iter time: 360.64 ms (step) remaining time: 0:40:07
113
+ Epoch 1 | iter 320 step 10 | loss train: 1.372, val: 1.439 | iter time: 357.91 ms (step) remaining time: 0:39:54
114
+ Epoch 1 | iter 352 step 11 | loss train: 1.486, val: 1.439 | iter time: 360.52 ms (step) remaining time: 0:39:41
115
+ Epoch 1 | iter 384 step 12 | loss train: 1.443, val: 1.439 | iter time: 358.10 ms (step) remaining time: 0:39:29
116
+ Epoch 1 | iter 416 step 13 | loss train: 1.522, val: 1.439 | iter time: 358.28 ms (step) remaining time: 0:39:17
117
+ Epoch 1 | iter 448 step 14 | loss train: 1.455, val: 1.439 | iter time: 360.51 ms (step) remaining time: 0:39:05
118
+ Epoch 1 | iter 480 step 15 | loss train: 1.406, val: 1.439 | iter time: 360.65 ms (step) remaining time: 0:38:53
119
+ Epoch 1 | iter 512 step 16 | loss train: 1.525, val: 1.439 | iter time: 358.56 ms (step) remaining time: 0:38:42
120
+ Epoch 1 | iter 544 step 17 | loss train: 1.481, val: 1.439 | iter time: 360.99 ms (step) remaining time: 0:38:30
121
+ Epoch 1 | iter 576 step 18 | loss train: 1.462, val: 1.439 | iter time: 361.35 ms (step) remaining time: 0:38:19
122
+ Epoch 1 | iter 608 step 19 | loss train: 1.465, val: 1.439 | iter time: 359.94 ms (step) remaining time: 0:38:07
123
+ Epoch 1 | iter 640 step 20 | loss train: 1.490, val: 1.439 | iter time: 359.97 ms (step) remaining time: 0:37:56
124
+ Epoch 1 | iter 672 step 21 | loss train: 1.507, val: 1.439 | iter time: 361.58 ms (step) remaining time: 0:37:45
125
+ Epoch 1 | iter 704 step 22 | loss train: 1.431, val: 1.439 | iter time: 358.75 ms (step) remaining time: 0:37:34
126
+ Epoch 1 | iter 736 step 23 | loss train: 1.510, val: 1.439 | iter time: 359.75 ms (step) remaining time: 0:37:23
127
+ Epoch 1 | iter 768 step 24 | loss train: 1.512, val: 1.439 | iter time: 360.48 ms (step) remaining time: 0:37:11
128
+ Epoch 1 | iter 800 step 25 | loss train: 1.461, val: 1.439 | iter time: 360.71 ms (step) remaining time: 0:37:00
129
+ Epoch 1 | iter 832 step 26 | loss train: 1.365, val: 1.439 | iter time: 361.25 ms (step) remaining time: 0:36:49
130
+ Epoch 1 | iter 864 step 27 | loss train: 1.463, val: 1.439 | iter time: 360.68 ms (step) remaining time: 0:36:38
131
+ Epoch 1 | iter 896 step 28 | loss train: 1.444, val: 1.439 | iter time: 360.16 ms (step) remaining time: 0:36:27
132
+ Epoch 1 | iter 928 step 29 | loss train: 1.409, val: 1.439 | iter time: 359.23 ms (step) remaining time: 0:36:16
133
+ Epoch 1 | iter 960 step 30 | loss train: 1.474, val: 1.439 | iter time: 359.96 ms (step) remaining time: 0:36:05
134
+ Epoch 1 | iter 992 step 31 | loss train: 1.516, val: 1.439 | iter time: 359.90 ms (step) remaining time: 0:35:54
135
+ Epoch 1 | iter 1024 step 32 | loss train: 1.440, val: 1.439 | iter time: 359.33 ms (step) remaining time: 0:35:43
136
+ Epoch 1 | iter 1056 step 33 | loss train: 1.495, val: 1.439 | iter time: 360.16 ms (step) remaining time: 0:35:32
137
+ Epoch 1 | iter 1088 step 34 | loss train: 1.481, val: 1.439 | iter time: 360.33 ms (step) remaining time: 0:35:21
138
+ Epoch 1 | iter 1120 step 35 | loss train: 1.436, val: 1.439 | iter time: 360.18 ms (step) remaining time: 0:35:10
139
+ Epoch 1 | iter 1152 step 36 | loss train: 1.436, val: 1.439 | iter time: 360.03 ms (step) remaining time: 0:34:59
140
+ Epoch 1 | iter 1184 step 37 | loss train: 1.420, val: 1.439 | iter time: 360.55 ms (step) remaining time: 0:34:48
141
+ Epoch 1 | iter 1216 step 38 | loss train: 1.494, val: 1.439 | iter time: 358.50 ms (step) remaining time: 0:34:37
142
+ Epoch 1 | iter 1248 step 39 | loss train: 1.452, val: 1.439 | iter time: 360.03 ms (step) remaining time: 0:34:26
143
+ Epoch 1 | iter 1280 step 40 | loss train: 1.440, val: 1.439 | iter time: 359.54 ms (step) remaining time: 0:34:15
144
+ Epoch 1 | iter 1312 step 41 | loss train: 1.490, val: 1.439 | iter time: 360.37 ms (step) remaining time: 0:34:04
145
+ Epoch 1 | iter 1344 step 42 | loss train: 1.483, val: 1.439 | iter time: 358.71 ms (step) remaining time: 0:33:54
146
+ Epoch 1 | iter 1376 step 43 | loss train: 1.516, val: 1.439 | iter time: 360.13 ms (step) remaining time: 0:33:43
147
+ Epoch 1 | iter 1408 step 44 | loss train: 1.404, val: 1.439 | iter time: 359.22 ms (step) remaining time: 0:33:33
148
+ Epoch 1 | iter 1440 step 45 | loss train: 1.449, val: 1.439 | iter time: 359.40 ms (step) remaining time: 0:33:22
149
+ Epoch 1 | iter 1472 step 46 | loss train: 1.486, val: 1.439 | iter time: 358.81 ms (step) remaining time: 0:33:11
150
+ Epoch 1 | iter 1504 step 47 | loss train: 1.552, val: 1.439 | iter time: 360.34 ms (step) remaining time: 0:33:00
151
+ Epoch 1 | iter 1536 step 48 | loss train: 1.391, val: 1.439 | iter time: 358.45 ms (step) remaining time: 0:32:49
152
+ Epoch 1 | iter 1568 step 49 | loss train: 1.471, val: 1.439 | iter time: 360.69 ms (step) remaining time: 0:32:38
153
+ Epoch 1 | iter 1600 step 50 | loss train: 1.493, val: 1.439 | iter time: 361.05 ms (step) remaining time: 0:32:27
154
+ Validating ...
155
+ iter 1600: val loss 1.4279, val time: 22370.93 ms
156
+ Epoch 1 | iter 1632 step 51 | loss train: 1.459, val: 1.428 | iter time: 359.62 ms (step) remaining time: 0:33:35
157
+ Epoch 1 | iter 1664 step 52 | loss train: 1.479, val: 1.428 | iter time: 357.51 ms (step) remaining time: 0:33:22
158
+ Epoch 1 | iter 1696 step 53 | loss train: 1.395, val: 1.428 | iter time: 358.65 ms (step) remaining time: 0:33:09
159
+ Epoch 1 | iter 1728 step 54 | loss train: 1.450, val: 1.428 | iter time: 360.68 ms (step) remaining time: 0:32:56
160
+ Epoch 1 | iter 1760 step 55 | loss train: 1.430, val: 1.428 | iter time: 360.09 ms (step) remaining time: 0:32:44
161
+ Epoch 1 | iter 1792 step 56 | loss train: 1.498, val: 1.428 | iter time: 360.17 ms (step) remaining time: 0:32:31
162
+ Epoch 1 | iter 1824 step 57 | loss train: 1.504, val: 1.428 | iter time: 360.66 ms (step) remaining time: 0:32:19
163
+ Epoch 1 | iter 1856 step 58 | loss train: 1.457, val: 1.428 | iter time: 360.74 ms (step) remaining time: 0:32:06
164
+ Epoch 1 | iter 1888 step 59 | loss train: 1.462, val: 1.428 | iter time: 361.99 ms (step) remaining time: 0:31:54
165
+ Epoch 1 | iter 1920 step 60 | loss train: 1.434, val: 1.428 | iter time: 360.23 ms (step) remaining time: 0:31:41
166
+ Epoch 1 | iter 1952 step 61 | loss train: 1.400, val: 1.428 | iter time: 360.74 ms (step) remaining time: 0:31:29
167
+ Epoch 1 | iter 1984 step 62 | loss train: 1.460, val: 1.428 | iter time: 360.52 ms (step) remaining time: 0:31:17
168
+ Epoch 1 | iter 2016 step 63 | loss train: 1.443, val: 1.428 | iter time: 360.72 ms (step) remaining time: 0:31:04
169
+ Epoch 1 | iter 2048 step 64 | loss train: 1.438, val: 1.428 | iter time: 359.46 ms (step) remaining time: 0:30:52
170
+ Epoch 1 | iter 2080 step 65 | loss train: 1.438, val: 1.428 | iter time: 359.72 ms (step) remaining time: 0:30:40
171
+ Epoch 1 | iter 2112 step 66 | loss train: 1.476, val: 1.428 | iter time: 361.45 ms (step) remaining time: 0:30:28
172
+ Epoch 1 | iter 2144 step 67 | loss train: 1.395, val: 1.428 | iter time: 360.85 ms (step) remaining time: 0:30:16
173
+ Epoch 1 | iter 2176 step 68 | loss train: 1.451, val: 1.428 | iter time: 359.96 ms (step) remaining time: 0:30:04
174
+ Epoch 1 | iter 2208 step 69 | loss train: 1.467, val: 1.428 | iter time: 358.44 ms (step) remaining time: 0:29:52
175
+ Epoch 1 | iter 2240 step 70 | loss train: 1.470, val: 1.428 | iter time: 360.59 ms (step) remaining time: 0:29:40
176
+ Epoch 1 | iter 2272 step 71 | loss train: 1.403, val: 1.428 | iter time: 358.49 ms (step) remaining time: 0:29:28
177
+ Epoch 1 | iter 2304 step 72 | loss train: 1.482, val: 1.428 | iter time: 359.42 ms (step) remaining time: 0:29:16
178
+ Epoch 1 | iter 2336 step 73 | loss train: 1.478, val: 1.428 | iter time: 360.80 ms (step) remaining time: 0:29:04
179
+ Epoch 1 | iter 2368 step 74 | loss train: 1.512, val: 1.428 | iter time: 359.14 ms (step) remaining time: 0:28:52
180
+ Epoch 1 | iter 2400 step 75 | loss train: 1.437, val: 1.428 | iter time: 360.57 ms (step) remaining time: 0:28:40
181
+ Epoch 1 | iter 2432 step 76 | loss train: 1.382, val: 1.428 | iter time: 362.33 ms (step) remaining time: 0:28:29
182
+ Epoch 1 | iter 2464 step 77 | loss train: 1.438, val: 1.428 | iter time: 359.09 ms (step) remaining time: 0:28:17
183
+ Epoch 1 | iter 2496 step 78 | loss train: 1.485, val: 1.428 | iter time: 361.08 ms (step) remaining time: 0:28:05
184
+ Epoch 1 | iter 2528 step 79 | loss train: 1.483, val: 1.428 | iter time: 358.16 ms (step) remaining time: 0:27:53
185
+ Epoch 1 | iter 2560 step 80 | loss train: 1.489, val: 1.428 | iter time: 360.06 ms (step) remaining time: 0:27:42
186
+ Epoch 1 | iter 2592 step 81 | loss train: 1.389, val: 1.428 | iter time: 359.91 ms (step) remaining time: 0:27:30
187
+ Epoch 1 | iter 2624 step 82 | loss train: 1.435, val: 1.428 | iter time: 359.27 ms (step) remaining time: 0:27:18
188
+ Epoch 1 | iter 2656 step 83 | loss train: 1.439, val: 1.428 | iter time: 360.49 ms (step) remaining time: 0:27:07
189
+ Epoch 1 | iter 2688 step 84 | loss train: 1.498, val: 1.428 | iter time: 361.39 ms (step) remaining time: 0:26:55
190
+ Epoch 1 | iter 2720 step 85 | loss train: 1.384, val: 1.428 | iter time: 358.58 ms (step) remaining time: 0:26:43
191
+ Epoch 1 | iter 2752 step 86 | loss train: 1.432, val: 1.428 | iter time: 358.85 ms (step) remaining time: 0:26:32
192
+ Epoch 1 | iter 2784 step 87 | loss train: 1.430, val: 1.428 | iter time: 360.26 ms (step) remaining time: 0:26:20
193
+ Epoch 1 | iter 2816 step 88 | loss train: 1.488, val: 1.428 | iter time: 359.59 ms (step) remaining time: 0:26:09
194
+ Epoch 1 | iter 2848 step 89 | loss train: 1.408, val: 1.428 | iter time: 359.99 ms (step) remaining time: 0:25:57
195
+ Epoch 1 | iter 2880 step 90 | loss train: 1.478, val: 1.428 | iter time: 360.41 ms (step) remaining time: 0:25:46
196
+ Epoch 1 | iter 2912 step 91 | loss train: 1.427, val: 1.428 | iter time: 359.02 ms (step) remaining time: 0:25:34
197
+ Epoch 1 | iter 2944 step 92 | loss train: 1.398, val: 1.428 | iter time: 359.95 ms (step) remaining time: 0:25:22
198
+ Epoch 1 | iter 2976 step 93 | loss train: 1.528, val: 1.428 | iter time: 359.32 ms (step) remaining time: 0:25:11
199
+ Epoch 1 | iter 3008 step 94 | loss train: 1.456, val: 1.428 | iter time: 358.77 ms (step) remaining time: 0:25:00
200
+ Epoch 1 | iter 3040 step 95 | loss train: 1.488, val: 1.428 | iter time: 361.67 ms (step) remaining time: 0:24:48
201
+ Epoch 1 | iter 3072 step 96 | loss train: 1.486, val: 1.428 | iter time: 358.49 ms (step) remaining time: 0:24:37
202
+ Epoch 1 | iter 3104 step 97 | loss train: 1.414, val: 1.428 | iter time: 362.04 ms (step) remaining time: 0:24:25
203
+ Epoch 1 | iter 3136 step 98 | loss train: 1.391, val: 1.428 | iter time: 358.65 ms (step) remaining time: 0:24:14
204
+ Epoch 1 | iter 3168 step 99 | loss train: 1.472, val: 1.428 | iter time: 359.13 ms (step) remaining time: 0:24:02
205
+ Epoch 1 | iter 3200 step 100 | loss train: 1.409, val: 1.428 | iter time: 360.12 ms (step) remaining time: 0:23:51
206
+ Validating ...
207
+ iter 3200: val loss 1.3742, val time: 22354.24 ms
208
+ Saving checkpoint to 'out/pretrain/2409_full/step-00000100/lit_model.pth'
209
+ Epoch 1 | iter 3232 step 101 | loss train: 1.449, val: 1.374 | iter time: 357.59 ms (step) remaining time: 0:24:29
210
+ Epoch 1 | iter 3264 step 102 | loss train: 1.449, val: 1.374 | iter time: 357.88 ms (step) remaining time: 0:24:17
211
+ Epoch 1 | iter 3296 step 103 | loss train: 1.492, val: 1.374 | iter time: 359.03 ms (step) remaining time: 0:24:05
212
+ Epoch 1 | iter 3328 step 104 | loss train: 1.441, val: 1.374 | iter time: 359.20 ms (step) remaining time: 0:23:53
213
+ Epoch 1 | iter 3360 step 105 | loss train: 1.439, val: 1.374 | iter time: 359.20 ms (step) remaining time: 0:23:40
214
+ Epoch 1 | iter 3392 step 106 | loss train: 1.409, val: 1.374 | iter time: 359.27 ms (step) remaining time: 0:23:28
215
+ Epoch 1 | iter 3424 step 107 | loss train: 1.403, val: 1.374 | iter time: 360.41 ms (step) remaining time: 0:23:16
216
+ Epoch 1 | iter 3456 step 108 | loss train: 1.461, val: 1.374 | iter time: 359.41 ms (step) remaining time: 0:23:04
217
+ Epoch 1 | iter 3488 step 109 | loss train: 1.417, val: 1.374 | iter time: 360.49 ms (step) remaining time: 0:22:52
218
+ Epoch 1 | iter 3520 step 110 | loss train: 1.439, val: 1.374 | iter time: 359.24 ms (step) remaining time: 0:22:40
219
+ Epoch 1 | iter 3552 step 111 | loss train: 1.391, val: 1.374 | iter time: 358.84 ms (step) remaining time: 0:22:28
220
+ Epoch 1 | iter 3584 step 112 | loss train: 1.442, val: 1.374 | iter time: 357.94 ms (step) remaining time: 0:22:16
221
+ Epoch 1 | iter 3616 step 113 | loss train: 1.433, val: 1.374 | iter time: 359.61 ms (step) remaining time: 0:22:04
222
+ Epoch 1 | iter 3648 step 114 | loss train: 1.371, val: 1.374 | iter time: 361.47 ms (step) remaining time: 0:21:52
223
+ Epoch 1 | iter 3680 step 115 | loss train: 1.525, val: 1.374 | iter time: 359.78 ms (step) remaining time: 0:21:40
224
+ Epoch 1 | iter 3712 step 116 | loss train: 1.445, val: 1.374 | iter time: 360.91 ms (step) remaining time: 0:21:28
225
+ Epoch 1 | iter 3744 step 117 | loss train: 1.500, val: 1.374 | iter time: 360.49 ms (step) remaining time: 0:21:17
226
+ Epoch 1 | iter 3776 step 118 | loss train: 1.409, val: 1.374 | iter time: 360.23 ms (step) remaining time: 0:21:05
227
+ Epoch 1 | iter 3808 step 119 | loss train: 1.398, val: 1.374 | iter time: 358.15 ms (step) remaining time: 0:20:53
228
+ Epoch 1 | iter 3840 step 120 | loss train: 1.438, val: 1.374 | iter time: 357.77 ms (step) remaining time: 0:20:41
229
+ Epoch 1 | iter 3872 step 121 | loss train: 1.486, val: 1.374 | iter time: 362.11 ms (step) remaining time: 0:20:29
230
+ Epoch 1 | iter 3904 step 122 | loss train: 1.431, val: 1.374 | iter time: 361.26 ms (step) remaining time: 0:20:17
231
+ Epoch 1 | iter 3936 step 123 | loss train: 1.366, val: 1.374 | iter time: 360.88 ms (step) remaining time: 0:20:05
232
+ Epoch 1 | iter 3968 step 124 | loss train: 1.379, val: 1.374 | iter time: 359.94 ms (step) remaining time: 0:19:53
233
+ Epoch 1 | iter 4000 step 125 | loss train: 1.492, val: 1.374 | iter time: 359.95 ms (step) remaining time: 0:19:42
234
+ Epoch 1 | iter 4032 step 126 | loss train: 1.510, val: 1.374 | iter time: 360.46 ms (step) remaining time: 0:19:30
235
+ Epoch 1 | iter 4064 step 127 | loss train: 1.388, val: 1.374 | iter time: 359.83 ms (step) remaining time: 0:19:18
236
+ Epoch 1 | iter 4096 step 128 | loss train: 1.392, val: 1.374 | iter time: 360.20 ms (step) remaining time: 0:19:06
237
+ Epoch 1 | iter 4128 step 129 | loss train: 1.424, val: 1.374 | iter time: 360.62 ms (step) remaining time: 0:18:55
238
+ Epoch 1 | iter 4160 step 130 | loss train: 1.378, val: 1.374 | iter time: 359.19 ms (step) remaining time: 0:18:43
239
+ Epoch 1 | iter 4192 step 131 | loss train: 1.381, val: 1.374 | iter time: 360.42 ms (step) remaining time: 0:18:31
240
+ Epoch 1 | iter 4224 step 132 | loss train: 1.448, val: 1.374 | iter time: 357.72 ms (step) remaining time: 0:18:20
241
+ Epoch 1 | iter 4256 step 133 | loss train: 1.352, val: 1.374 | iter time: 360.27 ms (step) remaining time: 0:18:08
242
+ Epoch 1 | iter 4288 step 134 | loss train: 1.428, val: 1.374 | iter time: 357.85 ms (step) remaining time: 0:17:56
243
+ Epoch 1 | iter 4320 step 135 | loss train: 1.455, val: 1.374 | iter time: 360.12 ms (step) remaining time: 0:17:44
244
+ Epoch 1 | iter 4352 step 136 | loss train: 1.404, val: 1.374 | iter time: 358.96 ms (step) remaining time: 0:17:33
245
+ Epoch 1 | iter 4384 step 137 | loss train: 1.408, val: 1.374 | iter time: 360.78 ms (step) remaining time: 0:17:21
246
+ Epoch 1 | iter 4416 step 138 | loss train: 1.399, val: 1.374 | iter time: 359.31 ms (step) remaining time: 0:17:10
247
+ Epoch 1 | iter 4448 step 139 | loss train: 1.377, val: 1.374 | iter time: 360.01 ms (step) remaining time: 0:16:58
248
+ Epoch 1 | iter 4480 step 140 | loss train: 1.430, val: 1.374 | iter time: 361.65 ms (step) remaining time: 0:16:46
249
+ Epoch 1 | iter 4512 step 141 | loss train: 1.430, val: 1.374 | iter time: 359.36 ms (step) remaining time: 0:16:35
250
+ Epoch 1 | iter 4544 step 142 | loss train: 1.393, val: 1.374 | iter time: 359.19 ms (step) remaining time: 0:16:23
251
+ Epoch 1 | iter 4576 step 143 | loss train: 1.414, val: 1.374 | iter time: 361.11 ms (step) remaining time: 0:16:12
252
+ Epoch 1 | iter 4608 step 144 | loss train: 1.416, val: 1.374 | iter time: 359.63 ms (step) remaining time: 0:16:00
253
+ Epoch 1 | iter 4640 step 145 | loss train: 1.502, val: 1.374 | iter time: 360.45 ms (step) remaining time: 0:15:48
254
+ Epoch 1 | iter 4672 step 146 | loss train: 1.401, val: 1.374 | iter time: 360.49 ms (step) remaining time: 0:15:37
255
+ Epoch 1 | iter 4704 step 147 | loss train: 1.426, val: 1.374 | iter time: 359.62 ms (step) remaining time: 0:15:25
256
+ Epoch 1 | iter 4736 step 148 | loss train: 1.469, val: 1.374 | iter time: 359.37 ms (step) remaining time: 0:15:14
257
+ Epoch 1 | iter 4768 step 149 | loss train: 1.379, val: 1.374 | iter time: 361.42 ms (step) remaining time: 0:15:02
258
+ Epoch 1 | iter 4800 step 150 | loss train: 1.434, val: 1.374 | iter time: 360.99 ms (step) remaining time: 0:14:51
259
+ Validating ...
260
+ iter 4800: val loss 1.3083, val time: 22361.91 ms
261
+ Epoch 1 | iter 4832 step 151 | loss train: 1.320, val: 1.308 | iter time: 360.14 ms (step) remaining time: 0:14:51
262
+ Epoch 1 | iter 4864 step 152 | loss train: 1.411, val: 1.308 | iter time: 359.02 ms (step) remaining time: 0:14:39
263
+ Epoch 1 | iter 4896 step 153 | loss train: 1.337, val: 1.308 | iter time: 360.17 ms (step) remaining time: 0:14:28
264
+ Epoch 1 | iter 4928 step 154 | loss train: 1.460, val: 1.308 | iter time: 360.78 ms (step) remaining time: 0:14:16
265
+ Epoch 1 | iter 4960 step 155 | loss train: 1.398, val: 1.308 | iter time: 360.32 ms (step) remaining time: 0:14:04
266
+ Epoch 1 | iter 4992 step 156 | loss train: 1.371, val: 1.308 | iter time: 361.74 ms (step) remaining time: 0:13:53
267
+ Epoch 1 | iter 5024 step 157 | loss train: 1.401, val: 1.308 | iter time: 361.55 ms (step) remaining time: 0:13:41
268
+ Epoch 1 | iter 5056 step 158 | loss train: 1.346, val: 1.308 | iter time: 361.36 ms (step) remaining time: 0:13:29
269
+ Epoch 1 | iter 5088 step 159 | loss train: 1.330, val: 1.308 | iter time: 359.34 ms (step) remaining time: 0:13:18
270
+ Epoch 1 | iter 5120 step 160 | loss train: 1.384, val: 1.308 | iter time: 357.55 ms (step) remaining time: 0:13:06
271
+ Epoch 1 | iter 5152 step 161 | loss train: 1.376, val: 1.308 | iter time: 359.88 ms (step) remaining time: 0:12:54
272
+ Epoch 1 | iter 5184 step 162 | loss train: 1.365, val: 1.308 | iter time: 359.41 ms (step) remaining time: 0:12:43
273
+ Epoch 1 | iter 5216 step 163 | loss train: 1.375, val: 1.308 | iter time: 359.51 ms (step) remaining time: 0:12:31
274
+ Epoch 1 | iter 5248 step 164 | loss train: 1.334, val: 1.308 | iter time: 360.30 ms (step) remaining time: 0:12:20
275
+ Epoch 1 | iter 5280 step 165 | loss train: 1.400, val: 1.308 | iter time: 361.56 ms (step) remaining time: 0:12:08
276
+ Epoch 1 | iter 5312 step 166 | loss train: 1.362, val: 1.308 | iter time: 362.95 ms (step) remaining time: 0:11:56
277
+ Epoch 1 | iter 5344 step 167 | loss train: 1.384, val: 1.308 | iter time: 360.51 ms (step) remaining time: 0:11:45
278
+ Epoch 1 | iter 5376 step 168 | loss train: 1.360, val: 1.308 | iter time: 360.44 ms (step) remaining time: 0:11:33
279
+ Epoch 1 | iter 5408 step 169 | loss train: 1.392, val: 1.308 | iter time: 360.19 ms (step) remaining time: 0:11:22
280
+ Epoch 1 | iter 5440 step 170 | loss train: 1.446, val: 1.308 | iter time: 359.28 ms (step) remaining time: 0:11:10
281
+ Epoch 1 | iter 5472 step 171 | loss train: 1.316, val: 1.308 | iter time: 359.45 ms (step) remaining time: 0:10:59
282
+ Epoch 1 | iter 5504 step 172 | loss train: 1.331, val: 1.308 | iter time: 360.01 ms (step) remaining time: 0:10:47
283
+ Epoch 1 | iter 5536 step 173 | loss train: 1.437, val: 1.308 | iter time: 358.74 ms (step) remaining time: 0:10:36
284
+ Epoch 1 | iter 5568 step 174 | loss train: 1.322, val: 1.308 | iter time: 359.81 ms (step) remaining time: 0:10:24
285
+ Epoch 1 | iter 5600 step 175 | loss train: 1.387, val: 1.308 | iter time: 361.16 ms (step) remaining time: 0:10:13
286
+ Epoch 1 | iter 5632 step 176 | loss train: 1.407, val: 1.308 | iter time: 360.28 ms (step) remaining time: 0:10:01
287
+ Epoch 1 | iter 5664 step 177 | loss train: 1.454, val: 1.308 | iter time: 359.64 ms (step) remaining time: 0:09:50
288
+ Epoch 1 | iter 5696 step 178 | loss train: 1.386, val: 1.308 | iter time: 359.70 ms (step) remaining time: 0:09:38
289
+ Epoch 1 | iter 5728 step 179 | loss train: 1.326, val: 1.308 | iter time: 362.08 ms (step) remaining time: 0:09:27
290
+ Epoch 1 | iter 5760 step 180 | loss train: 1.391, val: 1.308 | iter time: 359.35 ms (step) remaining time: 0:09:15
291
+ Epoch 1 | iter 5792 step 181 | loss train: 1.385, val: 1.308 | iter time: 361.94 ms (step) remaining time: 0:09:04
292
+ Epoch 1 | iter 5824 step 182 | loss train: 1.405, val: 1.308 | iter time: 359.97 ms (step) remaining time: 0:08:52
293
+ Epoch 1 | iter 5856 step 183 | loss train: 1.361, val: 1.308 | iter time: 360.45 ms (step) remaining time: 0:08:41
294
+ Epoch 1 | iter 5888 step 184 | loss train: 1.386, val: 1.308 | iter time: 358.21 ms (step) remaining time: 0:08:30
295
+ Epoch 1 | iter 5920 step 185 | loss train: 1.321, val: 1.308 | iter time: 359.45 ms (step) remaining time: 0:08:18
296
+ Epoch 1 | iter 5952 step 186 | loss train: 1.348, val: 1.308 | iter time: 361.57 ms (step) remaining time: 0:08:07
297
+ Epoch 1 | iter 5984 step 187 | loss train: 1.352, val: 1.308 | iter time: 359.83 ms (step) remaining time: 0:07:55
298
+ Epoch 1 | iter 6016 step 188 | loss train: 1.381, val: 1.308 | iter time: 360.83 ms (step) remaining time: 0:07:44
299
+ Epoch 1 | iter 6048 step 189 | loss train: 1.380, val: 1.308 | iter time: 359.87 ms (step) remaining time: 0:07:32
300
+ Epoch 1 | iter 6080 step 190 | loss train: 1.356, val: 1.308 | iter time: 359.57 ms (step) remaining time: 0:07:21
301
+ Epoch 1 | iter 6112 step 191 | loss train: 1.357, val: 1.308 | iter time: 360.99 ms (step) remaining time: 0:07:10
302
+ Epoch 1 | iter 6144 step 192 | loss train: 1.382, val: 1.308 | iter time: 359.41 ms (step) remaining time: 0:06:58
303
+ Epoch 1 | iter 6176 step 193 | loss train: 1.419, val: 1.308 | iter time: 362.76 ms (step) remaining time: 0:06:47
304
+ Epoch 1 | iter 6208 step 194 | loss train: 1.396, val: 1.308 | iter time: 357.99 ms (step) remaining time: 0:06:35
305
+ Epoch 1 | iter 6240 step 195 | loss train: 1.379, val: 1.308 | iter time: 360.59 ms (step) remaining time: 0:06:24
306
+ Epoch 1 | iter 6272 step 196 | loss train: 1.303, val: 1.308 | iter time: 360.35 ms (step) remaining time: 0:06:13
307
+ Epoch 1 | iter 6304 step 197 | loss train: 1.319, val: 1.308 | iter time: 361.73 ms (step) remaining time: 0:06:01
308
+ Epoch 1 | iter 6336 step 198 | loss train: 1.363, val: 1.308 | iter time: 361.68 ms (step) remaining time: 0:05:50
309
+ Epoch 1 | iter 6368 step 199 | loss train: 1.414, val: 1.308 | iter time: 360.85 ms (step) remaining time: 0:05:38
310
+ Epoch 1 | iter 6400 step 200 | loss train: 1.426, val: 1.308 | iter time: 359.01 ms (step) remaining time: 0:05:27
311
+ Validating ...
312
+ iter 6400: val loss 1.2540, val time: 22363.57 ms
313
+ Saving checkpoint to 'out/pretrain/2409_full/step-00000200/lit_model.pth'
314
+ Epoch 1 | iter 6432 step 201 | loss train: 1.349, val: 1.254 | iter time: 357.41 ms (step) remaining time: 0:05:21
315
+ Epoch 1 | iter 6464 step 202 | loss train: 1.325, val: 1.254 | iter time: 356.63 ms (step) remaining time: 0:05:10
316
+ Epoch 1 | iter 6496 step 203 | loss train: 1.364, val: 1.254 | iter time: 360.22 ms (step) remaining time: 0:04:58
317
+ Epoch 1 | iter 6528 step 204 | loss train: 1.339, val: 1.254 | iter time: 360.66 ms (step) remaining time: 0:04:47
318
+ Epoch 1 | iter 6560 step 205 | loss train: 1.444, val: 1.254 | iter time: 359.77 ms (step) remaining time: 0:04:35
319
+ Epoch 1 | iter 6592 step 206 | loss train: 1.366, val: 1.254 | iter time: 358.90 ms (step) remaining time: 0:04:23
320
+ Epoch 1 | iter 6624 step 207 | loss train: 1.327, val: 1.254 | iter time: 360.62 ms (step) remaining time: 0:04:12
321
+ Epoch 1 | iter 6656 step 208 | loss train: 1.382, val: 1.254 | iter time: 359.44 ms (step) remaining time: 0:04:00
322
+ Epoch 1 | iter 6688 step 209 | loss train: 1.393, val: 1.254 | iter time: 360.20 ms (step) remaining time: 0:03:49
323
+ Epoch 1 | iter 6720 step 210 | loss train: 1.318, val: 1.254 | iter time: 360.84 ms (step) remaining time: 0:03:37
324
+ Epoch 1 | iter 6752 step 211 | loss train: 1.447, val: 1.254 | iter time: 358.12 ms (step) remaining time: 0:03:26
325
+ Epoch 1 | iter 6784 step 212 | loss train: 1.398, val: 1.254 | iter time: 359.93 ms (step) remaining time: 0:03:14
326
+ Epoch 1 | iter 6816 step 213 | loss train: 1.424, val: 1.254 | iter time: 361.55 ms (step) remaining time: 0:03:03
327
+ Epoch 1 | iter 6848 step 214 | loss train: 1.363, val: 1.254 | iter time: 358.52 ms (step) remaining time: 0:02:51
328
+ Epoch 1 | iter 6880 step 215 | loss train: 1.384, val: 1.254 | iter time: 359.94 ms (step) remaining time: 0:02:40
329
+ Epoch 1 | iter 6912 step 216 | loss train: 1.361, val: 1.254 | iter time: 360.33 ms (step) remaining time: 0:02:28
330
+ Epoch 1 | iter 6944 step 217 | loss train: 1.370, val: 1.254 | iter time: 359.63 ms (step) remaining time: 0:02:17
331
+ Epoch 1 | iter 6976 step 218 | loss train: 1.361, val: 1.254 | iter time: 359.02 ms (step) remaining time: 0:02:05
332
+ Epoch 1 | iter 7008 step 219 | loss train: 1.324, val: 1.254 | iter time: 359.74 ms (step) remaining time: 0:01:54
333
+ Epoch 1 | iter 7040 step 220 | loss train: 1.320, val: 1.254 | iter time: 359.73 ms (step) remaining time: 0:01:42
334
+ Epoch 1 | iter 7072 step 221 | loss train: 1.405, val: 1.254 | iter time: 359.86 ms (step) remaining time: 0:01:31
335
+ Epoch 1 | iter 7104 step 222 | loss train: 1.370, val: 1.254 | iter time: 360.54 ms (step) remaining time: 0:01:20
336
+ Epoch 1 | iter 7136 step 223 | loss train: 1.352, val: 1.254 | iter time: 360.82 ms (step) remaining time: 0:01:08
337
+ Epoch 1 | iter 7168 step 224 | loss train: 1.415, val: 1.254 | iter time: 360.62 ms (step) remaining time: 0:00:57
338
+ Epoch 1 | iter 7200 step 225 | loss train: 1.401, val: 1.254 | iter time: 358.78 ms (step) remaining time: 0:00:45
339
+ Epoch 1 | iter 7232 step 226 | loss train: 1.302, val: 1.254 | iter time: 360.02 ms (step) remaining time: 0:00:34
340
+ Epoch 1 | iter 7264 step 227 | loss train: 1.352, val: 1.254 | iter time: 369.07 ms (step) remaining time: 0:00:22
341
+ Epoch 1 | iter 7296 step 228 | loss train: 1.362, val: 1.254 | iter time: 360.42 ms (step) remaining time: 0:00:11
342
+ Epoch 2 | iter 7328 step 229 | loss train: 1.286, val: 1.254 | iter time: 361.71 ms (step) remaining time: 0:00:00
343
+ Validating ...
344
+ Final evaluation | val loss: 1.232 | val ppl: 3.430
345
+ Saving checkpoint to 'out/pretrain/2409_full/final/lit_model.pth'
346
+ ----------------------------------------
347
+ | Performance
348
+ | - Total tokens : 240,123,904
349
+ | - Training Time : 2679.88 s
350
+ | - Tok/sec : 129.83 tok/s
351
+ | ----------------------------------------
352
+ | Memory Usage
353
+ | - Memory Used : 26.32 GB
354
+ ----------------------------------------
out/pretrain/tinyllama/teelogs/2409_lr4e-5.txt ADDED
@@ -0,0 +1,361 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Initializing distributed: GLOBAL_RANK: 0, MEMBER: 1/4
2
+ Initializing distributed: GLOBAL_RANK: 2, MEMBER: 3/4
3
+ [rank: 2] Seed set to 42
4
+ Initializing distributed: GLOBAL_RANK: 3, MEMBER: 4/4
5
+ [rank: 3] Seed set to 42
6
+ Initializing distributed: GLOBAL_RANK: 1, MEMBER: 2/4
7
+ ----------------------------------------------------------------------------------------------------
8
+ distributed_backend=nccl
9
+ All distributed processes registered. Starting with 4 processes
10
+ ----------------------------------------------------------------------------------------------------
11
+
12
+ [rank: 1] Seed set to 42
13
+ [rank: 0] Seed set to 42
14
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
15
+ train_dataloader = data.train_dataloader()
16
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
17
+ train_dataloader = data.train_dataloader()
18
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
19
+ train_dataloader = data.train_dataloader()
20
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
21
+ train_dataloader = data.train_dataloader()
22
+ All GPUs are fully connected via NVLink.
23
+ {'data': {'batch_size': 1,
24
+ 'data_path': PosixPath('litgpt/data/arxiv'),
25
+ 'num_workers': 0,
26
+ 'seed': 42,
27
+ 'seq_length': 2048,
28
+ 'use_starcoder': True},
29
+ 'data_dir': PosixPath('litgpt/data/arxiv/2409'),
30
+ 'devices': 'auto',
31
+ 'eval': {'evaluate_example': 'first',
32
+ 'final_validation': True,
33
+ 'initial_validation': True,
34
+ 'interval': 50,
35
+ 'max_iters': 200,
36
+ 'max_new_tokens': None},
37
+ 'initial_checkpoint_dir': PosixPath('out/pretrain/tinyllama/2408_full/final'),
38
+ 'log': {'group': None, 'project': None, 'run': None},
39
+ 'logger_name': 'tensorboard',
40
+ 'model_config': {'attention_logit_softcapping': None,
41
+ 'attention_scores_scalar': None,
42
+ 'attn_bias': False,
43
+ 'bias': False,
44
+ 'block_size': 2048,
45
+ 'final_logit_softcapping': None,
46
+ 'gelu_approximate': 'none',
47
+ 'head_size': 64,
48
+ 'hf_config': {'name': 'TinyLlama-1.1B-intermediate-step-1431k-3T',
49
+ 'org': 'TinyLlama'},
50
+ 'intermediate_size': 5632,
51
+ 'lm_head_bias': False,
52
+ 'mlp_class_name': 'LLaMAMLP',
53
+ 'moe_intermediate_size': None,
54
+ 'n_embd': 2048,
55
+ 'n_expert': 0,
56
+ 'n_expert_per_token': 0,
57
+ 'n_head': 32,
58
+ 'n_layer': 22,
59
+ 'n_query_groups': 4,
60
+ 'name': 'tiny-llama-1.1b',
61
+ 'norm_1': True,
62
+ 'norm_2': True,
63
+ 'norm_class_name': 'RMSNorm',
64
+ 'norm_eps': 1e-05,
65
+ 'norm_qk': False,
66
+ 'norm_qk_type': 'default',
67
+ 'padded_vocab_size': 32000,
68
+ 'padding_multiple': 64,
69
+ 'parallel_residual': False,
70
+ 'post_attention_norm': False,
71
+ 'post_mlp_norm': False,
72
+ 'rope_adjustments': None,
73
+ 'rope_base': 10000,
74
+ 'rope_condense_ratio': 1,
75
+ 'rope_indices': None,
76
+ 'rope_local_base_freq': None,
77
+ 'rotary_percentage': 1.0,
78
+ 'scale_embeddings': False,
79
+ 'shared_attention_norm': False,
80
+ 'sliding_window_indices': None,
81
+ 'sliding_window_size': None,
82
+ 'vocab_size': 32000},
83
+ 'model_name': 'tiny-llama-1.1b',
84
+ 'num_nodes': 1,
85
+ 'optimizer': "{'class_path': 'torch.optim.AdamW', 'init_args': {'lr': 4e-05, "
86
+ "'weight_decay': 0.1, 'betas': [0.9, 0.95]}}",
87
+ 'out_dir': PosixPath('out/pretrain/tinyllama/2409_lr4e-5'),
88
+ 'ppl': False,
89
+ 'precision': 'bf16-mixed',
90
+ 'resume': False,
91
+ 'seed': 42,
92
+ 'tokenizer_dir': PosixPath('checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T'),
93
+ 'train': {'epochs': None,
94
+ 'global_batch_size': 512,
95
+ 'log_interval': 1,
96
+ 'lr_warmup_fraction': None,
97
+ 'lr_warmup_steps': 20,
98
+ 'max_norm': 1.0,
99
+ 'max_seq_length': 2048,
100
+ 'max_steps': None,
101
+ 'max_tokens': 240123904,
102
+ 'micro_batch_size': 4,
103
+ 'min_lr': 4e-05,
104
+ 'save_interval': 100,
105
+ 'tie_embeddings': None}}
106
+ Time to instantiate model: 0.04 seconds.
107
+ Total parameters: 1,100,048,384
108
+ [ok] out/pretrain/tinyllama/2408_full/final/lit_model.pth 已是纯权重
109
+ Validating ...
110
+ Measured TFLOPs: 239.66
111
+ Epoch 1 | iter 32 step 1 | loss train: 1.492, val: 1.476 | iter time: 560.28 ms (step) remaining time: 0:44:23
112
+ Epoch 1 | iter 64 step 2 | loss train: 1.354, val: 1.476 | iter time: 356.78 ms (step) remaining time: 0:42:29
113
+ Epoch 1 | iter 96 step 3 | loss train: 1.430, val: 1.476 | iter time: 358.40 ms (step) remaining time: 0:41:52
114
+ Epoch 1 | iter 128 step 4 | loss train: 1.396, val: 1.476 | iter time: 358.67 ms (step) remaining time: 0:41:24
115
+ Epoch 1 | iter 160 step 5 | loss train: 1.335, val: 1.476 | iter time: 358.23 ms (step) remaining time: 0:41:03
116
+ Epoch 1 | iter 192 step 6 | loss train: 1.487, val: 1.476 | iter time: 358.89 ms (step) remaining time: 0:40:47
117
+ Epoch 1 | iter 224 step 7 | loss train: 1.408, val: 1.476 | iter time: 358.29 ms (step) remaining time: 0:40:32
118
+ Epoch 1 | iter 256 step 8 | loss train: 1.463, val: 1.476 | iter time: 359.94 ms (step) remaining time: 0:40:18
119
+ Epoch 1 | iter 288 step 9 | loss train: 1.382, val: 1.476 | iter time: 358.60 ms (step) remaining time: 0:40:05
120
+ Epoch 1 | iter 320 step 10 | loss train: 1.388, val: 1.476 | iter time: 359.31 ms (step) remaining time: 0:39:52
121
+ Epoch 1 | iter 352 step 11 | loss train: 1.447, val: 1.476 | iter time: 359.66 ms (step) remaining time: 0:39:40
122
+ Epoch 1 | iter 384 step 12 | loss train: 1.524, val: 1.476 | iter time: 360.06 ms (step) remaining time: 0:39:28
123
+ Epoch 1 | iter 416 step 13 | loss train: 1.444, val: 1.476 | iter time: 358.06 ms (step) remaining time: 0:39:16
124
+ Epoch 1 | iter 448 step 14 | loss train: 1.431, val: 1.476 | iter time: 358.18 ms (step) remaining time: 0:39:05
125
+ Epoch 1 | iter 480 step 15 | loss train: 1.472, val: 1.476 | iter time: 359.05 ms (step) remaining time: 0:38:53
126
+ Epoch 1 | iter 512 step 16 | loss train: 1.468, val: 1.476 | iter time: 360.11 ms (step) remaining time: 0:38:42
127
+ Epoch 1 | iter 544 step 17 | loss train: 1.402, val: 1.476 | iter time: 360.00 ms (step) remaining time: 0:38:30
128
+ Epoch 1 | iter 576 step 18 | loss train: 1.418, val: 1.476 | iter time: 357.75 ms (step) remaining time: 0:38:19
129
+ Epoch 1 | iter 608 step 19 | loss train: 1.514, val: 1.476 | iter time: 358.85 ms (step) remaining time: 0:38:08
130
+ Epoch 1 | iter 640 step 20 | loss train: 1.380, val: 1.476 | iter time: 359.64 ms (step) remaining time: 0:37:56
131
+ Epoch 1 | iter 672 step 21 | loss train: 1.510, val: 1.476 | iter time: 359.77 ms (step) remaining time: 0:37:45
132
+ Epoch 1 | iter 704 step 22 | loss train: 1.443, val: 1.476 | iter time: 359.87 ms (step) remaining time: 0:37:34
133
+ Epoch 1 | iter 736 step 23 | loss train: 1.398, val: 1.476 | iter time: 363.44 ms (step) remaining time: 0:37:23
134
+ Epoch 1 | iter 768 step 24 | loss train: 1.455, val: 1.476 | iter time: 358.04 ms (step) remaining time: 0:37:12
135
+ Epoch 1 | iter 800 step 25 | loss train: 1.467, val: 1.476 | iter time: 359.13 ms (step) remaining time: 0:37:01
136
+ Epoch 1 | iter 832 step 26 | loss train: 1.532, val: 1.476 | iter time: 360.82 ms (step) remaining time: 0:36:50
137
+ Epoch 1 | iter 864 step 27 | loss train: 1.453, val: 1.476 | iter time: 360.87 ms (step) remaining time: 0:36:39
138
+ Epoch 1 | iter 896 step 28 | loss train: 1.494, val: 1.476 | iter time: 360.39 ms (step) remaining time: 0:36:28
139
+ Epoch 1 | iter 928 step 29 | loss train: 1.432, val: 1.476 | iter time: 360.38 ms (step) remaining time: 0:36:17
140
+ Epoch 1 | iter 960 step 30 | loss train: 1.460, val: 1.476 | iter time: 358.57 ms (step) remaining time: 0:36:06
141
+ Epoch 1 | iter 992 step 31 | loss train: 1.431, val: 1.476 | iter time: 359.41 ms (step) remaining time: 0:35:55
142
+ Epoch 1 | iter 1024 step 32 | loss train: 1.461, val: 1.476 | iter time: 361.23 ms (step) remaining time: 0:35:44
143
+ Epoch 1 | iter 1056 step 33 | loss train: 1.468, val: 1.476 | iter time: 361.40 ms (step) remaining time: 0:35:33
144
+ Epoch 1 | iter 1088 step 34 | loss train: 1.357, val: 1.476 | iter time: 361.12 ms (step) remaining time: 0:35:22
145
+ Epoch 1 | iter 1120 step 35 | loss train: 1.400, val: 1.476 | iter time: 359.38 ms (step) remaining time: 0:35:11
146
+ Epoch 1 | iter 1152 step 36 | loss train: 1.451, val: 1.476 | iter time: 360.41 ms (step) remaining time: 0:35:00
147
+ Epoch 1 | iter 1184 step 37 | loss train: 1.423, val: 1.476 | iter time: 358.08 ms (step) remaining time: 0:34:49
148
+ Epoch 1 | iter 1216 step 38 | loss train: 1.326, val: 1.476 | iter time: 359.41 ms (step) remaining time: 0:34:38
149
+ Epoch 1 | iter 1248 step 39 | loss train: 1.431, val: 1.476 | iter time: 360.46 ms (step) remaining time: 0:34:27
150
+ Epoch 1 | iter 1280 step 40 | loss train: 1.335, val: 1.476 | iter time: 358.80 ms (step) remaining time: 0:34:16
151
+ Epoch 1 | iter 1312 step 41 | loss train: 1.472, val: 1.476 | iter time: 359.76 ms (step) remaining time: 0:34:05
152
+ Epoch 1 | iter 1344 step 42 | loss train: 1.416, val: 1.476 | iter time: 360.22 ms (step) remaining time: 0:33:54
153
+ Epoch 1 | iter 1376 step 43 | loss train: 1.439, val: 1.476 | iter time: 361.07 ms (step) remaining time: 0:33:43
154
+ Epoch 1 | iter 1408 step 44 | loss train: 1.322, val: 1.476 | iter time: 358.30 ms (step) remaining time: 0:33:32
155
+ Epoch 1 | iter 1440 step 45 | loss train: 1.465, val: 1.476 | iter time: 358.79 ms (step) remaining time: 0:33:21
156
+ Epoch 1 | iter 1472 step 46 | loss train: 1.410, val: 1.476 | iter time: 359.17 ms (step) remaining time: 0:33:11
157
+ Epoch 1 | iter 1504 step 47 | loss train: 1.391, val: 1.476 | iter time: 359.52 ms (step) remaining time: 0:33:00
158
+ Epoch 1 | iter 1536 step 48 | loss train: 1.442, val: 1.476 | iter time: 358.06 ms (step) remaining time: 0:32:49
159
+ Epoch 1 | iter 1568 step 49 | loss train: 1.426, val: 1.476 | iter time: 359.79 ms (step) remaining time: 0:32:38
160
+ Epoch 1 | iter 1600 step 50 | loss train: 1.405, val: 1.476 | iter time: 357.34 ms (step) remaining time: 0:32:27
161
+ Validating ...
162
+ iter 1600: val loss 1.4626, val time: 21934.90 ms
163
+ Epoch 1 | iter 1632 step 51 | loss train: 1.409, val: 1.463 | iter time: 360.03 ms (step) remaining time: 0:33:33
164
+ Epoch 1 | iter 1664 step 52 | loss train: 1.472, val: 1.463 | iter time: 361.39 ms (step) remaining time: 0:33:20
165
+ Epoch 1 | iter 1696 step 53 | loss train: 1.347, val: 1.463 | iter time: 359.22 ms (step) remaining time: 0:33:07
166
+ Epoch 1 | iter 1728 step 54 | loss train: 1.342, val: 1.463 | iter time: 357.80 ms (step) remaining time: 0:32:55
167
+ Epoch 1 | iter 1760 step 55 | loss train: 1.430, val: 1.463 | iter time: 605.16 ms (step) remaining time: 0:32:43
168
+ Epoch 1 | iter 1792 step 56 | loss train: 1.435, val: 1.463 | iter time: 359.22 ms (step) remaining time: 0:32:30
169
+ Epoch 1 | iter 1824 step 57 | loss train: 1.413, val: 1.463 | iter time: 358.84 ms (step) remaining time: 0:32:18
170
+ Epoch 1 | iter 1856 step 58 | loss train: 1.386, val: 1.463 | iter time: 358.99 ms (step) remaining time: 0:32:05
171
+ Epoch 1 | iter 1888 step 59 | loss train: 1.427, val: 1.463 | iter time: 360.41 ms (step) remaining time: 0:31:53
172
+ Epoch 1 | iter 1920 step 60 | loss train: 1.400, val: 1.463 | iter time: 357.77 ms (step) remaining time: 0:31:40
173
+ Epoch 1 | iter 1952 step 61 | loss train: 1.450, val: 1.463 | iter time: 358.48 ms (step) remaining time: 0:31:28
174
+ Epoch 1 | iter 1984 step 62 | loss train: 1.381, val: 1.463 | iter time: 360.31 ms (step) remaining time: 0:31:16
175
+ Epoch 1 | iter 2016 step 63 | loss train: 1.372, val: 1.463 | iter time: 361.03 ms (step) remaining time: 0:31:04
176
+ Epoch 1 | iter 2048 step 64 | loss train: 1.455, val: 1.463 | iter time: 361.10 ms (step) remaining time: 0:30:51
177
+ Epoch 1 | iter 2080 step 65 | loss train: 1.453, val: 1.463 | iter time: 359.55 ms (step) remaining time: 0:30:39
178
+ Epoch 1 | iter 2112 step 66 | loss train: 1.407, val: 1.463 | iter time: 359.98 ms (step) remaining time: 0:30:27
179
+ Epoch 1 | iter 2144 step 67 | loss train: 1.403, val: 1.463 | iter time: 361.10 ms (step) remaining time: 0:30:15
180
+ Epoch 1 | iter 2176 step 68 | loss train: 1.389, val: 1.463 | iter time: 360.72 ms (step) remaining time: 0:30:03
181
+ Epoch 1 | iter 2208 step 69 | loss train: 1.431, val: 1.463 | iter time: 359.26 ms (step) remaining time: 0:29:51
182
+ Epoch 1 | iter 2240 step 70 | loss train: 1.421, val: 1.463 | iter time: 358.02 ms (step) remaining time: 0:29:39
183
+ Epoch 1 | iter 2272 step 71 | loss train: 1.408, val: 1.463 | iter time: 358.99 ms (step) remaining time: 0:29:27
184
+ Epoch 1 | iter 2304 step 72 | loss train: 1.349, val: 1.463 | iter time: 359.51 ms (step) remaining time: 0:29:15
185
+ Epoch 1 | iter 2336 step 73 | loss train: 1.480, val: 1.463 | iter time: 359.38 ms (step) remaining time: 0:29:03
186
+ Epoch 1 | iter 2368 step 74 | loss train: 1.400, val: 1.463 | iter time: 360.56 ms (step) remaining time: 0:28:52
187
+ Epoch 1 | iter 2400 step 75 | loss train: 1.423, val: 1.463 | iter time: 359.39 ms (step) remaining time: 0:28:40
188
+ Epoch 1 | iter 2432 step 76 | loss train: 1.468, val: 1.463 | iter time: 358.38 ms (step) remaining time: 0:28:28
189
+ Epoch 1 | iter 2464 step 77 | loss train: 1.457, val: 1.463 | iter time: 360.23 ms (step) remaining time: 0:28:16
190
+ Epoch 1 | iter 2496 step 78 | loss train: 1.464, val: 1.463 | iter time: 359.34 ms (step) remaining time: 0:28:04
191
+ Epoch 1 | iter 2528 step 79 | loss train: 1.422, val: 1.463 | iter time: 360.36 ms (step) remaining time: 0:27:53
192
+ Epoch 1 | iter 2560 step 80 | loss train: 1.396, val: 1.463 | iter time: 359.33 ms (step) remaining time: 0:27:41
193
+ Epoch 1 | iter 2592 step 81 | loss train: 1.370, val: 1.463 | iter time: 358.77 ms (step) remaining time: 0:27:29
194
+ Epoch 1 | iter 2624 step 82 | loss train: 1.462, val: 1.463 | iter time: 357.94 ms (step) remaining time: 0:27:18
195
+ Epoch 1 | iter 2656 step 83 | loss train: 1.440, val: 1.463 | iter time: 360.16 ms (step) remaining time: 0:27:06
196
+ Epoch 1 | iter 2688 step 84 | loss train: 1.407, val: 1.463 | iter time: 360.14 ms (step) remaining time: 0:26:54
197
+ Epoch 1 | iter 2720 step 85 | loss train: 1.381, val: 1.463 | iter time: 357.96 ms (step) remaining time: 0:26:43
198
+ Epoch 1 | iter 2752 step 86 | loss train: 1.398, val: 1.463 | iter time: 360.35 ms (step) remaining time: 0:26:31
199
+ Epoch 1 | iter 2784 step 87 | loss train: 1.412, val: 1.463 | iter time: 358.09 ms (step) remaining time: 0:26:20
200
+ Epoch 1 | iter 2816 step 88 | loss train: 1.441, val: 1.463 | iter time: 359.66 ms (step) remaining time: 0:26:08
201
+ Epoch 1 | iter 2848 step 89 | loss train: 1.427, val: 1.463 | iter time: 358.37 ms (step) remaining time: 0:25:56
202
+ Epoch 1 | iter 2880 step 90 | loss train: 1.425, val: 1.463 | iter time: 359.89 ms (step) remaining time: 0:25:45
203
+ Epoch 1 | iter 2912 step 91 | loss train: 1.437, val: 1.463 | iter time: 357.72 ms (step) remaining time: 0:25:33
204
+ Epoch 1 | iter 2944 step 92 | loss train: 1.421, val: 1.463 | iter time: 357.77 ms (step) remaining time: 0:25:22
205
+ Epoch 1 | iter 2976 step 93 | loss train: 1.427, val: 1.463 | iter time: 359.69 ms (step) remaining time: 0:25:10
206
+ Epoch 1 | iter 3008 step 94 | loss train: 1.365, val: 1.463 | iter time: 360.43 ms (step) remaining time: 0:24:59
207
+ Epoch 1 | iter 3040 step 95 | loss train: 1.356, val: 1.463 | iter time: 358.34 ms (step) remaining time: 0:24:47
208
+ Epoch 1 | iter 3072 step 96 | loss train: 1.374, val: 1.463 | iter time: 359.81 ms (step) remaining time: 0:24:36
209
+ Epoch 1 | iter 3104 step 97 | loss train: 1.399, val: 1.463 | iter time: 359.25 ms (step) remaining time: 0:24:25
210
+ Epoch 1 | iter 3136 step 98 | loss train: 1.420, val: 1.463 | iter time: 360.10 ms (step) remaining time: 0:24:13
211
+ Epoch 1 | iter 3168 step 99 | loss train: 1.341, val: 1.463 | iter time: 360.92 ms (step) remaining time: 0:24:02
212
+ Epoch 1 | iter 3200 step 100 | loss train: 1.384, val: 1.463 | iter time: 358.23 ms (step) remaining time: 0:23:50
213
+ Validating ...
214
+ iter 3200: val loss 1.4580, val time: 21920.12 ms
215
+ Saving checkpoint to 'out/pretrain/tinyllama/2409_lr4e-5/step-00000100/lit_model.pth'
216
+ Epoch 1 | iter 3232 step 101 | loss train: 1.406, val: 1.458 | iter time: 357.85 ms (step) remaining time: 0:24:27
217
+ Epoch 1 | iter 3264 step 102 | loss train: 1.475, val: 1.458 | iter time: 359.15 ms (step) remaining time: 0:24:15
218
+ Epoch 1 | iter 3296 step 103 | loss train: 1.458, val: 1.458 | iter time: 356.86 ms (step) remaining time: 0:24:03
219
+ Epoch 1 | iter 3328 step 104 | loss train: 1.400, val: 1.458 | iter time: 359.59 ms (step) remaining time: 0:23:51
220
+ Epoch 1 | iter 3360 step 105 | loss train: 1.399, val: 1.458 | iter time: 360.45 ms (step) remaining time: 0:23:39
221
+ Epoch 1 | iter 3392 step 106 | loss train: 1.417, val: 1.458 | iter time: 359.52 ms (step) remaining time: 0:23:26
222
+ Epoch 1 | iter 3424 step 107 | loss train: 1.451, val: 1.458 | iter time: 359.45 ms (step) remaining time: 0:23:15
223
+ Epoch 1 | iter 3456 step 108 | loss train: 1.403, val: 1.458 | iter time: 357.44 ms (step) remaining time: 0:23:02
224
+ Epoch 1 | iter 3488 step 109 | loss train: 1.366, val: 1.458 | iter time: 358.17 ms (step) remaining time: 0:22:50
225
+ Epoch 1 | iter 3520 step 110 | loss train: 1.384, val: 1.458 | iter time: 359.12 ms (step) remaining time: 0:22:38
226
+ Epoch 1 | iter 3552 step 111 | loss train: 1.378, val: 1.458 | iter time: 360.84 ms (step) remaining time: 0:22:26
227
+ Epoch 1 | iter 3584 step 112 | loss train: 1.423, val: 1.458 | iter time: 357.46 ms (step) remaining time: 0:22:14
228
+ Epoch 1 | iter 3616 step 113 | loss train: 1.323, val: 1.458 | iter time: 596.10 ms (step) remaining time: 0:22:03
229
+ Epoch 1 | iter 3648 step 114 | loss train: 1.356, val: 1.458 | iter time: 359.73 ms (step) remaining time: 0:21:51
230
+ Epoch 1 | iter 3680 step 115 | loss train: 1.409, val: 1.458 | iter time: 361.04 ms (step) remaining time: 0:21:39
231
+ Epoch 1 | iter 3712 step 116 | loss train: 1.437, val: 1.458 | iter time: 359.73 ms (step) remaining time: 0:21:27
232
+ Epoch 1 | iter 3744 step 117 | loss train: 1.408, val: 1.458 | iter time: 359.68 ms (step) remaining time: 0:21:15
233
+ Epoch 1 | iter 3776 step 118 | loss train: 1.356, val: 1.458 | iter time: 360.61 ms (step) remaining time: 0:21:03
234
+ Epoch 1 | iter 3808 step 119 | loss train: 1.425, val: 1.458 | iter time: 359.65 ms (step) remaining time: 0:20:51
235
+ Epoch 1 | iter 3840 step 120 | loss train: 1.417, val: 1.458 | iter time: 359.37 ms (step) remaining time: 0:20:39
236
+ Epoch 1 | iter 3872 step 121 | loss train: 1.429, val: 1.458 | iter time: 359.00 ms (step) remaining time: 0:20:28
237
+ Epoch 1 | iter 3904 step 122 | loss train: 1.404, val: 1.458 | iter time: 359.91 ms (step) remaining time: 0:20:16
238
+ Epoch 1 | iter 3936 step 123 | loss train: 1.426, val: 1.458 | iter time: 361.13 ms (step) remaining time: 0:20:04
239
+ Epoch 1 | iter 3968 step 124 | loss train: 1.385, val: 1.458 | iter time: 359.71 ms (step) remaining time: 0:19:52
240
+ Epoch 1 | iter 4000 step 125 | loss train: 1.389, val: 1.458 | iter time: 358.60 ms (step) remaining time: 0:19:41
241
+ Epoch 1 | iter 4032 step 126 | loss train: 1.388, val: 1.458 | iter time: 359.52 ms (step) remaining time: 0:19:29
242
+ Epoch 1 | iter 4064 step 127 | loss train: 1.445, val: 1.458 | iter time: 359.61 ms (step) remaining time: 0:19:17
243
+ Epoch 1 | iter 4096 step 128 | loss train: 1.386, val: 1.458 | iter time: 358.50 ms (step) remaining time: 0:19:06
244
+ Epoch 1 | iter 4128 step 129 | loss train: 1.404, val: 1.458 | iter time: 358.11 ms (step) remaining time: 0:18:54
245
+ Epoch 1 | iter 4160 step 130 | loss train: 1.408, val: 1.458 | iter time: 357.88 ms (step) remaining time: 0:18:42
246
+ Epoch 1 | iter 4192 step 131 | loss train: 1.440, val: 1.458 | iter time: 360.37 ms (step) remaining time: 0:18:30
247
+ Epoch 1 | iter 4224 step 132 | loss train: 1.380, val: 1.458 | iter time: 358.55 ms (step) remaining time: 0:18:19
248
+ Epoch 1 | iter 4256 step 133 | loss train: 1.398, val: 1.458 | iter time: 359.42 ms (step) remaining time: 0:18:07
249
+ Epoch 1 | iter 4288 step 134 | loss train: 1.378, val: 1.458 | iter time: 360.42 ms (step) remaining time: 0:17:55
250
+ Epoch 1 | iter 4320 step 135 | loss train: 1.364, val: 1.458 | iter time: 358.11 ms (step) remaining time: 0:17:44
251
+ Epoch 1 | iter 4352 step 136 | loss train: 1.416, val: 1.458 | iter time: 360.24 ms (step) remaining time: 0:17:32
252
+ Epoch 1 | iter 4384 step 137 | loss train: 1.386, val: 1.458 | iter time: 359.45 ms (step) remaining time: 0:17:21
253
+ Epoch 1 | iter 4416 step 138 | loss train: 1.408, val: 1.458 | iter time: 361.19 ms (step) remaining time: 0:17:09
254
+ Epoch 1 | iter 4448 step 139 | loss train: 1.459, val: 1.458 | iter time: 360.82 ms (step) remaining time: 0:16:57
255
+ Epoch 1 | iter 4480 step 140 | loss train: 1.402, val: 1.458 | iter time: 358.89 ms (step) remaining time: 0:16:46
256
+ Epoch 1 | iter 4512 step 141 | loss train: 1.382, val: 1.458 | iter time: 360.97 ms (step) remaining time: 0:16:34
257
+ Epoch 1 | iter 4544 step 142 | loss train: 1.371, val: 1.458 | iter time: 358.42 ms (step) remaining time: 0:16:23
258
+ Epoch 1 | iter 4576 step 143 | loss train: 1.418, val: 1.458 | iter time: 359.63 ms (step) remaining time: 0:16:11
259
+ Epoch 1 | iter 4608 step 144 | loss train: 1.351, val: 1.458 | iter time: 359.74 ms (step) remaining time: 0:15:59
260
+ Epoch 1 | iter 4640 step 145 | loss train: 1.333, val: 1.458 | iter time: 360.27 ms (step) remaining time: 0:15:48
261
+ Epoch 1 | iter 4672 step 146 | loss train: 1.401, val: 1.458 | iter time: 358.44 ms (step) remaining time: 0:15:36
262
+ Epoch 1 | iter 4704 step 147 | loss train: 1.430, val: 1.458 | iter time: 359.89 ms (step) remaining time: 0:15:25
263
+ Epoch 1 | iter 4736 step 148 | loss train: 1.327, val: 1.458 | iter time: 359.88 ms (step) remaining time: 0:15:13
264
+ Epoch 1 | iter 4768 step 149 | loss train: 1.426, val: 1.458 | iter time: 360.46 ms (step) remaining time: 0:15:02
265
+ Epoch 1 | iter 4800 step 150 | loss train: 1.401, val: 1.458 | iter time: 360.05 ms (step) remaining time: 0:14:50
266
+ Validating ...
267
+ iter 4800: val loss 1.4516, val time: 21924.21 ms
268
+ Epoch 1 | iter 4832 step 151 | loss train: 1.370, val: 1.452 | iter time: 360.84 ms (step) remaining time: 0:14:50
269
+ Epoch 1 | iter 4864 step 152 | loss train: 1.437, val: 1.452 | iter time: 360.27 ms (step) remaining time: 0:14:38
270
+ Epoch 1 | iter 4896 step 153 | loss train: 1.412, val: 1.452 | iter time: 358.42 ms (step) remaining time: 0:14:27
271
+ Epoch 1 | iter 4928 step 154 | loss train: 1.325, val: 1.452 | iter time: 361.22 ms (step) remaining time: 0:14:15
272
+ Epoch 1 | iter 4960 step 155 | loss train: 1.377, val: 1.452 | iter time: 358.15 ms (step) remaining time: 0:14:03
273
+ Epoch 1 | iter 4992 step 156 | loss train: 1.376, val: 1.452 | iter time: 359.50 ms (step) remaining time: 0:13:52
274
+ Epoch 1 | iter 5024 step 157 | loss train: 1.419, val: 1.452 | iter time: 360.80 ms (step) remaining time: 0:13:40
275
+ Epoch 1 | iter 5056 step 158 | loss train: 1.383, val: 1.452 | iter time: 358.62 ms (step) remaining time: 0:13:28
276
+ Epoch 1 | iter 5088 step 159 | loss train: 1.442, val: 1.452 | iter time: 360.67 ms (step) remaining time: 0:13:17
277
+ Epoch 1 | iter 5120 step 160 | loss train: 1.416, val: 1.452 | iter time: 361.06 ms (step) remaining time: 0:13:05
278
+ Epoch 1 | iter 5152 step 161 | loss train: 1.381, val: 1.452 | iter time: 359.42 ms (step) remaining time: 0:12:54
279
+ Epoch 1 | iter 5184 step 162 | loss train: 1.408, val: 1.452 | iter time: 359.96 ms (step) remaining time: 0:12:42
280
+ Epoch 1 | iter 5216 step 163 | loss train: 1.434, val: 1.452 | iter time: 359.23 ms (step) remaining time: 0:12:30
281
+ Epoch 1 | iter 5248 step 164 | loss train: 1.373, val: 1.452 | iter time: 358.71 ms (step) remaining time: 0:12:19
282
+ Epoch 1 | iter 5280 step 165 | loss train: 1.435, val: 1.452 | iter time: 361.07 ms (step) remaining time: 0:12:07
283
+ Epoch 1 | iter 5312 step 166 | loss train: 1.414, val: 1.452 | iter time: 360.17 ms (step) remaining time: 0:11:56
284
+ Epoch 1 | iter 5344 step 167 | loss train: 1.380, val: 1.452 | iter time: 360.52 ms (step) remaining time: 0:11:44
285
+ Epoch 1 | iter 5376 step 168 | loss train: 1.317, val: 1.452 | iter time: 360.58 ms (step) remaining time: 0:11:33
286
+ Epoch 1 | iter 5408 step 169 | loss train: 1.351, val: 1.452 | iter time: 360.37 ms (step) remaining time: 0:11:21
287
+ Epoch 1 | iter 5440 step 170 | loss train: 1.424, val: 1.452 | iter time: 359.67 ms (step) remaining time: 0:11:10
288
+ Epoch 1 | iter 5472 step 171 | loss train: 1.347, val: 1.452 | iter time: 358.93 ms (step) remaining time: 0:10:58
289
+ Epoch 1 | iter 5504 step 172 | loss train: 1.405, val: 1.452 | iter time: 357.71 ms (step) remaining time: 0:10:46
290
+ Epoch 1 | iter 5536 step 173 | loss train: 1.371, val: 1.452 | iter time: 361.79 ms (step) remaining time: 0:10:35
291
+ Epoch 1 | iter 5568 step 174 | loss train: 1.490, val: 1.452 | iter time: 359.08 ms (step) remaining time: 0:10:23
292
+ Epoch 1 | iter 5600 step 175 | loss train: 1.386, val: 1.452 | iter time: 359.42 ms (step) remaining time: 0:10:12
293
+ Epoch 1 | iter 5632 step 176 | loss train: 1.329, val: 1.452 | iter time: 358.66 ms (step) remaining time: 0:10:00
294
+ Epoch 1 | iter 5664 step 177 | loss train: 1.423, val: 1.452 | iter time: 360.97 ms (step) remaining time: 0:09:49
295
+ Epoch 1 | iter 5696 step 178 | loss train: 1.364, val: 1.452 | iter time: 360.18 ms (step) remaining time: 0:09:38
296
+ Epoch 1 | iter 5728 step 179 | loss train: 1.362, val: 1.452 | iter time: 360.96 ms (step) remaining time: 0:09:26
297
+ Epoch 1 | iter 5760 step 180 | loss train: 1.379, val: 1.452 | iter time: 359.09 ms (step) remaining time: 0:09:15
298
+ Epoch 1 | iter 5792 step 181 | loss train: 1.378, val: 1.452 | iter time: 359.78 ms (step) remaining time: 0:09:03
299
+ Epoch 1 | iter 5824 step 182 | loss train: 1.406, val: 1.452 | iter time: 360.20 ms (step) remaining time: 0:08:52
300
+ Epoch 1 | iter 5856 step 183 | loss train: 1.436, val: 1.452 | iter time: 359.22 ms (step) remaining time: 0:08:40
301
+ Epoch 1 | iter 5888 step 184 | loss train: 1.378, val: 1.452 | iter time: 360.23 ms (step) remaining time: 0:08:29
302
+ Epoch 1 | iter 5920 step 185 | loss train: 1.415, val: 1.452 | iter time: 360.64 ms (step) remaining time: 0:08:17
303
+ Epoch 1 | iter 5952 step 186 | loss train: 1.337, val: 1.452 | iter time: 359.25 ms (step) remaining time: 0:08:06
304
+ Epoch 1 | iter 5984 step 187 | loss train: 1.320, val: 1.452 | iter time: 360.09 ms (step) remaining time: 0:07:55
305
+ Epoch 1 | iter 6016 step 188 | loss train: 1.446, val: 1.452 | iter time: 360.62 ms (step) remaining time: 0:07:43
306
+ Epoch 1 | iter 6048 step 189 | loss train: 1.330, val: 1.452 | iter time: 360.41 ms (step) remaining time: 0:07:32
307
+ Epoch 1 | iter 6080 step 190 | loss train: 1.385, val: 1.452 | iter time: 361.33 ms (step) remaining time: 0:07:20
308
+ Epoch 1 | iter 6112 step 191 | loss train: 1.373, val: 1.452 | iter time: 360.04 ms (step) remaining time: 0:07:09
309
+ Epoch 1 | iter 6144 step 192 | loss train: 1.391, val: 1.452 | iter time: 361.01 ms (step) remaining time: 0:06:58
310
+ Epoch 1 | iter 6176 step 193 | loss train: 1.398, val: 1.452 | iter time: 360.77 ms (step) remaining time: 0:06:46
311
+ Epoch 1 | iter 6208 step 194 | loss train: 1.407, val: 1.452 | iter time: 604.28 ms (step) remaining time: 0:06:35
312
+ Epoch 1 | iter 6240 step 195 | loss train: 1.342, val: 1.452 | iter time: 359.34 ms (step) remaining time: 0:06:24
313
+ Epoch 1 | iter 6272 step 196 | loss train: 1.441, val: 1.452 | iter time: 362.32 ms (step) remaining time: 0:06:12
314
+ Epoch 1 | iter 6304 step 197 | loss train: 1.441, val: 1.452 | iter time: 359.83 ms (step) remaining time: 0:06:01
315
+ Epoch 1 | iter 6336 step 198 | loss train: 1.373, val: 1.452 | iter time: 358.74 ms (step) remaining time: 0:05:49
316
+ Epoch 1 | iter 6368 step 199 | loss train: 1.436, val: 1.452 | iter time: 359.92 ms (step) remaining time: 0:05:38
317
+ Epoch 1 | iter 6400 step 200 | loss train: 1.352, val: 1.452 | iter time: 359.42 ms (step) remaining time: 0:05:27
318
+ Validating ...
319
+ iter 6400: val loss 1.4481, val time: 21929.64 ms
320
+ Saving checkpoint to 'out/pretrain/tinyllama/2409_lr4e-5/step-00000200/lit_model.pth'
321
+ Epoch 1 | iter 6432 step 201 | loss train: 1.359, val: 1.448 | iter time: 358.05 ms (step) remaining time: 0:05:21
322
+ Epoch 1 | iter 6464 step 202 | loss train: 1.331, val: 1.448 | iter time: 357.87 ms (step) remaining time: 0:05:09
323
+ Epoch 1 | iter 6496 step 203 | loss train: 1.441, val: 1.448 | iter time: 357.22 ms (step) remaining time: 0:04:58
324
+ Epoch 1 | iter 6528 step 204 | loss train: 1.402, val: 1.448 | iter time: 359.81 ms (step) remaining time: 0:04:46
325
+ Epoch 1 | iter 6560 step 205 | loss train: 1.405, val: 1.448 | iter time: 360.59 ms (step) remaining time: 0:04:35
326
+ Epoch 1 | iter 6592 step 206 | loss train: 1.356, val: 1.448 | iter time: 358.81 ms (step) remaining time: 0:04:23
327
+ Epoch 1 | iter 6624 step 207 | loss train: 1.385, val: 1.448 | iter time: 360.74 ms (step) remaining time: 0:04:11
328
+ Epoch 1 | iter 6656 step 208 | loss train: 1.364, val: 1.448 | iter time: 360.52 ms (step) remaining time: 0:04:00
329
+ Epoch 1 | iter 6688 step 209 | loss train: 1.383, val: 1.448 | iter time: 359.06 ms (step) remaining time: 0:03:48
330
+ Epoch 1 | iter 6720 step 210 | loss train: 1.430, val: 1.448 | iter time: 359.83 ms (step) remaining time: 0:03:37
331
+ Epoch 1 | iter 6752 step 211 | loss train: 1.330, val: 1.448 | iter time: 361.80 ms (step) remaining time: 0:03:25
332
+ Epoch 1 | iter 6784 step 212 | loss train: 1.422, val: 1.448 | iter time: 359.61 ms (step) remaining time: 0:03:14
333
+ Epoch 1 | iter 6816 step 213 | loss train: 1.382, val: 1.448 | iter time: 358.98 ms (step) remaining time: 0:03:02
334
+ Epoch 1 | iter 6848 step 214 | loss train: 1.410, val: 1.448 | iter time: 359.65 ms (step) remaining time: 0:02:51
335
+ Epoch 1 | iter 6880 step 215 | loss train: 1.425, val: 1.448 | iter time: 359.73 ms (step) remaining time: 0:02:40
336
+ Epoch 1 | iter 6912 step 216 | loss train: 1.337, val: 1.448 | iter time: 360.00 ms (step) remaining time: 0:02:28
337
+ Epoch 1 | iter 6944 step 217 | loss train: 1.382, val: 1.448 | iter time: 361.35 ms (step) remaining time: 0:02:17
338
+ Epoch 1 | iter 6976 step 218 | loss train: 1.289, val: 1.448 | iter time: 360.33 ms (step) remaining time: 0:02:05
339
+ Epoch 1 | iter 7008 step 219 | loss train: 1.404, val: 1.448 | iter time: 359.64 ms (step) remaining time: 0:01:54
340
+ Epoch 1 | iter 7040 step 220 | loss train: 1.327, val: 1.448 | iter time: 458.26 ms (step) remaining time: 0:01:42
341
+ Epoch 1 | iter 7072 step 221 | loss train: 1.339, val: 1.448 | iter time: 359.29 ms (step) remaining time: 0:01:31
342
+ Epoch 1 | iter 7104 step 222 | loss train: 1.324, val: 1.448 | iter time: 360.20 ms (step) remaining time: 0:01:19
343
+ Epoch 1 | iter 7136 step 223 | loss train: 1.408, val: 1.448 | iter time: 360.63 ms (step) remaining time: 0:01:08
344
+ Epoch 1 | iter 7168 step 224 | loss train: 1.403, val: 1.448 | iter time: 357.99 ms (step) remaining time: 0:00:57
345
+ Epoch 1 | iter 7200 step 225 | loss train: 1.408, val: 1.448 | iter time: 361.14 ms (step) remaining time: 0:00:45
346
+ Epoch 1 | iter 7232 step 226 | loss train: 1.354, val: 1.448 | iter time: 359.66 ms (step) remaining time: 0:00:34
347
+ Epoch 1 | iter 7264 step 227 | loss train: 1.414, val: 1.448 | iter time: 358.26 ms (step) remaining time: 0:00:22
348
+ Epoch 1 | iter 7296 step 228 | loss train: 1.287, val: 1.448 | iter time: 360.67 ms (step) remaining time: 0:00:11
349
+ Epoch 2 | iter 7328 step 229 | loss train: 1.357, val: 1.448 | iter time: 359.97 ms (step) remaining time: 0:00:00
350
+ Validating ...
351
+ Final evaluation | val loss: 1.388 | val ppl: 4.009
352
+ Saving checkpoint to 'out/pretrain/tinyllama/2409_lr4e-5/final/lit_model.pth'
353
+ ----------------------------------------
354
+ | Performance
355
+ | - Total tokens : 240,123,904
356
+ | - Training Time : 2672.28 s
357
+ | - Tok/sec : 124.03 tok/s
358
+ | ----------------------------------------
359
+ | Memory Usage
360
+ | - Memory Used : 26.32 GB
361
+ ----------------------------------------
out/pretrain/tinyllama/teelogs/2410.txt ADDED
@@ -0,0 +1,368 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Initializing distributed: GLOBAL_RANK: 0, MEMBER: 1/4
2
+ Initializing distributed: GLOBAL_RANK: 3, MEMBER: 4/4
3
+ [rank: 3] Seed set to 42
4
+ Initializing distributed: GLOBAL_RANK: 2, MEMBER: 3/4
5
+ [rank: 2] Seed set to 42
6
+ Initializing distributed: GLOBAL_RANK: 1, MEMBER: 2/4
7
+ [rank: 1] Seed set to 42
8
+ ----------------------------------------------------------------------------------------------------
9
+ distributed_backend=nccl
10
+ All distributed processes registered. Starting with 4 processes
11
+ ----------------------------------------------------------------------------------------------------
12
+
13
+ [rank: 0] Seed set to 42
14
+ All GPUs are fully connected via NVLink.
15
+ {'data': {'batch_size': 1,
16
+ 'data_path': PosixPath('litgpt/data/arxiv'),
17
+ 'num_workers': 8,
18
+ 'seed': 42,
19
+ 'seq_length': 2048,
20
+ 'use_starcoder': True},
21
+ 'data_dir': PosixPath('litgpt/data/arxiv/2410'),
22
+ 'devices': 'auto',
23
+ 'eval': {'evaluate_example': 'first',
24
+ 'final_validation': True,
25
+ 'initial_validation': True,
26
+ 'interval': 50,
27
+ 'max_iters': 100,
28
+ 'max_new_tokens': None},
29
+ 'initial_checkpoint_dir': PosixPath('out/pretrain/2409/final'),
30
+ 'log': {'group': None, 'project': None, 'run': None},
31
+ 'logger_name': 'tensorboard',
32
+ 'model_config': {'attention_logit_softcapping': None,
33
+ 'attention_scores_scalar': None,
34
+ 'attn_bias': False,
35
+ 'bias': False,
36
+ 'block_size': 2048,
37
+ 'final_logit_softcapping': None,
38
+ 'gelu_approximate': 'none',
39
+ 'head_size': 64,
40
+ 'hf_config': {'name': 'TinyLlama-1.1B-intermediate-step-1431k-3T',
41
+ 'org': 'TinyLlama'},
42
+ 'intermediate_size': 5632,
43
+ 'lm_head_bias': False,
44
+ 'mlp_class_name': 'LLaMAMLP',
45
+ 'moe_intermediate_size': None,
46
+ 'n_embd': 2048,
47
+ 'n_expert': 0,
48
+ 'n_expert_per_token': 0,
49
+ 'n_head': 32,
50
+ 'n_layer': 22,
51
+ 'n_query_groups': 4,
52
+ 'name': 'tiny-llama-1.1b',
53
+ 'norm_1': True,
54
+ 'norm_2': True,
55
+ 'norm_class_name': 'RMSNorm',
56
+ 'norm_eps': 1e-05,
57
+ 'norm_qk': False,
58
+ 'norm_qk_type': 'default',
59
+ 'padded_vocab_size': 32000,
60
+ 'padding_multiple': 64,
61
+ 'parallel_residual': False,
62
+ 'post_attention_norm': False,
63
+ 'post_mlp_norm': False,
64
+ 'rope_adjustments': None,
65
+ 'rope_base': 10000,
66
+ 'rope_condense_ratio': 1,
67
+ 'rope_indices': None,
68
+ 'rope_local_base_freq': None,
69
+ 'rotary_percentage': 1.0,
70
+ 'scale_embeddings': False,
71
+ 'shared_attention_norm': False,
72
+ 'sliding_window_indices': None,
73
+ 'sliding_window_size': None,
74
+ 'vocab_size': 32000},
75
+ 'model_name': 'tiny-llama-1.1b',
76
+ 'num_nodes': 1,
77
+ 'optimizer': "{'class_path': 'torch.optim.AdamW', 'init_args': {'lr': 0.0004, "
78
+ "'weight_decay': 0.1, 'betas': [0.9, 0.95]}}",
79
+ 'out_dir': PosixPath('out/pretrain/2410'),
80
+ 'precision': 'bf16-mixed',
81
+ 'resume': False,
82
+ 'seed': 42,
83
+ 'tokenizer_dir': PosixPath('checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T'),
84
+ 'train': {'epochs': None,
85
+ 'global_batch_size': 512,
86
+ 'log_interval': 1,
87
+ 'lr_warmup_fraction': None,
88
+ 'lr_warmup_steps': 20,
89
+ 'max_norm': 1.0,
90
+ 'max_seq_length': 2048,
91
+ 'max_steps': None,
92
+ 'max_tokens': 255852544,
93
+ 'micro_batch_size': 4,
94
+ 'min_lr': 4e-05,
95
+ 'save_interval': 100,
96
+ 'tie_embeddings': None}}
97
+ Time to instantiate model: 0.02 seconds.
98
+ Total parameters: 1,100,048,384
99
+ [fix] out/pretrain/2409/final/lit_model.pth 是整包 state,提取 model 权重并原子覆盖...
100
+ [fix] 已覆盖为纯权重: out/pretrain/2409/final/lit_model.pth
101
+ Validating ...
102
+ Measured TFLOPs: 239.66
103
+ Epoch 1 | iter 32 step 1 | loss train: 1.466, val: 1.433 | iter time: 538.97 ms (step) remaining time: 0:48:06
104
+ Epoch 1 | iter 64 step 2 | loss train: 1.464, val: 1.433 | iter time: 357.93 ms (step) remaining time: 0:45:38
105
+ Epoch 1 | iter 96 step 3 | loss train: 1.443, val: 1.433 | iter time: 356.34 ms (step) remaining time: 0:44:42
106
+ Epoch 1 | iter 128 step 4 | loss train: 1.445, val: 1.433 | iter time: 360.75 ms (step) remaining time: 0:44:13
107
+ Epoch 1 | iter 160 step 5 | loss train: 1.402, val: 1.433 | iter time: 359.07 ms (step) remaining time: 0:43:52
108
+ Epoch 1 | iter 192 step 6 | loss train: 1.479, val: 1.433 | iter time: 359.01 ms (step) remaining time: 0:43:34
109
+ Epoch 1 | iter 224 step 7 | loss train: 1.492, val: 1.433 | iter time: 358.53 ms (step) remaining time: 0:43:18
110
+ Epoch 1 | iter 256 step 8 | loss train: 1.459, val: 1.433 | iter time: 359.98 ms (step) remaining time: 0:43:03
111
+ Epoch 1 | iter 288 step 9 | loss train: 1.464, val: 1.433 | iter time: 358.87 ms (step) remaining time: 0:42:49
112
+ Epoch 1 | iter 320 step 10 | loss train: 1.448, val: 1.433 | iter time: 358.91 ms (step) remaining time: 0:42:36
113
+ Epoch 1 | iter 352 step 11 | loss train: 1.456, val: 1.433 | iter time: 357.55 ms (step) remaining time: 0:42:24
114
+ Epoch 1 | iter 384 step 12 | loss train: 1.470, val: 1.433 | iter time: 359.43 ms (step) remaining time: 0:42:12
115
+ Epoch 1 | iter 416 step 13 | loss train: 1.490, val: 1.433 | iter time: 360.05 ms (step) remaining time: 0:41:59
116
+ Epoch 1 | iter 448 step 14 | loss train: 1.440, val: 1.433 | iter time: 360.50 ms (step) remaining time: 0:41:47
117
+ Epoch 1 | iter 480 step 15 | loss train: 1.453, val: 1.433 | iter time: 360.78 ms (step) remaining time: 0:41:36
118
+ Epoch 1 | iter 512 step 16 | loss train: 1.419, val: 1.433 | iter time: 360.18 ms (step) remaining time: 0:41:24
119
+ Epoch 1 | iter 544 step 17 | loss train: 1.485, val: 1.433 | iter time: 360.97 ms (step) remaining time: 0:41:13
120
+ Epoch 1 | iter 576 step 18 | loss train: 1.457, val: 1.433 | iter time: 359.06 ms (step) remaining time: 0:41:02
121
+ Epoch 1 | iter 608 step 19 | loss train: 1.491, val: 1.433 | iter time: 358.67 ms (step) remaining time: 0:40:50
122
+ Epoch 1 | iter 640 step 20 | loss train: 1.485, val: 1.433 | iter time: 359.53 ms (step) remaining time: 0:40:39
123
+ Epoch 1 | iter 672 step 21 | loss train: 1.460, val: 1.433 | iter time: 360.01 ms (step) remaining time: 0:40:28
124
+ Epoch 1 | iter 704 step 22 | loss train: 1.486, val: 1.433 | iter time: 358.20 ms (step) remaining time: 0:40:17
125
+ Epoch 1 | iter 736 step 23 | loss train: 1.431, val: 1.433 | iter time: 358.85 ms (step) remaining time: 0:40:06
126
+ Epoch 1 | iter 768 step 24 | loss train: 1.432, val: 1.433 | iter time: 358.47 ms (step) remaining time: 0:39:55
127
+ Epoch 1 | iter 800 step 25 | loss train: 1.430, val: 1.433 | iter time: 359.47 ms (step) remaining time: 0:39:44
128
+ Epoch 1 | iter 832 step 26 | loss train: 1.482, val: 1.433 | iter time: 360.26 ms (step) remaining time: 0:39:32
129
+ Epoch 1 | iter 864 step 27 | loss train: 1.422, val: 1.433 | iter time: 358.35 ms (step) remaining time: 0:39:21
130
+ Epoch 1 | iter 896 step 28 | loss train: 1.398, val: 1.433 | iter time: 359.50 ms (step) remaining time: 0:39:10
131
+ Epoch 1 | iter 928 step 29 | loss train: 1.424, val: 1.433 | iter time: 360.78 ms (step) remaining time: 0:38:59
132
+ Epoch 1 | iter 960 step 30 | loss train: 1.510, val: 1.433 | iter time: 360.66 ms (step) remaining time: 0:38:49
133
+ Epoch 1 | iter 992 step 31 | loss train: 1.376, val: 1.433 | iter time: 358.93 ms (step) remaining time: 0:38:38
134
+ Epoch 1 | iter 1024 step 32 | loss train: 1.508, val: 1.433 | iter time: 359.29 ms (step) remaining time: 0:38:28
135
+ Epoch 1 | iter 1056 step 33 | loss train: 1.424, val: 1.433 | iter time: 360.59 ms (step) remaining time: 0:38:17
136
+ Epoch 1 | iter 1088 step 34 | loss train: 1.426, val: 1.433 | iter time: 361.23 ms (step) remaining time: 0:38:06
137
+ Epoch 1 | iter 1120 step 35 | loss train: 1.462, val: 1.433 | iter time: 361.19 ms (step) remaining time: 0:37:55
138
+ Epoch 1 | iter 1152 step 36 | loss train: 1.465, val: 1.433 | iter time: 360.33 ms (step) remaining time: 0:37:44
139
+ Epoch 1 | iter 1184 step 37 | loss train: 1.506, val: 1.433 | iter time: 360.63 ms (step) remaining time: 0:37:33
140
+ Epoch 1 | iter 1216 step 38 | loss train: 1.472, val: 1.433 | iter time: 359.68 ms (step) remaining time: 0:37:22
141
+ Epoch 1 | iter 1248 step 39 | loss train: 1.373, val: 1.433 | iter time: 574.68 ms (step) remaining time: 0:37:12
142
+ Epoch 1 | iter 1280 step 40 | loss train: 1.489, val: 1.433 | iter time: 361.46 ms (step) remaining time: 0:37:01
143
+ Epoch 1 | iter 1312 step 41 | loss train: 1.445, val: 1.433 | iter time: 360.36 ms (step) remaining time: 0:36:50
144
+ Epoch 1 | iter 1344 step 42 | loss train: 1.536, val: 1.433 | iter time: 360.28 ms (step) remaining time: 0:36:39
145
+ Epoch 1 | iter 1376 step 43 | loss train: 1.466, val: 1.433 | iter time: 360.25 ms (step) remaining time: 0:36:28
146
+ Epoch 1 | iter 1408 step 44 | loss train: 1.475, val: 1.433 | iter time: 360.48 ms (step) remaining time: 0:36:17
147
+ Epoch 1 | iter 1440 step 45 | loss train: 1.467, val: 1.433 | iter time: 360.39 ms (step) remaining time: 0:36:06
148
+ Epoch 1 | iter 1472 step 46 | loss train: 1.461, val: 1.433 | iter time: 360.02 ms (step) remaining time: 0:35:55
149
+ Epoch 1 | iter 1504 step 47 | loss train: 1.331, val: 1.433 | iter time: 360.49 ms (step) remaining time: 0:35:44
150
+ Epoch 1 | iter 1536 step 48 | loss train: 1.468, val: 1.433 | iter time: 358.78 ms (step) remaining time: 0:35:33
151
+ Epoch 1 | iter 1568 step 49 | loss train: 1.398, val: 1.433 | iter time: 358.70 ms (step) remaining time: 0:35:22
152
+ Epoch 1 | iter 1600 step 50 | loss train: 1.427, val: 1.433 | iter time: 359.64 ms (step) remaining time: 0:35:11
153
+ Validating ...
154
+ iter 1600: val loss 1.4872, val time: 9390.22 ms
155
+ Epoch 1 | iter 1632 step 51 | loss train: 1.448, val: 1.487 | iter time: 361.16 ms (step) remaining time: 0:35:36
156
+ Epoch 1 | iter 1664 step 52 | loss train: 1.374, val: 1.487 | iter time: 361.60 ms (step) remaining time: 0:35:24
157
+ Epoch 1 | iter 1696 step 53 | loss train: 1.434, val: 1.487 | iter time: 358.98 ms (step) remaining time: 0:35:12
158
+ Epoch 1 | iter 1728 step 54 | loss train: 1.340, val: 1.487 | iter time: 359.62 ms (step) remaining time: 0:35:00
159
+ Epoch 1 | iter 1760 step 55 | loss train: 1.432, val: 1.487 | iter time: 359.79 ms (step) remaining time: 0:34:49
160
+ Epoch 1 | iter 1792 step 56 | loss train: 1.433, val: 1.487 | iter time: 358.30 ms (step) remaining time: 0:34:37
161
+ Epoch 1 | iter 1824 step 57 | loss train: 1.463, val: 1.487 | iter time: 357.91 ms (step) remaining time: 0:34:25
162
+ Epoch 1 | iter 1856 step 58 | loss train: 1.435, val: 1.487 | iter time: 361.36 ms (step) remaining time: 0:34:14
163
+ Epoch 1 | iter 1888 step 59 | loss train: 1.523, val: 1.487 | iter time: 360.72 ms (step) remaining time: 0:34:02
164
+ Epoch 1 | iter 1920 step 60 | loss train: 1.391, val: 1.487 | iter time: 358.96 ms (step) remaining time: 0:33:51
165
+ Epoch 1 | iter 1952 step 61 | loss train: 1.361, val: 1.487 | iter time: 358.85 ms (step) remaining time: 0:33:39
166
+ Epoch 1 | iter 1984 step 62 | loss train: 1.409, val: 1.487 | iter time: 360.52 ms (step) remaining time: 0:33:27
167
+ Epoch 1 | iter 2016 step 63 | loss train: 1.473, val: 1.487 | iter time: 359.89 ms (step) remaining time: 0:33:16
168
+ Epoch 1 | iter 2048 step 64 | loss train: 1.455, val: 1.487 | iter time: 360.09 ms (step) remaining time: 0:33:04
169
+ Epoch 1 | iter 2080 step 65 | loss train: 1.425, val: 1.487 | iter time: 360.64 ms (step) remaining time: 0:32:53
170
+ Epoch 1 | iter 2112 step 66 | loss train: 1.549, val: 1.487 | iter time: 361.60 ms (step) remaining time: 0:32:42
171
+ Epoch 1 | iter 2144 step 67 | loss train: 1.412, val: 1.487 | iter time: 361.50 ms (step) remaining time: 0:32:30
172
+ Epoch 1 | iter 2176 step 68 | loss train: 1.370, val: 1.487 | iter time: 358.55 ms (step) remaining time: 0:32:19
173
+ Epoch 1 | iter 2208 step 69 | loss train: 1.455, val: 1.487 | iter time: 360.57 ms (step) remaining time: 0:32:07
174
+ Epoch 1 | iter 2240 step 70 | loss train: 1.382, val: 1.487 | iter time: 360.11 ms (step) remaining time: 0:31:56
175
+ Epoch 1 | iter 2272 step 71 | loss train: 1.461, val: 1.487 | iter time: 359.48 ms (step) remaining time: 0:31:45
176
+ Epoch 1 | iter 2304 step 72 | loss train: 1.423, val: 1.487 | iter time: 359.50 ms (step) remaining time: 0:31:33
177
+ Epoch 1 | iter 2336 step 73 | loss train: 1.441, val: 1.487 | iter time: 358.78 ms (step) remaining time: 0:31:22
178
+ Epoch 1 | iter 2368 step 74 | loss train: 1.479, val: 1.487 | iter time: 359.35 ms (step) remaining time: 0:31:11
179
+ Epoch 1 | iter 2400 step 75 | loss train: 1.377, val: 1.487 | iter time: 360.39 ms (step) remaining time: 0:30:59
180
+ Epoch 1 | iter 2432 step 76 | loss train: 1.402, val: 1.487 | iter time: 360.92 ms (step) remaining time: 0:30:48
181
+ Epoch 1 | iter 2464 step 77 | loss train: 1.344, val: 1.487 | iter time: 360.01 ms (step) remaining time: 0:30:37
182
+ Epoch 1 | iter 2496 step 78 | loss train: 1.350, val: 1.487 | iter time: 359.29 ms (step) remaining time: 0:30:25
183
+ Epoch 1 | iter 2528 step 79 | loss train: 1.396, val: 1.487 | iter time: 359.02 ms (step) remaining time: 0:30:14
184
+ Epoch 1 | iter 2560 step 80 | loss train: 1.420, val: 1.487 | iter time: 358.99 ms (step) remaining time: 0:30:03
185
+ Epoch 1 | iter 2592 step 81 | loss train: 1.383, val: 1.487 | iter time: 360.35 ms (step) remaining time: 0:29:52
186
+ Epoch 1 | iter 2624 step 82 | loss train: 1.333, val: 1.487 | iter time: 361.65 ms (step) remaining time: 0:29:40
187
+ Epoch 1 | iter 2656 step 83 | loss train: 1.346, val: 1.487 | iter time: 361.18 ms (step) remaining time: 0:29:29
188
+ Epoch 1 | iter 2688 step 84 | loss train: 1.514, val: 1.487 | iter time: 360.03 ms (step) remaining time: 0:29:18
189
+ Epoch 1 | iter 2720 step 85 | loss train: 1.409, val: 1.487 | iter time: 360.69 ms (step) remaining time: 0:29:07
190
+ Epoch 1 | iter 2752 step 86 | loss train: 1.385, val: 1.487 | iter time: 362.28 ms (step) remaining time: 0:28:56
191
+ Epoch 1 | iter 2784 step 87 | loss train: 1.415, val: 1.487 | iter time: 359.15 ms (step) remaining time: 0:28:44
192
+ Epoch 1 | iter 2816 step 88 | loss train: 1.414, val: 1.487 | iter time: 358.59 ms (step) remaining time: 0:28:33
193
+ Epoch 1 | iter 2848 step 89 | loss train: 1.431, val: 1.487 | iter time: 361.58 ms (step) remaining time: 0:28:22
194
+ Epoch 1 | iter 2880 step 90 | loss train: 1.403, val: 1.487 | iter time: 360.82 ms (step) remaining time: 0:28:11
195
+ Epoch 1 | iter 2912 step 91 | loss train: 1.380, val: 1.487 | iter time: 361.98 ms (step) remaining time: 0:28:00
196
+ Epoch 1 | iter 2944 step 92 | loss train: 1.343, val: 1.487 | iter time: 359.92 ms (step) remaining time: 0:27:49
197
+ Epoch 1 | iter 2976 step 93 | loss train: 1.462, val: 1.487 | iter time: 362.02 ms (step) remaining time: 0:27:38
198
+ Epoch 1 | iter 3008 step 94 | loss train: 1.381, val: 1.487 | iter time: 359.42 ms (step) remaining time: 0:27:27
199
+ Epoch 1 | iter 3040 step 95 | loss train: 1.448, val: 1.487 | iter time: 359.42 ms (step) remaining time: 0:27:16
200
+ Epoch 1 | iter 3072 step 96 | loss train: 1.388, val: 1.487 | iter time: 361.60 ms (step) remaining time: 0:27:04
201
+ Epoch 1 | iter 3104 step 97 | loss train: 1.327, val: 1.487 | iter time: 359.88 ms (step) remaining time: 0:26:53
202
+ Epoch 1 | iter 3136 step 98 | loss train: 1.466, val: 1.487 | iter time: 361.19 ms (step) remaining time: 0:26:42
203
+ Epoch 1 | iter 3168 step 99 | loss train: 1.329, val: 1.487 | iter time: 360.18 ms (step) remaining time: 0:26:31
204
+ Epoch 1 | iter 3200 step 100 | loss train: 1.350, val: 1.487 | iter time: 359.05 ms (step) remaining time: 0:26:20
205
+ Validating ...
206
+ iter 3200: val loss 1.4740, val time: 9384.93 ms
207
+ Saving checkpoint to 'out/pretrain/2410/step-00000100/lit_model.pth'
208
+ Epoch 1 | iter 3232 step 101 | loss train: 1.296, val: 1.474 | iter time: 357.38 ms (step) remaining time: 0:26:46
209
+ Epoch 1 | iter 3264 step 102 | loss train: 1.428, val: 1.474 | iter time: 355.69 ms (step) remaining time: 0:26:34
210
+ Epoch 1 | iter 3296 step 103 | loss train: 1.448, val: 1.474 | iter time: 359.48 ms (step) remaining time: 0:26:22
211
+ Epoch 1 | iter 3328 step 104 | loss train: 1.404, val: 1.474 | iter time: 359.12 ms (step) remaining time: 0:26:11
212
+ Epoch 1 | iter 3360 step 105 | loss train: 1.388, val: 1.474 | iter time: 358.96 ms (step) remaining time: 0:25:59
213
+ Epoch 1 | iter 3392 step 106 | loss train: 1.379, val: 1.474 | iter time: 359.76 ms (step) remaining time: 0:25:47
214
+ Epoch 1 | iter 3424 step 107 | loss train: 1.381, val: 1.474 | iter time: 358.67 ms (step) remaining time: 0:25:36
215
+ Epoch 1 | iter 3456 step 108 | loss train: 1.447, val: 1.474 | iter time: 361.40 ms (step) remaining time: 0:25:24
216
+ Epoch 1 | iter 3488 step 109 | loss train: 1.409, val: 1.474 | iter time: 361.35 ms (step) remaining time: 0:25:12
217
+ Epoch 1 | iter 3520 step 110 | loss train: 1.466, val: 1.474 | iter time: 359.23 ms (step) remaining time: 0:25:01
218
+ Epoch 1 | iter 3552 step 111 | loss train: 1.318, val: 1.474 | iter time: 360.34 ms (step) remaining time: 0:24:49
219
+ Epoch 1 | iter 3584 step 112 | loss train: 1.326, val: 1.474 | iter time: 360.75 ms (step) remaining time: 0:24:38
220
+ Epoch 1 | iter 3616 step 113 | loss train: 1.371, val: 1.474 | iter time: 360.46 ms (step) remaining time: 0:24:26
221
+ Epoch 1 | iter 3648 step 114 | loss train: 1.399, val: 1.474 | iter time: 360.60 ms (step) remaining time: 0:24:14
222
+ Epoch 1 | iter 3680 step 115 | loss train: 1.436, val: 1.474 | iter time: 359.09 ms (step) remaining time: 0:24:03
223
+ Epoch 1 | iter 3712 step 116 | loss train: 1.362, val: 1.474 | iter time: 360.72 ms (step) remaining time: 0:23:51
224
+ Epoch 1 | iter 3744 step 117 | loss train: 1.312, val: 1.474 | iter time: 362.13 ms (step) remaining time: 0:23:40
225
+ Epoch 1 | iter 3776 step 118 | loss train: 1.358, val: 1.474 | iter time: 360.14 ms (step) remaining time: 0:23:28
226
+ Epoch 1 | iter 3808 step 119 | loss train: 1.335, val: 1.474 | iter time: 359.47 ms (step) remaining time: 0:23:17
227
+ Epoch 1 | iter 3840 step 120 | loss train: 1.353, val: 1.474 | iter time: 361.59 ms (step) remaining time: 0:23:05
228
+ Epoch 1 | iter 3872 step 121 | loss train: 1.338, val: 1.474 | iter time: 358.56 ms (step) remaining time: 0:22:54
229
+ Epoch 1 | iter 3904 step 122 | loss train: 1.269, val: 1.474 | iter time: 357.74 ms (step) remaining time: 0:22:42
230
+ Epoch 1 | iter 3936 step 123 | loss train: 1.350, val: 1.474 | iter time: 359.74 ms (step) remaining time: 0:22:31
231
+ Epoch 1 | iter 3968 step 124 | loss train: 1.408, val: 1.474 | iter time: 363.03 ms (step) remaining time: 0:22:19
232
+ Epoch 1 | iter 4000 step 125 | loss train: 1.413, val: 1.474 | iter time: 361.10 ms (step) remaining time: 0:22:08
233
+ Epoch 1 | iter 4032 step 126 | loss train: 1.361, val: 1.474 | iter time: 361.09 ms (step) remaining time: 0:21:56
234
+ Epoch 1 | iter 4064 step 127 | loss train: 1.339, val: 1.474 | iter time: 359.38 ms (step) remaining time: 0:21:45
235
+ Epoch 1 | iter 4096 step 128 | loss train: 1.353, val: 1.474 | iter time: 359.48 ms (step) remaining time: 0:21:34
236
+ Epoch 1 | iter 4128 step 129 | loss train: 1.406, val: 1.474 | iter time: 360.43 ms (step) remaining time: 0:21:22
237
+ Epoch 1 | iter 4160 step 130 | loss train: 1.392, val: 1.474 | iter time: 361.54 ms (step) remaining time: 0:21:11
238
+ Epoch 1 | iter 4192 step 131 | loss train: 1.394, val: 1.474 | iter time: 361.00 ms (step) remaining time: 0:20:59
239
+ Epoch 1 | iter 4224 step 132 | loss train: 1.390, val: 1.474 | iter time: 360.47 ms (step) remaining time: 0:20:48
240
+ Epoch 1 | iter 4256 step 133 | loss train: 1.364, val: 1.474 | iter time: 361.32 ms (step) remaining time: 0:20:37
241
+ Epoch 1 | iter 4288 step 134 | loss train: 1.315, val: 1.474 | iter time: 360.98 ms (step) remaining time: 0:20:25
242
+ Epoch 1 | iter 4320 step 135 | loss train: 1.338, val: 1.474 | iter time: 360.69 ms (step) remaining time: 0:20:14
243
+ Epoch 1 | iter 4352 step 136 | loss train: 1.309, val: 1.474 | iter time: 361.63 ms (step) remaining time: 0:20:03
244
+ Epoch 1 | iter 4384 step 137 | loss train: 1.281, val: 1.474 | iter time: 360.86 ms (step) remaining time: 0:19:51
245
+ Epoch 1 | iter 4416 step 138 | loss train: 1.396, val: 1.474 | iter time: 359.86 ms (step) remaining time: 0:19:40
246
+ Epoch 1 | iter 4448 step 139 | loss train: 1.356, val: 1.474 | iter time: 359.44 ms (step) remaining time: 0:19:29
247
+ Epoch 1 | iter 4480 step 140 | loss train: 1.416, val: 1.474 | iter time: 360.20 ms (step) remaining time: 0:19:17
248
+ Epoch 1 | iter 4512 step 141 | loss train: 1.449, val: 1.474 | iter time: 362.18 ms (step) remaining time: 0:19:06
249
+ Epoch 1 | iter 4544 step 142 | loss train: 1.456, val: 1.474 | iter time: 358.76 ms (step) remaining time: 0:18:55
250
+ Epoch 1 | iter 4576 step 143 | loss train: 1.363, val: 1.474 | iter time: 360.93 ms (step) remaining time: 0:18:44
251
+ Epoch 1 | iter 4608 step 144 | loss train: 1.339, val: 1.474 | iter time: 359.04 ms (step) remaining time: 0:18:32
252
+ Epoch 1 | iter 4640 step 145 | loss train: 1.378, val: 1.474 | iter time: 361.47 ms (step) remaining time: 0:18:21
253
+ Epoch 1 | iter 4672 step 146 | loss train: 1.417, val: 1.474 | iter time: 361.31 ms (step) remaining time: 0:18:10
254
+ Epoch 1 | iter 4704 step 147 | loss train: 1.392, val: 1.474 | iter time: 359.46 ms (step) remaining time: 0:17:58
255
+ Epoch 1 | iter 4736 step 148 | loss train: 1.326, val: 1.474 | iter time: 359.44 ms (step) remaining time: 0:17:47
256
+ Epoch 1 | iter 4768 step 149 | loss train: 1.336, val: 1.474 | iter time: 360.85 ms (step) remaining time: 0:17:36
257
+ Epoch 1 | iter 4800 step 150 | loss train: 1.320, val: 1.474 | iter time: 361.59 ms (step) remaining time: 0:17:25
258
+ Validating ...
259
+ iter 4800: val loss 1.4520, val time: 9382.39 ms
260
+ Epoch 1 | iter 4832 step 151 | loss train: 1.357, val: 1.452 | iter time: 361.72 ms (step) remaining time: 0:17:19
261
+ Epoch 1 | iter 4864 step 152 | loss train: 1.390, val: 1.452 | iter time: 361.82 ms (step) remaining time: 0:17:08
262
+ Epoch 1 | iter 4896 step 153 | loss train: 1.325, val: 1.452 | iter time: 359.57 ms (step) remaining time: 0:16:56
263
+ Epoch 1 | iter 4928 step 154 | loss train: 1.330, val: 1.452 | iter time: 358.68 ms (step) remaining time: 0:16:45
264
+ Epoch 1 | iter 4960 step 155 | loss train: 1.378, val: 1.452 | iter time: 359.74 ms (step) remaining time: 0:16:34
265
+ Epoch 1 | iter 4992 step 156 | loss train: 1.380, val: 1.452 | iter time: 359.44 ms (step) remaining time: 0:16:22
266
+ Epoch 1 | iter 5024 step 157 | loss train: 1.326, val: 1.452 | iter time: 360.72 ms (step) remaining time: 0:16:11
267
+ Epoch 1 | iter 5056 step 158 | loss train: 1.331, val: 1.452 | iter time: 359.16 ms (step) remaining time: 0:16:00
268
+ Epoch 1 | iter 5088 step 159 | loss train: 1.312, val: 1.452 | iter time: 360.73 ms (step) remaining time: 0:15:48
269
+ Epoch 1 | iter 5120 step 160 | loss train: 1.320, val: 1.452 | iter time: 358.48 ms (step) remaining time: 0:15:37
270
+ Epoch 1 | iter 5152 step 161 | loss train: 1.425, val: 1.452 | iter time: 361.04 ms (step) remaining time: 0:15:26
271
+ Epoch 1 | iter 5184 step 162 | loss train: 1.359, val: 1.452 | iter time: 359.68 ms (step) remaining time: 0:15:15
272
+ Epoch 1 | iter 5216 step 163 | loss train: 1.322, val: 1.452 | iter time: 359.98 ms (step) remaining time: 0:15:03
273
+ Epoch 1 | iter 5248 step 164 | loss train: 1.362, val: 1.452 | iter time: 360.02 ms (step) remaining time: 0:14:52
274
+ Epoch 1 | iter 5280 step 165 | loss train: 1.416, val: 1.452 | iter time: 360.96 ms (step) remaining time: 0:14:41
275
+ Epoch 1 | iter 5312 step 166 | loss train: 1.361, val: 1.452 | iter time: 360.85 ms (step) remaining time: 0:14:29
276
+ Epoch 1 | iter 5344 step 167 | loss train: 1.284, val: 1.452 | iter time: 361.93 ms (step) remaining time: 0:14:18
277
+ Epoch 1 | iter 5376 step 168 | loss train: 1.341, val: 1.452 | iter time: 361.32 ms (step) remaining time: 0:14:07
278
+ Epoch 1 | iter 5408 step 169 | loss train: 1.354, val: 1.452 | iter time: 360.35 ms (step) remaining time: 0:13:56
279
+ Epoch 1 | iter 5440 step 170 | loss train: 1.324, val: 1.452 | iter time: 360.10 ms (step) remaining time: 0:13:44
280
+ Epoch 1 | iter 5472 step 171 | loss train: 1.276, val: 1.452 | iter time: 360.18 ms (step) remaining time: 0:13:33
281
+ Epoch 1 | iter 5504 step 172 | loss train: 1.378, val: 1.452 | iter time: 358.51 ms (step) remaining time: 0:13:22
282
+ Epoch 1 | iter 5536 step 173 | loss train: 1.440, val: 1.452 | iter time: 360.06 ms (step) remaining time: 0:13:11
283
+ Epoch 1 | iter 5568 step 174 | loss train: 1.353, val: 1.452 | iter time: 361.86 ms (step) remaining time: 0:12:59
284
+ Epoch 1 | iter 5600 step 175 | loss train: 1.355, val: 1.452 | iter time: 359.46 ms (step) remaining time: 0:12:48
285
+ Epoch 1 | iter 5632 step 176 | loss train: 1.359, val: 1.452 | iter time: 360.40 ms (step) remaining time: 0:12:37
286
+ Epoch 1 | iter 5664 step 177 | loss train: 1.381, val: 1.452 | iter time: 362.49 ms (step) remaining time: 0:12:26
287
+ Epoch 1 | iter 5696 step 178 | loss train: 1.301, val: 1.452 | iter time: 360.00 ms (step) remaining time: 0:12:14
288
+ Epoch 1 | iter 5728 step 179 | loss train: 1.365, val: 1.452 | iter time: 361.17 ms (step) remaining time: 0:12:03
289
+ Epoch 1 | iter 5760 step 180 | loss train: 1.239, val: 1.452 | iter time: 361.51 ms (step) remaining time: 0:11:52
290
+ Epoch 1 | iter 5792 step 181 | loss train: 1.363, val: 1.452 | iter time: 358.99 ms (step) remaining time: 0:11:41
291
+ Epoch 1 | iter 5824 step 182 | loss train: 1.392, val: 1.452 | iter time: 361.74 ms (step) remaining time: 0:11:29
292
+ Epoch 1 | iter 5856 step 183 | loss train: 1.356, val: 1.452 | iter time: 359.16 ms (step) remaining time: 0:11:18
293
+ Epoch 1 | iter 5888 step 184 | loss train: 1.277, val: 1.452 | iter time: 360.23 ms (step) remaining time: 0:11:07
294
+ Epoch 1 | iter 5920 step 185 | loss train: 1.300, val: 1.452 | iter time: 360.14 ms (step) remaining time: 0:10:56
295
+ Epoch 1 | iter 5952 step 186 | loss train: 1.326, val: 1.452 | iter time: 361.37 ms (step) remaining time: 0:10:45
296
+ Epoch 1 | iter 5984 step 187 | loss train: 1.311, val: 1.452 | iter time: 360.23 ms (step) remaining time: 0:10:33
297
+ Epoch 1 | iter 6016 step 188 | loss train: 1.275, val: 1.452 | iter time: 359.46 ms (step) remaining time: 0:10:22
298
+ Epoch 1 | iter 6048 step 189 | loss train: 1.363, val: 1.452 | iter time: 360.62 ms (step) remaining time: 0:10:11
299
+ Epoch 1 | iter 6080 step 190 | loss train: 1.369, val: 1.452 | iter time: 360.38 ms (step) remaining time: 0:10:00
300
+ Epoch 1 | iter 6112 step 191 | loss train: 1.333, val: 1.452 | iter time: 360.70 ms (step) remaining time: 0:09:49
301
+ Epoch 1 | iter 6144 step 192 | loss train: 1.371, val: 1.452 | iter time: 360.33 ms (step) remaining time: 0:09:38
302
+ Epoch 1 | iter 6176 step 193 | loss train: 1.425, val: 1.452 | iter time: 360.49 ms (step) remaining time: 0:09:26
303
+ Epoch 1 | iter 6208 step 194 | loss train: 1.347, val: 1.452 | iter time: 360.85 ms (step) remaining time: 0:09:15
304
+ Epoch 1 | iter 6240 step 195 | loss train: 1.347, val: 1.452 | iter time: 359.56 ms (step) remaining time: 0:09:04
305
+ Epoch 1 | iter 6272 step 196 | loss train: 1.288, val: 1.452 | iter time: 360.48 ms (step) remaining time: 0:08:53
306
+ Epoch 1 | iter 6304 step 197 | loss train: 1.360, val: 1.452 | iter time: 360.64 ms (step) remaining time: 0:08:42
307
+ Epoch 1 | iter 6336 step 198 | loss train: 1.344, val: 1.452 | iter time: 361.55 ms (step) remaining time: 0:08:31
308
+ Epoch 1 | iter 6368 step 199 | loss train: 1.407, val: 1.452 | iter time: 360.21 ms (step) remaining time: 0:08:19
309
+ Epoch 1 | iter 6400 step 200 | loss train: 1.399, val: 1.452 | iter time: 362.19 ms (step) remaining time: 0:08:08
310
+ Validating ...
311
+ iter 6400: val loss 1.4320, val time: 9372.73 ms
312
+ Saving checkpoint to 'out/pretrain/2410/step-00000200/lit_model.pth'
313
+ Epoch 1 | iter 6432 step 201 | loss train: 1.332, val: 1.432 | iter time: 357.97 ms (step) remaining time: 0:08:03
314
+ Epoch 1 | iter 6464 step 202 | loss train: 1.368, val: 1.432 | iter time: 358.87 ms (step) remaining time: 0:07:51
315
+ Epoch 1 | iter 6496 step 203 | loss train: 1.373, val: 1.432 | iter time: 358.71 ms (step) remaining time: 0:07:40
316
+ Epoch 1 | iter 6528 step 204 | loss train: 1.342, val: 1.432 | iter time: 361.87 ms (step) remaining time: 0:07:29
317
+ Epoch 1 | iter 6560 step 205 | loss train: 1.275, val: 1.432 | iter time: 360.69 ms (step) remaining time: 0:07:17
318
+ Epoch 1 | iter 6592 step 206 | loss train: 1.338, val: 1.432 | iter time: 363.39 ms (step) remaining time: 0:07:06
319
+ Epoch 1 | iter 6624 step 207 | loss train: 1.325, val: 1.432 | iter time: 361.00 ms (step) remaining time: 0:06:55
320
+ Epoch 1 | iter 6656 step 208 | loss train: 1.316, val: 1.432 | iter time: 361.32 ms (step) remaining time: 0:06:44
321
+ Epoch 1 | iter 6688 step 209 | loss train: 1.314, val: 1.432 | iter time: 358.59 ms (step) remaining time: 0:06:32
322
+ Epoch 1 | iter 6720 step 210 | loss train: 1.293, val: 1.432 | iter time: 359.86 ms (step) remaining time: 0:06:21
323
+ Epoch 1 | iter 6752 step 211 | loss train: 1.294, val: 1.432 | iter time: 359.83 ms (step) remaining time: 0:06:10
324
+ Epoch 1 | iter 6784 step 212 | loss train: 1.266, val: 1.432 | iter time: 359.81 ms (step) remaining time: 0:05:58
325
+ Epoch 1 | iter 6816 step 213 | loss train: 1.242, val: 1.432 | iter time: 359.95 ms (step) remaining time: 0:05:47
326
+ Epoch 1 | iter 6848 step 214 | loss train: 1.295, val: 1.432 | iter time: 360.34 ms (step) remaining time: 0:05:36
327
+ Epoch 1 | iter 6880 step 215 | loss train: 1.375, val: 1.432 | iter time: 358.12 ms (step) remaining time: 0:05:25
328
+ Epoch 1 | iter 6912 step 216 | loss train: 1.295, val: 1.432 | iter time: 359.80 ms (step) remaining time: 0:05:13
329
+ Epoch 1 | iter 6944 step 217 | loss train: 1.303, val: 1.432 | iter time: 358.80 ms (step) remaining time: 0:05:02
330
+ Epoch 1 | iter 6976 step 218 | loss train: 1.301, val: 1.432 | iter time: 358.60 ms (step) remaining time: 0:04:51
331
+ Epoch 1 | iter 7008 step 219 | loss train: 1.315, val: 1.432 | iter time: 359.93 ms (step) remaining time: 0:04:40
332
+ Epoch 1 | iter 7040 step 220 | loss train: 1.302, val: 1.432 | iter time: 359.42 ms (step) remaining time: 0:04:28
333
+ Epoch 1 | iter 7072 step 221 | loss train: 1.305, val: 1.432 | iter time: 358.54 ms (step) remaining time: 0:04:17
334
+ Epoch 1 | iter 7104 step 222 | loss train: 1.300, val: 1.432 | iter time: 360.42 ms (step) remaining time: 0:04:06
335
+ Epoch 1 | iter 7136 step 223 | loss train: 1.326, val: 1.432 | iter time: 360.95 ms (step) remaining time: 0:03:55
336
+ Epoch 1 | iter 7168 step 224 | loss train: 1.359, val: 1.432 | iter time: 359.79 ms (step) remaining time: 0:03:43
337
+ Epoch 1 | iter 7200 step 225 | loss train: 1.401, val: 1.432 | iter time: 360.84 ms (step) remaining time: 0:03:32
338
+ Epoch 1 | iter 7232 step 226 | loss train: 1.359, val: 1.432 | iter time: 361.45 ms (step) remaining time: 0:03:21
339
+ Epoch 1 | iter 7264 step 227 | loss train: 1.390, val: 1.432 | iter time: 360.51 ms (step) remaining time: 0:03:10
340
+ Epoch 1 | iter 7296 step 228 | loss train: 1.358, val: 1.432 | iter time: 360.34 ms (step) remaining time: 0:02:59
341
+ Epoch 1 | iter 7328 step 229 | loss train: 1.322, val: 1.432 | iter time: 359.47 ms (step) remaining time: 0:02:47
342
+ Epoch 1 | iter 7360 step 230 | loss train: 1.335, val: 1.432 | iter time: 359.69 ms (step) remaining time: 0:02:36
343
+ Epoch 1 | iter 7392 step 231 | loss train: 1.316, val: 1.432 | iter time: 361.00 ms (step) remaining time: 0:02:25
344
+ Epoch 1 | iter 7424 step 232 | loss train: 1.289, val: 1.432 | iter time: 359.70 ms (step) remaining time: 0:02:14
345
+ Epoch 1 | iter 7456 step 233 | loss train: 1.184, val: 1.432 | iter time: 358.77 ms (step) remaining time: 0:02:03
346
+ Epoch 1 | iter 7488 step 234 | loss train: 1.281, val: 1.432 | iter time: 361.00 ms (step) remaining time: 0:01:51
347
+ Epoch 1 | iter 7520 step 235 | loss train: 1.274, val: 1.432 | iter time: 360.44 ms (step) remaining time: 0:01:40
348
+ Epoch 1 | iter 7552 step 236 | loss train: 1.313, val: 1.432 | iter time: 359.87 ms (step) remaining time: 0:01:29
349
+ Epoch 1 | iter 7584 step 237 | loss train: 1.313, val: 1.432 | iter time: 360.86 ms (step) remaining time: 0:01:18
350
+ Epoch 1 | iter 7616 step 238 | loss train: 1.296, val: 1.432 | iter time: 361.32 ms (step) remaining time: 0:01:07
351
+ Epoch 1 | iter 7648 step 239 | loss train: 1.318, val: 1.432 | iter time: 359.93 ms (step) remaining time: 0:00:55
352
+ Epoch 1 | iter 7680 step 240 | loss train: 1.270, val: 1.432 | iter time: 358.38 ms (step) remaining time: 0:00:44
353
+ Epoch 1 | iter 7712 step 241 | loss train: 1.349, val: 1.432 | iter time: 361.13 ms (step) remaining time: 0:00:33
354
+ Epoch 1 | iter 7744 step 242 | loss train: 1.299, val: 1.432 | iter time: 359.81 ms (step) remaining time: 0:00:22
355
+ Epoch 1 | iter 7776 step 243 | loss train: 1.362, val: 1.432 | iter time: 360.26 ms (step) remaining time: 0:00:11
356
+ Epoch 1 | iter 7808 step 244 | loss train: 1.311, val: 1.432 | iter time: 360.30 ms (step) remaining time: 0:00:00
357
+ Validating ...
358
+ Final evaluation | val loss: 1.422 | val ppl: 4.145
359
+ Saving checkpoint to 'out/pretrain/2410/final/lit_model.pth'
360
+ ----------------------------------------
361
+ | Performance
362
+ | - Total tokens : 255,852,544
363
+ | - Training Time : 2764.77 s
364
+ | - Tok/sec : 229.35 tok/s
365
+ | ----------------------------------------
366
+ | Memory Usage
367
+ | - Memory Used : 26.32 GB
368
+ ----------------------------------------
out/pretrain/tinyllama/teelogs/2410_full.txt ADDED
@@ -0,0 +1,380 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Initializing distributed: GLOBAL_RANK: 0, MEMBER: 1/4
2
+ Initializing distributed: GLOBAL_RANK: 3, MEMBER: 4/4
3
+ [rank: 3] Seed set to 42
4
+ Initializing distributed: GLOBAL_RANK: 2, MEMBER: 3/4
5
+ Initializing distributed: GLOBAL_RANK: 1, MEMBER: 2/4
6
+ [rank: 2] Seed set to 42
7
+ ----------------------------------------------------------------------------------------------------
8
+ distributed_backend=nccl
9
+ All distributed processes registered. Starting with 4 processes
10
+ ----------------------------------------------------------------------------------------------------
11
+
12
+ [rank: 1] Seed set to 42
13
+ [rank: 0] Seed set to 42
14
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
15
+ train_dataloader = data.train_dataloader()
16
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
17
+ train_dataloader = data.train_dataloader()
18
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
19
+ train_dataloader = data.train_dataloader()
20
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
21
+ train_dataloader = data.train_dataloader()
22
+ All GPUs are fully connected via NVLink.
23
+ {'data': {'batch_size': 1,
24
+ 'data_path': PosixPath('litgpt/data/arxiv'),
25
+ 'num_workers': 8,
26
+ 'seed': 42,
27
+ 'seq_length': 2048,
28
+ 'use_starcoder': True},
29
+ 'data_dir': PosixPath('litgpt/data/arxiv/2410'),
30
+ 'devices': 'auto',
31
+ 'eval': {'evaluate_example': 'first',
32
+ 'final_validation': True,
33
+ 'initial_validation': True,
34
+ 'interval': 50,
35
+ 'max_iters': 200,
36
+ 'max_new_tokens': None},
37
+ 'initial_checkpoint_dir': PosixPath('out/pretrain/2409_full/final'),
38
+ 'log': {'group': None, 'project': None, 'run': None},
39
+ 'logger_name': 'tensorboard',
40
+ 'model_config': {'attention_logit_softcapping': None,
41
+ 'attention_scores_scalar': None,
42
+ 'attn_bias': False,
43
+ 'bias': False,
44
+ 'block_size': 2048,
45
+ 'final_logit_softcapping': None,
46
+ 'gelu_approximate': 'none',
47
+ 'head_size': 64,
48
+ 'hf_config': {'name': 'TinyLlama-1.1B-intermediate-step-1431k-3T',
49
+ 'org': 'TinyLlama'},
50
+ 'intermediate_size': 5632,
51
+ 'lm_head_bias': False,
52
+ 'mlp_class_name': 'LLaMAMLP',
53
+ 'moe_intermediate_size': None,
54
+ 'n_embd': 2048,
55
+ 'n_expert': 0,
56
+ 'n_expert_per_token': 0,
57
+ 'n_head': 32,
58
+ 'n_layer': 22,
59
+ 'n_query_groups': 4,
60
+ 'name': 'tiny-llama-1.1b',
61
+ 'norm_1': True,
62
+ 'norm_2': True,
63
+ 'norm_class_name': 'RMSNorm',
64
+ 'norm_eps': 1e-05,
65
+ 'norm_qk': False,
66
+ 'norm_qk_type': 'default',
67
+ 'padded_vocab_size': 32000,
68
+ 'padding_multiple': 64,
69
+ 'parallel_residual': False,
70
+ 'post_attention_norm': False,
71
+ 'post_mlp_norm': False,
72
+ 'rope_adjustments': None,
73
+ 'rope_base': 10000,
74
+ 'rope_condense_ratio': 1,
75
+ 'rope_indices': None,
76
+ 'rope_local_base_freq': None,
77
+ 'rotary_percentage': 1.0,
78
+ 'scale_embeddings': False,
79
+ 'shared_attention_norm': False,
80
+ 'sliding_window_indices': None,
81
+ 'sliding_window_size': None,
82
+ 'vocab_size': 32000},
83
+ 'model_name': 'tiny-llama-1.1b',
84
+ 'num_nodes': 1,
85
+ 'optimizer': "{'class_path': 'torch.optim.AdamW', 'init_args': {'lr': 0.0004, "
86
+ "'weight_decay': 0.1, 'betas': [0.9, 0.95]}}",
87
+ 'out_dir': PosixPath('out/pretrain/2410_full'),
88
+ 'ppl': False,
89
+ 'precision': 'bf16-mixed',
90
+ 'resume': False,
91
+ 'seed': 42,
92
+ 'tokenizer_dir': PosixPath('checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T'),
93
+ 'train': {'epochs': None,
94
+ 'global_batch_size': 512,
95
+ 'log_interval': 1,
96
+ 'lr_warmup_fraction': None,
97
+ 'lr_warmup_steps': 20,
98
+ 'max_norm': 1.0,
99
+ 'max_seq_length': 2048,
100
+ 'max_steps': None,
101
+ 'max_tokens': 258998272,
102
+ 'micro_batch_size': 4,
103
+ 'min_lr': 4e-05,
104
+ 'save_interval': 100,
105
+ 'tie_embeddings': None}}
106
+ Time to instantiate model: 0.03 seconds.
107
+ Total parameters: 1,100,048,384
108
+ [fix] out/pretrain/2409_full/final/lit_model.pth 是整包 state,提取 model 权重并原子覆盖...
109
+ [fix] 已覆盖为纯权重: out/pretrain/2409_full/final/lit_model.pth
110
+ Validating ...
111
+ Measured TFLOPs: 239.66
112
+ Epoch 1 | iter 32 step 1 | loss train: 1.468, val: 1.434 | iter time: 549.16 ms (step) remaining time: 0:49:22
113
+ Epoch 1 | iter 64 step 2 | loss train: 1.464, val: 1.434 | iter time: 357.02 ms (step) remaining time: 0:46:36
114
+ Epoch 1 | iter 96 step 3 | loss train: 1.444, val: 1.434 | iter time: 358.26 ms (step) remaining time: 0:45:35
115
+ Epoch 1 | iter 128 step 4 | loss train: 1.446, val: 1.434 | iter time: 359.15 ms (step) remaining time: 0:45:01
116
+ Epoch 1 | iter 160 step 5 | loss train: 1.403, val: 1.434 | iter time: 358.13 ms (step) remaining time: 0:44:36
117
+ Epoch 1 | iter 192 step 6 | loss train: 1.478, val: 1.434 | iter time: 358.23 ms (step) remaining time: 0:44:16
118
+ Epoch 1 | iter 224 step 7 | loss train: 1.494, val: 1.434 | iter time: 358.74 ms (step) remaining time: 0:43:59
119
+ Epoch 1 | iter 256 step 8 | loss train: 1.459, val: 1.434 | iter time: 358.28 ms (step) remaining time: 0:43:43
120
+ Epoch 1 | iter 288 step 9 | loss train: 1.464, val: 1.434 | iter time: 359.45 ms (step) remaining time: 0:43:29
121
+ Epoch 1 | iter 320 step 10 | loss train: 1.448, val: 1.434 | iter time: 358.44 ms (step) remaining time: 0:43:15
122
+ Epoch 1 | iter 352 step 11 | loss train: 1.456, val: 1.434 | iter time: 358.89 ms (step) remaining time: 0:43:02
123
+ Epoch 1 | iter 384 step 12 | loss train: 1.468, val: 1.434 | iter time: 359.20 ms (step) remaining time: 0:42:50
124
+ Epoch 1 | iter 416 step 13 | loss train: 1.488, val: 1.434 | iter time: 358.92 ms (step) remaining time: 0:42:37
125
+ Epoch 1 | iter 448 step 14 | loss train: 1.440, val: 1.434 | iter time: 359.96 ms (step) remaining time: 0:42:25
126
+ Epoch 1 | iter 480 step 15 | loss train: 1.453, val: 1.434 | iter time: 359.32 ms (step) remaining time: 0:42:13
127
+ Epoch 1 | iter 512 step 16 | loss train: 1.418, val: 1.434 | iter time: 359.56 ms (step) remaining time: 0:42:01
128
+ Epoch 1 | iter 544 step 17 | loss train: 1.485, val: 1.434 | iter time: 360.81 ms (step) remaining time: 0:41:50
129
+ Epoch 1 | iter 576 step 18 | loss train: 1.456, val: 1.434 | iter time: 361.53 ms (step) remaining time: 0:41:38
130
+ Epoch 1 | iter 608 step 19 | loss train: 1.488, val: 1.434 | iter time: 358.12 ms (step) remaining time: 0:41:27
131
+ Epoch 1 | iter 640 step 20 | loss train: 1.483, val: 1.434 | iter time: 358.79 ms (step) remaining time: 0:41:15
132
+ Epoch 1 | iter 672 step 21 | loss train: 1.458, val: 1.434 | iter time: 360.80 ms (step) remaining time: 0:41:04
133
+ Epoch 1 | iter 704 step 22 | loss train: 1.487, val: 1.434 | iter time: 360.74 ms (step) remaining time: 0:40:53
134
+ Epoch 1 | iter 736 step 23 | loss train: 1.429, val: 1.434 | iter time: 360.30 ms (step) remaining time: 0:40:41
135
+ Epoch 1 | iter 768 step 24 | loss train: 1.426, val: 1.434 | iter time: 359.60 ms (step) remaining time: 0:40:30
136
+ Epoch 1 | iter 800 step 25 | loss train: 1.426, val: 1.434 | iter time: 360.50 ms (step) remaining time: 0:40:19
137
+ Epoch 1 | iter 832 step 26 | loss train: 1.480, val: 1.434 | iter time: 360.38 ms (step) remaining time: 0:40:08
138
+ Epoch 1 | iter 864 step 27 | loss train: 1.421, val: 1.434 | iter time: 359.78 ms (step) remaining time: 0:39:57
139
+ Epoch 1 | iter 896 step 28 | loss train: 1.395, val: 1.434 | iter time: 361.38 ms (step) remaining time: 0:39:45
140
+ Epoch 1 | iter 928 step 29 | loss train: 1.423, val: 1.434 | iter time: 359.38 ms (step) remaining time: 0:39:34
141
+ Epoch 1 | iter 960 step 30 | loss train: 1.510, val: 1.434 | iter time: 360.48 ms (step) remaining time: 0:39:23
142
+ Epoch 1 | iter 992 step 31 | loss train: 1.375, val: 1.434 | iter time: 357.96 ms (step) remaining time: 0:39:12
143
+ Epoch 1 | iter 1024 step 32 | loss train: 1.507, val: 1.434 | iter time: 362.19 ms (step) remaining time: 0:39:01
144
+ Epoch 1 | iter 1056 step 33 | loss train: 1.426, val: 1.434 | iter time: 360.28 ms (step) remaining time: 0:38:50
145
+ Epoch 1 | iter 1088 step 34 | loss train: 1.427, val: 1.434 | iter time: 360.18 ms (step) remaining time: 0:38:39
146
+ Epoch 1 | iter 1120 step 35 | loss train: 1.463, val: 1.434 | iter time: 361.08 ms (step) remaining time: 0:38:28
147
+ Epoch 1 | iter 1152 step 36 | loss train: 1.465, val: 1.434 | iter time: 361.35 ms (step) remaining time: 0:38:17
148
+ Epoch 1 | iter 1184 step 37 | loss train: 1.501, val: 1.434 | iter time: 361.16 ms (step) remaining time: 0:38:06
149
+ Epoch 1 | iter 1216 step 38 | loss train: 1.468, val: 1.434 | iter time: 361.31 ms (step) remaining time: 0:37:55
150
+ Epoch 1 | iter 1248 step 39 | loss train: 1.369, val: 1.434 | iter time: 361.01 ms (step) remaining time: 0:37:44
151
+ Epoch 1 | iter 1280 step 40 | loss train: 1.486, val: 1.434 | iter time: 359.81 ms (step) remaining time: 0:37:33
152
+ Epoch 1 | iter 1312 step 41 | loss train: 1.444, val: 1.434 | iter time: 360.68 ms (step) remaining time: 0:37:22
153
+ Epoch 1 | iter 1344 step 42 | loss train: 1.534, val: 1.434 | iter time: 358.77 ms (step) remaining time: 0:37:13
154
+ Epoch 1 | iter 1376 step 43 | loss train: 1.465, val: 1.434 | iter time: 358.65 ms (step) remaining time: 0:37:02
155
+ Epoch 1 | iter 1408 step 44 | loss train: 1.473, val: 1.434 | iter time: 359.95 ms (step) remaining time: 0:36:51
156
+ Epoch 1 | iter 1440 step 45 | loss train: 1.463, val: 1.434 | iter time: 359.10 ms (step) remaining time: 0:36:40
157
+ Epoch 1 | iter 1472 step 46 | loss train: 1.460, val: 1.434 | iter time: 358.77 ms (step) remaining time: 0:36:29
158
+ Epoch 1 | iter 1504 step 47 | loss train: 1.331, val: 1.434 | iter time: 360.40 ms (step) remaining time: 0:36:18
159
+ Epoch 1 | iter 1536 step 48 | loss train: 1.467, val: 1.434 | iter time: 363.89 ms (step) remaining time: 0:36:07
160
+ Epoch 1 | iter 1568 step 49 | loss train: 1.397, val: 1.434 | iter time: 360.78 ms (step) remaining time: 0:35:56
161
+ Epoch 1 | iter 1600 step 50 | loss train: 1.426, val: 1.434 | iter time: 359.65 ms (step) remaining time: 0:35:45
162
+ Validating ...
163
+ iter 1600: val loss 1.4374, val time: 22390.38 ms
164
+ Epoch 1 | iter 1632 step 51 | loss train: 1.448, val: 1.437 | iter time: 361.75 ms (step) remaining time: 0:37:03
165
+ Epoch 1 | iter 1664 step 52 | loss train: 1.374, val: 1.437 | iter time: 360.65 ms (step) remaining time: 0:36:49
166
+ Epoch 1 | iter 1696 step 53 | loss train: 1.432, val: 1.437 | iter time: 358.93 ms (step) remaining time: 0:36:36
167
+ Epoch 1 | iter 1728 step 54 | loss train: 1.338, val: 1.437 | iter time: 360.48 ms (step) remaining time: 0:36:23
168
+ Epoch 1 | iter 1760 step 55 | loss train: 1.431, val: 1.437 | iter time: 359.98 ms (step) remaining time: 0:36:11
169
+ Epoch 1 | iter 1792 step 56 | loss train: 1.432, val: 1.437 | iter time: 361.41 ms (step) remaining time: 0:35:58
170
+ Epoch 1 | iter 1824 step 57 | loss train: 1.460, val: 1.437 | iter time: 359.26 ms (step) remaining time: 0:35:45
171
+ Epoch 1 | iter 1856 step 58 | loss train: 1.434, val: 1.437 | iter time: 361.14 ms (step) remaining time: 0:35:32
172
+ Epoch 1 | iter 1888 step 59 | loss train: 1.520, val: 1.437 | iter time: 358.19 ms (step) remaining time: 0:35:20
173
+ Epoch 1 | iter 1920 step 60 | loss train: 1.389, val: 1.437 | iter time: 359.54 ms (step) remaining time: 0:35:07
174
+ Epoch 1 | iter 1952 step 61 | loss train: 1.361, val: 1.437 | iter time: 358.54 ms (step) remaining time: 0:34:55
175
+ Epoch 1 | iter 1984 step 62 | loss train: 1.410, val: 1.437 | iter time: 359.31 ms (step) remaining time: 0:34:42
176
+ Epoch 1 | iter 2016 step 63 | loss train: 1.472, val: 1.437 | iter time: 359.21 ms (step) remaining time: 0:34:30
177
+ Epoch 1 | iter 2048 step 64 | loss train: 1.452, val: 1.437 | iter time: 359.36 ms (step) remaining time: 0:34:17
178
+ Epoch 1 | iter 2080 step 65 | loss train: 1.424, val: 1.437 | iter time: 360.00 ms (step) remaining time: 0:34:05
179
+ Epoch 1 | iter 2112 step 66 | loss train: 1.547, val: 1.437 | iter time: 358.38 ms (step) remaining time: 0:33:53
180
+ Epoch 1 | iter 2144 step 67 | loss train: 1.412, val: 1.437 | iter time: 361.05 ms (step) remaining time: 0:33:41
181
+ Epoch 1 | iter 2176 step 68 | loss train: 1.370, val: 1.437 | iter time: 360.14 ms (step) remaining time: 0:33:28
182
+ Epoch 1 | iter 2208 step 69 | loss train: 1.456, val: 1.437 | iter time: 358.71 ms (step) remaining time: 0:33:16
183
+ Epoch 1 | iter 2240 step 70 | loss train: 1.383, val: 1.437 | iter time: 357.78 ms (step) remaining time: 0:33:04
184
+ Epoch 1 | iter 2272 step 71 | loss train: 1.462, val: 1.437 | iter time: 360.52 ms (step) remaining time: 0:32:52
185
+ Epoch 1 | iter 2304 step 72 | loss train: 1.421, val: 1.437 | iter time: 359.15 ms (step) remaining time: 0:32:40
186
+ Epoch 1 | iter 2336 step 73 | loss train: 1.441, val: 1.437 | iter time: 359.55 ms (step) remaining time: 0:32:28
187
+ Epoch 1 | iter 2368 step 74 | loss train: 1.481, val: 1.437 | iter time: 358.50 ms (step) remaining time: 0:32:16
188
+ Epoch 1 | iter 2400 step 75 | loss train: 1.378, val: 1.437 | iter time: 359.39 ms (step) remaining time: 0:32:04
189
+ Epoch 1 | iter 2432 step 76 | loss train: 1.404, val: 1.437 | iter time: 358.19 ms (step) remaining time: 0:31:52
190
+ Epoch 1 | iter 2464 step 77 | loss train: 1.344, val: 1.437 | iter time: 360.52 ms (step) remaining time: 0:31:40
191
+ Epoch 1 | iter 2496 step 78 | loss train: 1.351, val: 1.437 | iter time: 360.53 ms (step) remaining time: 0:31:28
192
+ Epoch 1 | iter 2528 step 79 | loss train: 1.395, val: 1.437 | iter time: 359.22 ms (step) remaining time: 0:31:17
193
+ Epoch 1 | iter 2560 step 80 | loss train: 1.420, val: 1.437 | iter time: 360.32 ms (step) remaining time: 0:31:05
194
+ Epoch 1 | iter 2592 step 81 | loss train: 1.384, val: 1.437 | iter time: 358.82 ms (step) remaining time: 0:30:53
195
+ Epoch 1 | iter 2624 step 82 | loss train: 1.332, val: 1.437 | iter time: 360.52 ms (step) remaining time: 0:30:41
196
+ Epoch 1 | iter 2656 step 83 | loss train: 1.345, val: 1.437 | iter time: 359.73 ms (step) remaining time: 0:30:29
197
+ Epoch 1 | iter 2688 step 84 | loss train: 1.511, val: 1.437 | iter time: 359.19 ms (step) remaining time: 0:30:18
198
+ Epoch 1 | iter 2720 step 85 | loss train: 1.409, val: 1.437 | iter time: 360.14 ms (step) remaining time: 0:30:06
199
+ Epoch 1 | iter 2752 step 86 | loss train: 1.385, val: 1.437 | iter time: 359.63 ms (step) remaining time: 0:29:54
200
+ Epoch 1 | iter 2784 step 87 | loss train: 1.413, val: 1.437 | iter time: 359.79 ms (step) remaining time: 0:29:43
201
+ Epoch 1 | iter 2816 step 88 | loss train: 1.415, val: 1.437 | iter time: 358.56 ms (step) remaining time: 0:29:31
202
+ Epoch 1 | iter 2848 step 89 | loss train: 1.431, val: 1.437 | iter time: 360.75 ms (step) remaining time: 0:29:19
203
+ Epoch 1 | iter 2880 step 90 | loss train: 1.402, val: 1.437 | iter time: 362.31 ms (step) remaining time: 0:29:08
204
+ Epoch 1 | iter 2912 step 91 | loss train: 1.379, val: 1.437 | iter time: 360.99 ms (step) remaining time: 0:28:56
205
+ Epoch 1 | iter 2944 step 92 | loss train: 1.340, val: 1.437 | iter time: 360.45 ms (step) remaining time: 0:28:45
206
+ Epoch 1 | iter 2976 step 93 | loss train: 1.460, val: 1.437 | iter time: 358.90 ms (step) remaining time: 0:28:33
207
+ Epoch 1 | iter 3008 step 94 | loss train: 1.381, val: 1.437 | iter time: 360.97 ms (step) remaining time: 0:28:21
208
+ Epoch 1 | iter 3040 step 95 | loss train: 1.449, val: 1.437 | iter time: 361.44 ms (step) remaining time: 0:28:10
209
+ Epoch 1 | iter 3072 step 96 | loss train: 1.392, val: 1.437 | iter time: 360.08 ms (step) remaining time: 0:27:58
210
+ Epoch 1 | iter 3104 step 97 | loss train: 1.328, val: 1.437 | iter time: 358.03 ms (step) remaining time: 0:27:47
211
+ Epoch 1 | iter 3136 step 98 | loss train: 1.466, val: 1.437 | iter time: 360.39 ms (step) remaining time: 0:27:35
212
+ Epoch 1 | iter 3168 step 99 | loss train: 1.329, val: 1.437 | iter time: 359.32 ms (step) remaining time: 0:27:24
213
+ Epoch 1 | iter 3200 step 100 | loss train: 1.350, val: 1.437 | iter time: 360.09 ms (step) remaining time: 0:27:12
214
+ Validating ...
215
+ iter 3200: val loss 1.3736, val time: 22397.14 ms
216
+ Saving checkpoint to 'out/pretrain/2410_full/step-00000100/lit_model.pth'
217
+ Epoch 1 | iter 3232 step 101 | loss train: 1.296, val: 1.374 | iter time: 356.74 ms (step) remaining time: 0:27:58
218
+ Epoch 1 | iter 3264 step 102 | loss train: 1.429, val: 1.374 | iter time: 358.51 ms (step) remaining time: 0:27:46
219
+ Epoch 1 | iter 3296 step 103 | loss train: 1.447, val: 1.374 | iter time: 360.03 ms (step) remaining time: 0:27:33
220
+ Epoch 1 | iter 3328 step 104 | loss train: 1.404, val: 1.374 | iter time: 361.34 ms (step) remaining time: 0:27:21
221
+ Epoch 1 | iter 3360 step 105 | loss train: 1.390, val: 1.374 | iter time: 358.50 ms (step) remaining time: 0:27:09
222
+ Epoch 1 | iter 3392 step 106 | loss train: 1.379, val: 1.374 | iter time: 360.28 ms (step) remaining time: 0:26:57
223
+ Epoch 1 | iter 3424 step 107 | loss train: 1.382, val: 1.374 | iter time: 361.12 ms (step) remaining time: 0:26:45
224
+ Epoch 1 | iter 3456 step 108 | loss train: 1.447, val: 1.374 | iter time: 361.09 ms (step) remaining time: 0:26:32
225
+ Epoch 1 | iter 3488 step 109 | loss train: 1.408, val: 1.374 | iter time: 360.69 ms (step) remaining time: 0:26:20
226
+ Epoch 1 | iter 3520 step 110 | loss train: 1.467, val: 1.374 | iter time: 362.05 ms (step) remaining time: 0:26:08
227
+ Epoch 1 | iter 3552 step 111 | loss train: 1.317, val: 1.374 | iter time: 358.93 ms (step) remaining time: 0:25:56
228
+ Epoch 1 | iter 3584 step 112 | loss train: 1.327, val: 1.374 | iter time: 358.60 ms (step) remaining time: 0:25:44
229
+ Epoch 1 | iter 3616 step 113 | loss train: 1.371, val: 1.374 | iter time: 358.74 ms (step) remaining time: 0:25:31
230
+ Epoch 1 | iter 3648 step 114 | loss train: 1.401, val: 1.374 | iter time: 357.60 ms (step) remaining time: 0:25:19
231
+ Epoch 1 | iter 3680 step 115 | loss train: 1.435, val: 1.374 | iter time: 360.41 ms (step) remaining time: 0:25:08
232
+ Epoch 1 | iter 3712 step 116 | loss train: 1.363, val: 1.374 | iter time: 360.91 ms (step) remaining time: 0:24:56
233
+ Epoch 1 | iter 3744 step 117 | loss train: 1.313, val: 1.374 | iter time: 360.18 ms (step) remaining time: 0:24:44
234
+ Epoch 1 | iter 3776 step 118 | loss train: 1.358, val: 1.374 | iter time: 358.93 ms (step) remaining time: 0:24:32
235
+ Epoch 1 | iter 3808 step 119 | loss train: 1.335, val: 1.374 | iter time: 358.73 ms (step) remaining time: 0:24:20
236
+ Epoch 1 | iter 3840 step 120 | loss train: 1.353, val: 1.374 | iter time: 358.42 ms (step) remaining time: 0:24:08
237
+ Epoch 1 | iter 3872 step 121 | loss train: 1.339, val: 1.374 | iter time: 358.72 ms (step) remaining time: 0:23:56
238
+ Epoch 1 | iter 3904 step 122 | loss train: 1.269, val: 1.374 | iter time: 358.77 ms (step) remaining time: 0:23:44
239
+ Epoch 1 | iter 3936 step 123 | loss train: 1.350, val: 1.374 | iter time: 360.59 ms (step) remaining time: 0:23:32
240
+ Epoch 1 | iter 3968 step 124 | loss train: 1.408, val: 1.374 | iter time: 357.90 ms (step) remaining time: 0:23:20
241
+ Epoch 1 | iter 4000 step 125 | loss train: 1.412, val: 1.374 | iter time: 360.35 ms (step) remaining time: 0:23:08
242
+ Epoch 1 | iter 4032 step 126 | loss train: 1.362, val: 1.374 | iter time: 360.72 ms (step) remaining time: 0:22:56
243
+ Epoch 1 | iter 4064 step 127 | loss train: 1.340, val: 1.374 | iter time: 359.27 ms (step) remaining time: 0:22:44
244
+ Epoch 1 | iter 4096 step 128 | loss train: 1.354, val: 1.374 | iter time: 360.72 ms (step) remaining time: 0:22:32
245
+ Epoch 1 | iter 4128 step 129 | loss train: 1.406, val: 1.374 | iter time: 359.88 ms (step) remaining time: 0:22:21
246
+ Epoch 1 | iter 4160 step 130 | loss train: 1.392, val: 1.374 | iter time: 360.16 ms (step) remaining time: 0:22:09
247
+ Epoch 1 | iter 4192 step 131 | loss train: 1.393, val: 1.374 | iter time: 359.85 ms (step) remaining time: 0:21:57
248
+ Epoch 1 | iter 4224 step 132 | loss train: 1.390, val: 1.374 | iter time: 360.97 ms (step) remaining time: 0:21:45
249
+ Epoch 1 | iter 4256 step 133 | loss train: 1.364, val: 1.374 | iter time: 359.47 ms (step) remaining time: 0:21:33
250
+ Epoch 1 | iter 4288 step 134 | loss train: 1.314, val: 1.374 | iter time: 360.33 ms (step) remaining time: 0:21:22
251
+ Epoch 1 | iter 4320 step 135 | loss train: 1.338, val: 1.374 | iter time: 359.13 ms (step) remaining time: 0:21:10
252
+ Epoch 1 | iter 4352 step 136 | loss train: 1.309, val: 1.374 | iter time: 360.45 ms (step) remaining time: 0:20:58
253
+ Epoch 1 | iter 4384 step 137 | loss train: 1.281, val: 1.374 | iter time: 361.21 ms (step) remaining time: 0:20:46
254
+ Epoch 1 | iter 4416 step 138 | loss train: 1.397, val: 1.374 | iter time: 360.56 ms (step) remaining time: 0:20:35
255
+ Epoch 1 | iter 4448 step 139 | loss train: 1.357, val: 1.374 | iter time: 360.16 ms (step) remaining time: 0:20:23
256
+ Epoch 1 | iter 4480 step 140 | loss train: 1.417, val: 1.374 | iter time: 359.42 ms (step) remaining time: 0:20:11
257
+ Epoch 1 | iter 4512 step 141 | loss train: 1.448, val: 1.374 | iter time: 361.35 ms (step) remaining time: 0:20:00
258
+ Epoch 1 | iter 4544 step 142 | loss train: 1.455, val: 1.374 | iter time: 359.39 ms (step) remaining time: 0:19:48
259
+ Epoch 1 | iter 4576 step 143 | loss train: 1.363, val: 1.374 | iter time: 360.30 ms (step) remaining time: 0:19:36
260
+ Epoch 1 | iter 4608 step 144 | loss train: 1.340, val: 1.374 | iter time: 361.42 ms (step) remaining time: 0:19:25
261
+ Epoch 1 | iter 4640 step 145 | loss train: 1.379, val: 1.374 | iter time: 360.63 ms (step) remaining time: 0:19:13
262
+ Epoch 1 | iter 4672 step 146 | loss train: 1.417, val: 1.374 | iter time: 358.98 ms (step) remaining time: 0:19:02
263
+ Epoch 1 | iter 4704 step 147 | loss train: 1.391, val: 1.374 | iter time: 360.11 ms (step) remaining time: 0:18:50
264
+ Epoch 1 | iter 4736 step 148 | loss train: 1.326, val: 1.374 | iter time: 359.22 ms (step) remaining time: 0:18:38
265
+ Epoch 1 | iter 4768 step 149 | loss train: 1.335, val: 1.374 | iter time: 360.32 ms (step) remaining time: 0:18:27
266
+ Epoch 1 | iter 4800 step 150 | loss train: 1.320, val: 1.374 | iter time: 359.80 ms (step) remaining time: 0:18:15
267
+ Validating ...
268
+ iter 4800: val loss 1.3219, val time: 22430.10 ms
269
+ Epoch 1 | iter 4832 step 151 | loss train: 1.357, val: 1.322 | iter time: 361.07 ms (step) remaining time: 0:18:18
270
+ Epoch 1 | iter 4864 step 152 | loss train: 1.390, val: 1.322 | iter time: 359.62 ms (step) remaining time: 0:18:06
271
+ Epoch 1 | iter 4896 step 153 | loss train: 1.325, val: 1.322 | iter time: 358.47 ms (step) remaining time: 0:17:54
272
+ Epoch 1 | iter 4928 step 154 | loss train: 1.332, val: 1.322 | iter time: 359.33 ms (step) remaining time: 0:17:42
273
+ Epoch 1 | iter 4960 step 155 | loss train: 1.378, val: 1.322 | iter time: 360.33 ms (step) remaining time: 0:17:31
274
+ Epoch 1 | iter 4992 step 156 | loss train: 1.380, val: 1.322 | iter time: 361.17 ms (step) remaining time: 0:17:19
275
+ Epoch 1 | iter 5024 step 157 | loss train: 1.327, val: 1.322 | iter time: 359.73 ms (step) remaining time: 0:17:07
276
+ Epoch 1 | iter 5056 step 158 | loss train: 1.332, val: 1.322 | iter time: 360.24 ms (step) remaining time: 0:16:55
277
+ Epoch 1 | iter 5088 step 159 | loss train: 1.312, val: 1.322 | iter time: 358.35 ms (step) remaining time: 0:16:44
278
+ Epoch 1 | iter 5120 step 160 | loss train: 1.322, val: 1.322 | iter time: 359.24 ms (step) remaining time: 0:16:32
279
+ Epoch 1 | iter 5152 step 161 | loss train: 1.427, val: 1.322 | iter time: 359.44 ms (step) remaining time: 0:16:20
280
+ Epoch 1 | iter 5184 step 162 | loss train: 1.359, val: 1.322 | iter time: 360.82 ms (step) remaining time: 0:16:09
281
+ Epoch 1 | iter 5216 step 163 | loss train: 1.322, val: 1.322 | iter time: 360.22 ms (step) remaining time: 0:15:57
282
+ Epoch 1 | iter 5248 step 164 | loss train: 1.361, val: 1.322 | iter time: 360.12 ms (step) remaining time: 0:15:45
283
+ Epoch 1 | iter 5280 step 165 | loss train: 1.417, val: 1.322 | iter time: 360.10 ms (step) remaining time: 0:15:34
284
+ Epoch 1 | iter 5312 step 166 | loss train: 1.361, val: 1.322 | iter time: 360.82 ms (step) remaining time: 0:15:22
285
+ Epoch 1 | iter 5344 step 167 | loss train: 1.283, val: 1.322 | iter time: 361.00 ms (step) remaining time: 0:15:10
286
+ Epoch 1 | iter 5376 step 168 | loss train: 1.343, val: 1.322 | iter time: 360.32 ms (step) remaining time: 0:14:59
287
+ Epoch 1 | iter 5408 step 169 | loss train: 1.354, val: 1.322 | iter time: 358.65 ms (step) remaining time: 0:14:47
288
+ Epoch 1 | iter 5440 step 170 | loss train: 1.324, val: 1.322 | iter time: 359.05 ms (step) remaining time: 0:14:36
289
+ Epoch 1 | iter 5472 step 171 | loss train: 1.276, val: 1.322 | iter time: 359.95 ms (step) remaining time: 0:14:24
290
+ Epoch 1 | iter 5504 step 172 | loss train: 1.377, val: 1.322 | iter time: 360.81 ms (step) remaining time: 0:14:12
291
+ Epoch 1 | iter 5536 step 173 | loss train: 1.440, val: 1.322 | iter time: 360.85 ms (step) remaining time: 0:14:01
292
+ Epoch 1 | iter 5568 step 174 | loss train: 1.354, val: 1.322 | iter time: 360.22 ms (step) remaining time: 0:13:49
293
+ Epoch 1 | iter 5600 step 175 | loss train: 1.356, val: 1.322 | iter time: 361.09 ms (step) remaining time: 0:13:38
294
+ Epoch 1 | iter 5632 step 176 | loss train: 1.360, val: 1.322 | iter time: 359.31 ms (step) remaining time: 0:13:26
295
+ Epoch 1 | iter 5664 step 177 | loss train: 1.380, val: 1.322 | iter time: 359.61 ms (step) remaining time: 0:13:15
296
+ Epoch 1 | iter 5696 step 178 | loss train: 1.301, val: 1.322 | iter time: 359.06 ms (step) remaining time: 0:13:03
297
+ Epoch 1 | iter 5728 step 179 | loss train: 1.366, val: 1.322 | iter time: 589.34 ms (step) remaining time: 0:12:52
298
+ Epoch 1 | iter 5760 step 180 | loss train: 1.240, val: 1.322 | iter time: 358.76 ms (step) remaining time: 0:12:40
299
+ Epoch 1 | iter 5792 step 181 | loss train: 1.363, val: 1.322 | iter time: 360.55 ms (step) remaining time: 0:12:29
300
+ Epoch 1 | iter 5824 step 182 | loss train: 1.392, val: 1.322 | iter time: 361.28 ms (step) remaining time: 0:12:17
301
+ Epoch 1 | iter 5856 step 183 | loss train: 1.356, val: 1.322 | iter time: 358.77 ms (step) remaining time: 0:12:06
302
+ Epoch 1 | iter 5888 step 184 | loss train: 1.278, val: 1.322 | iter time: 359.74 ms (step) remaining time: 0:11:54
303
+ Epoch 1 | iter 5920 step 185 | loss train: 1.300, val: 1.322 | iter time: 360.26 ms (step) remaining time: 0:11:43
304
+ Epoch 1 | iter 5952 step 186 | loss train: 1.326, val: 1.322 | iter time: 358.85 ms (step) remaining time: 0:11:31
305
+ Epoch 1 | iter 5984 step 187 | loss train: 1.312, val: 1.322 | iter time: 360.39 ms (step) remaining time: 0:11:20
306
+ Epoch 1 | iter 6016 step 188 | loss train: 1.275, val: 1.322 | iter time: 360.87 ms (step) remaining time: 0:11:08
307
+ Epoch 1 | iter 6048 step 189 | loss train: 1.363, val: 1.322 | iter time: 358.40 ms (step) remaining time: 0:10:57
308
+ Epoch 1 | iter 6080 step 190 | loss train: 1.369, val: 1.322 | iter time: 360.01 ms (step) remaining time: 0:10:45
309
+ Epoch 1 | iter 6112 step 191 | loss train: 1.333, val: 1.322 | iter time: 358.96 ms (step) remaining time: 0:10:34
310
+ Epoch 1 | iter 6144 step 192 | loss train: 1.372, val: 1.322 | iter time: 360.33 ms (step) remaining time: 0:10:22
311
+ Epoch 1 | iter 6176 step 193 | loss train: 1.426, val: 1.322 | iter time: 361.20 ms (step) remaining time: 0:10:11
312
+ Epoch 1 | iter 6208 step 194 | loss train: 1.349, val: 1.322 | iter time: 359.21 ms (step) remaining time: 0:09:59
313
+ Epoch 1 | iter 6240 step 195 | loss train: 1.348, val: 1.322 | iter time: 361.67 ms (step) remaining time: 0:09:48
314
+ Epoch 1 | iter 6272 step 196 | loss train: 1.288, val: 1.322 | iter time: 360.96 ms (step) remaining time: 0:09:36
315
+ Epoch 1 | iter 6304 step 197 | loss train: 1.360, val: 1.322 | iter time: 360.50 ms (step) remaining time: 0:09:25
316
+ Epoch 1 | iter 6336 step 198 | loss train: 1.343, val: 1.322 | iter time: 359.55 ms (step) remaining time: 0:09:14
317
+ Epoch 1 | iter 6368 step 199 | loss train: 1.407, val: 1.322 | iter time: 359.32 ms (step) remaining time: 0:09:02
318
+ Epoch 1 | iter 6400 step 200 | loss train: 1.399, val: 1.322 | iter time: 358.83 ms (step) remaining time: 0:08:51
319
+ Validating ...
320
+ iter 6400: val loss 1.2789, val time: 22400.75 ms
321
+ Saving checkpoint to 'out/pretrain/2410_full/step-00000200/lit_model.pth'
322
+ Epoch 1 | iter 6432 step 201 | loss train: 1.332, val: 1.279 | iter time: 355.09 ms (step) remaining time: 0:08:48
323
+ Epoch 1 | iter 6464 step 202 | loss train: 1.369, val: 1.279 | iter time: 356.26 ms (step) remaining time: 0:08:37
324
+ Epoch 1 | iter 6496 step 203 | loss train: 1.373, val: 1.279 | iter time: 358.11 ms (step) remaining time: 0:08:25
325
+ Epoch 1 | iter 6528 step 204 | loss train: 1.341, val: 1.279 | iter time: 358.52 ms (step) remaining time: 0:08:13
326
+ Epoch 1 | iter 6560 step 205 | loss train: 1.276, val: 1.279 | iter time: 358.79 ms (step) remaining time: 0:08:02
327
+ Epoch 1 | iter 6592 step 206 | loss train: 1.338, val: 1.279 | iter time: 359.64 ms (step) remaining time: 0:07:50
328
+ Epoch 1 | iter 6624 step 207 | loss train: 1.326, val: 1.279 | iter time: 360.01 ms (step) remaining time: 0:07:39
329
+ Epoch 1 | iter 6656 step 208 | loss train: 1.316, val: 1.279 | iter time: 360.84 ms (step) remaining time: 0:07:27
330
+ Epoch 1 | iter 6688 step 209 | loss train: 1.314, val: 1.279 | iter time: 360.15 ms (step) remaining time: 0:07:15
331
+ Epoch 1 | iter 6720 step 210 | loss train: 1.293, val: 1.279 | iter time: 359.60 ms (step) remaining time: 0:07:04
332
+ Epoch 1 | iter 6752 step 211 | loss train: 1.293, val: 1.279 | iter time: 359.32 ms (step) remaining time: 0:06:52
333
+ Epoch 1 | iter 6784 step 212 | loss train: 1.265, val: 1.279 | iter time: 360.92 ms (step) remaining time: 0:06:41
334
+ Epoch 1 | iter 6816 step 213 | loss train: 1.241, val: 1.279 | iter time: 358.27 ms (step) remaining time: 0:06:29
335
+ Epoch 1 | iter 6848 step 214 | loss train: 1.295, val: 1.279 | iter time: 359.99 ms (step) remaining time: 0:06:18
336
+ Epoch 1 | iter 6880 step 215 | loss train: 1.375, val: 1.279 | iter time: 361.18 ms (step) remaining time: 0:06:06
337
+ Epoch 1 | iter 6912 step 216 | loss train: 1.296, val: 1.279 | iter time: 360.69 ms (step) remaining time: 0:05:55
338
+ Epoch 1 | iter 6944 step 217 | loss train: 1.302, val: 1.279 | iter time: 359.04 ms (step) remaining time: 0:05:43
339
+ Epoch 1 | iter 6976 step 218 | loss train: 1.301, val: 1.279 | iter time: 361.16 ms (step) remaining time: 0:05:31
340
+ Epoch 1 | iter 7008 step 219 | loss train: 1.315, val: 1.279 | iter time: 359.14 ms (step) remaining time: 0:05:20
341
+ Epoch 1 | iter 7040 step 220 | loss train: 1.302, val: 1.279 | iter time: 359.59 ms (step) remaining time: 0:05:08
342
+ Epoch 1 | iter 7072 step 221 | loss train: 1.304, val: 1.279 | iter time: 360.41 ms (step) remaining time: 0:04:57
343
+ Epoch 1 | iter 7104 step 222 | loss train: 1.300, val: 1.279 | iter time: 359.67 ms (step) remaining time: 0:04:45
344
+ Epoch 1 | iter 7136 step 223 | loss train: 1.325, val: 1.279 | iter time: 358.14 ms (step) remaining time: 0:04:34
345
+ Epoch 1 | iter 7168 step 224 | loss train: 1.360, val: 1.279 | iter time: 358.44 ms (step) remaining time: 0:04:22
346
+ Epoch 1 | iter 7200 step 225 | loss train: 1.401, val: 1.279 | iter time: 360.85 ms (step) remaining time: 0:04:11
347
+ Epoch 1 | iter 7232 step 226 | loss train: 1.359, val: 1.279 | iter time: 360.32 ms (step) remaining time: 0:03:59
348
+ Epoch 1 | iter 7264 step 227 | loss train: 1.389, val: 1.279 | iter time: 360.97 ms (step) remaining time: 0:03:48
349
+ Epoch 1 | iter 7296 step 228 | loss train: 1.358, val: 1.279 | iter time: 359.79 ms (step) remaining time: 0:03:36
350
+ Epoch 1 | iter 7328 step 229 | loss train: 1.323, val: 1.279 | iter time: 358.85 ms (step) remaining time: 0:03:25
351
+ Epoch 1 | iter 7360 step 230 | loss train: 1.334, val: 1.279 | iter time: 359.17 ms (step) remaining time: 0:03:14
352
+ Epoch 1 | iter 7392 step 231 | loss train: 1.316, val: 1.279 | iter time: 360.60 ms (step) remaining time: 0:03:02
353
+ Epoch 1 | iter 7424 step 232 | loss train: 1.289, val: 1.279 | iter time: 359.59 ms (step) remaining time: 0:02:51
354
+ Epoch 1 | iter 7456 step 233 | loss train: 1.183, val: 1.279 | iter time: 360.47 ms (step) remaining time: 0:02:39
355
+ Epoch 1 | iter 7488 step 234 | loss train: 1.281, val: 1.279 | iter time: 360.38 ms (step) remaining time: 0:02:28
356
+ Epoch 1 | iter 7520 step 235 | loss train: 1.275, val: 1.279 | iter time: 358.83 ms (step) remaining time: 0:02:16
357
+ Epoch 1 | iter 7552 step 236 | loss train: 1.313, val: 1.279 | iter time: 359.05 ms (step) remaining time: 0:02:05
358
+ Epoch 1 | iter 7584 step 237 | loss train: 1.312, val: 1.279 | iter time: 359.89 ms (step) remaining time: 0:01:54
359
+ Epoch 1 | iter 7616 step 238 | loss train: 1.295, val: 1.279 | iter time: 359.48 ms (step) remaining time: 0:01:42
360
+ Epoch 1 | iter 7648 step 239 | loss train: 1.317, val: 1.279 | iter time: 361.31 ms (step) remaining time: 0:01:31
361
+ Epoch 1 | iter 7680 step 240 | loss train: 1.271, val: 1.279 | iter time: 359.98 ms (step) remaining time: 0:01:19
362
+ Epoch 1 | iter 7712 step 241 | loss train: 1.348, val: 1.279 | iter time: 360.80 ms (step) remaining time: 0:01:08
363
+ Epoch 1 | iter 7744 step 242 | loss train: 1.297, val: 1.279 | iter time: 360.05 ms (step) remaining time: 0:00:56
364
+ Epoch 1 | iter 7776 step 243 | loss train: 1.360, val: 1.279 | iter time: 360.56 ms (step) remaining time: 0:00:45
365
+ Epoch 1 | iter 7808 step 244 | loss train: 1.309, val: 1.279 | iter time: 358.93 ms (step) remaining time: 0:00:34
366
+ Epoch 1 | iter 7840 step 245 | loss train: 1.294, val: 1.279 | iter time: 360.23 ms (step) remaining time: 0:00:22
367
+ Epoch 1 | iter 7872 step 246 | loss train: 1.249, val: 1.279 | iter time: 360.23 ms (step) remaining time: 0:00:11
368
+ Epoch 2 | iter 7904 step 247 | loss train: 1.242, val: 1.279 | iter time: 359.87 ms (step) remaining time: 0:00:00
369
+ Validating ...
370
+ Final evaluation | val loss: 1.247 | val ppl: 3.480
371
+ Saving checkpoint to 'out/pretrain/2410_full/final/lit_model.pth'
372
+ ----------------------------------------
373
+ | Performance
374
+ | - Total tokens : 258,998,272
375
+ | - Training Time : 2876.07 s
376
+ | - Tok/sec : 139.82 tok/s
377
+ | ----------------------------------------
378
+ | Memory Usage
379
+ | - Memory Used : 26.32 GB
380
+ ----------------------------------------
out/pretrain/tinyllama/teelogs/2410_lr4e-5.txt ADDED
@@ -0,0 +1,379 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Initializing distributed: GLOBAL_RANK: 0, MEMBER: 1/4
2
+ Initializing distributed: GLOBAL_RANK: 1, MEMBER: 2/4
3
+ [rank: 1] Seed set to 42
4
+ Initializing distributed: GLOBAL_RANK: 2, MEMBER: 3/4
5
+ [rank: 2] Seed set to 42
6
+ Initializing distributed: GLOBAL_RANK: 3, MEMBER: 4/4
7
+ [rank: 3] Seed set to 42
8
+ ----------------------------------------------------------------------------------------------------
9
+ distributed_backend=nccl
10
+ All distributed processes registered. Starting with 4 processes
11
+ ----------------------------------------------------------------------------------------------------
12
+
13
+ [rank: 0] Seed set to 42
14
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
15
+ train_dataloader = data.train_dataloader()
16
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
17
+ train_dataloader = data.train_dataloader()
18
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
19
+ train_dataloader = data.train_dataloader()
20
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
21
+ train_dataloader = data.train_dataloader()
22
+ All GPUs are fully connected via NVLink.
23
+ {'data': {'batch_size': 1,
24
+ 'data_path': PosixPath('litgpt/data/arxiv'),
25
+ 'num_workers': 0,
26
+ 'seed': 42,
27
+ 'seq_length': 2048,
28
+ 'use_starcoder': True},
29
+ 'data_dir': PosixPath('litgpt/data/arxiv/2410'),
30
+ 'devices': 'auto',
31
+ 'eval': {'evaluate_example': 'first',
32
+ 'final_validation': True,
33
+ 'initial_validation': True,
34
+ 'interval': 50,
35
+ 'max_iters': 200,
36
+ 'max_new_tokens': None},
37
+ 'initial_checkpoint_dir': PosixPath('out/pretrain/tinyllama/2409_full/final'),
38
+ 'log': {'group': None, 'project': None, 'run': None},
39
+ 'logger_name': 'tensorboard',
40
+ 'model_config': {'attention_logit_softcapping': None,
41
+ 'attention_scores_scalar': None,
42
+ 'attn_bias': False,
43
+ 'bias': False,
44
+ 'block_size': 2048,
45
+ 'final_logit_softcapping': None,
46
+ 'gelu_approximate': 'none',
47
+ 'head_size': 64,
48
+ 'hf_config': {'name': 'TinyLlama-1.1B-intermediate-step-1431k-3T',
49
+ 'org': 'TinyLlama'},
50
+ 'intermediate_size': 5632,
51
+ 'lm_head_bias': False,
52
+ 'mlp_class_name': 'LLaMAMLP',
53
+ 'moe_intermediate_size': None,
54
+ 'n_embd': 2048,
55
+ 'n_expert': 0,
56
+ 'n_expert_per_token': 0,
57
+ 'n_head': 32,
58
+ 'n_layer': 22,
59
+ 'n_query_groups': 4,
60
+ 'name': 'tiny-llama-1.1b',
61
+ 'norm_1': True,
62
+ 'norm_2': True,
63
+ 'norm_class_name': 'RMSNorm',
64
+ 'norm_eps': 1e-05,
65
+ 'norm_qk': False,
66
+ 'norm_qk_type': 'default',
67
+ 'padded_vocab_size': 32000,
68
+ 'padding_multiple': 64,
69
+ 'parallel_residual': False,
70
+ 'post_attention_norm': False,
71
+ 'post_mlp_norm': False,
72
+ 'rope_adjustments': None,
73
+ 'rope_base': 10000,
74
+ 'rope_condense_ratio': 1,
75
+ 'rope_indices': None,
76
+ 'rope_local_base_freq': None,
77
+ 'rotary_percentage': 1.0,
78
+ 'scale_embeddings': False,
79
+ 'shared_attention_norm': False,
80
+ 'sliding_window_indices': None,
81
+ 'sliding_window_size': None,
82
+ 'vocab_size': 32000},
83
+ 'model_name': 'tiny-llama-1.1b',
84
+ 'num_nodes': 1,
85
+ 'optimizer': "{'class_path': 'torch.optim.AdamW', 'init_args': {'lr': 4e-05, "
86
+ "'weight_decay': 0.1, 'betas': [0.9, 0.95]}}",
87
+ 'out_dir': PosixPath('out/pretrain/tinyllama/2410_lr4e-5'),
88
+ 'ppl': False,
89
+ 'precision': 'bf16-mixed',
90
+ 'resume': False,
91
+ 'seed': 42,
92
+ 'tokenizer_dir': PosixPath('checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T'),
93
+ 'train': {'epochs': None,
94
+ 'global_batch_size': 512,
95
+ 'log_interval': 1,
96
+ 'lr_warmup_fraction': None,
97
+ 'lr_warmup_steps': 20,
98
+ 'max_norm': 1.0,
99
+ 'max_seq_length': 2048,
100
+ 'max_steps': None,
101
+ 'max_tokens': 258998272,
102
+ 'micro_batch_size': 4,
103
+ 'min_lr': 4e-05,
104
+ 'save_interval': 100,
105
+ 'tie_embeddings': None}}
106
+ Time to instantiate model: 0.02 seconds.
107
+ Total parameters: 1,100,048,384
108
+ [ok] out/pretrain/tinyllama/2409_full/final/lit_model.pth 已是纯权重
109
+ Validating ...
110
+ Measured TFLOPs: 239.66
111
+ Epoch 1 | iter 32 step 1 | loss train: 1.579, val: 1.379 | iter time: 536.97 ms (step) remaining time: 0:47:57
112
+ Epoch 1 | iter 64 step 2 | loss train: 1.535, val: 1.379 | iter time: 357.22 ms (step) remaining time: 0:45:54
113
+ Epoch 1 | iter 96 step 3 | loss train: 1.516, val: 1.379 | iter time: 359.31 ms (step) remaining time: 0:45:07
114
+ Epoch 1 | iter 128 step 4 | loss train: 1.584, val: 1.379 | iter time: 358.93 ms (step) remaining time: 0:44:39
115
+ Epoch 1 | iter 160 step 5 | loss train: 1.584, val: 1.379 | iter time: 358.89 ms (step) remaining time: 0:44:19
116
+ Epoch 1 | iter 192 step 6 | loss train: 1.559, val: 1.379 | iter time: 358.77 ms (step) remaining time: 0:44:02
117
+ Epoch 1 | iter 224 step 7 | loss train: 1.581, val: 1.379 | iter time: 359.92 ms (step) remaining time: 0:43:48
118
+ Epoch 1 | iter 256 step 8 | loss train: 1.523, val: 1.379 | iter time: 358.95 ms (step) remaining time: 0:43:40
119
+ Epoch 1 | iter 288 step 9 | loss train: 1.454, val: 1.379 | iter time: 359.82 ms (step) remaining time: 0:43:26
120
+ Epoch 1 | iter 320 step 10 | loss train: 1.428, val: 1.379 | iter time: 359.87 ms (step) remaining time: 0:43:13
121
+ Epoch 1 | iter 352 step 11 | loss train: 1.415, val: 1.379 | iter time: 359.11 ms (step) remaining time: 0:43:00
122
+ Epoch 1 | iter 384 step 12 | loss train: 1.433, val: 1.379 | iter time: 358.79 ms (step) remaining time: 0:42:48
123
+ Epoch 1 | iter 416 step 13 | loss train: 1.412, val: 1.379 | iter time: 359.60 ms (step) remaining time: 0:42:36
124
+ Epoch 1 | iter 448 step 14 | loss train: 1.447, val: 1.379 | iter time: 359.45 ms (step) remaining time: 0:42:24
125
+ Epoch 1 | iter 480 step 15 | loss train: 1.393, val: 1.379 | iter time: 360.32 ms (step) remaining time: 0:42:12
126
+ Epoch 1 | iter 512 step 16 | loss train: 1.469, val: 1.379 | iter time: 359.34 ms (step) remaining time: 0:42:01
127
+ Epoch 1 | iter 544 step 17 | loss train: 1.330, val: 1.379 | iter time: 359.33 ms (step) remaining time: 0:41:49
128
+ Epoch 1 | iter 576 step 18 | loss train: 1.370, val: 1.379 | iter time: 359.64 ms (step) remaining time: 0:41:38
129
+ Epoch 1 | iter 608 step 19 | loss train: 1.357, val: 1.379 | iter time: 360.17 ms (step) remaining time: 0:41:26
130
+ Epoch 1 | iter 640 step 20 | loss train: 1.517, val: 1.379 | iter time: 360.47 ms (step) remaining time: 0:41:15
131
+ Epoch 1 | iter 672 step 21 | loss train: 1.409, val: 1.379 | iter time: 361.47 ms (step) remaining time: 0:41:03
132
+ Epoch 1 | iter 704 step 22 | loss train: 1.443, val: 1.379 | iter time: 359.09 ms (step) remaining time: 0:40:52
133
+ Epoch 1 | iter 736 step 23 | loss train: 1.436, val: 1.379 | iter time: 360.50 ms (step) remaining time: 0:40:41
134
+ Epoch 1 | iter 768 step 24 | loss train: 1.375, val: 1.379 | iter time: 359.23 ms (step) remaining time: 0:40:30
135
+ Epoch 1 | iter 800 step 25 | loss train: 1.423, val: 1.379 | iter time: 358.73 ms (step) remaining time: 0:40:19
136
+ Epoch 1 | iter 832 step 26 | loss train: 1.417, val: 1.379 | iter time: 360.01 ms (step) remaining time: 0:40:08
137
+ Epoch 1 | iter 864 step 27 | loss train: 1.422, val: 1.379 | iter time: 360.65 ms (step) remaining time: 0:39:57
138
+ Epoch 1 | iter 896 step 28 | loss train: 1.372, val: 1.379 | iter time: 359.86 ms (step) remaining time: 0:39:46
139
+ Epoch 1 | iter 928 step 29 | loss train: 1.409, val: 1.379 | iter time: 360.36 ms (step) remaining time: 0:39:34
140
+ Epoch 1 | iter 960 step 30 | loss train: 1.403, val: 1.379 | iter time: 357.53 ms (step) remaining time: 0:39:23
141
+ Epoch 1 | iter 992 step 31 | loss train: 1.405, val: 1.379 | iter time: 359.55 ms (step) remaining time: 0:39:12
142
+ Epoch 1 | iter 1024 step 32 | loss train: 1.332, val: 1.379 | iter time: 360.30 ms (step) remaining time: 0:39:01
143
+ Epoch 1 | iter 1056 step 33 | loss train: 1.379, val: 1.379 | iter time: 360.59 ms (step) remaining time: 0:38:50
144
+ Epoch 1 | iter 1088 step 34 | loss train: 1.369, val: 1.379 | iter time: 360.79 ms (step) remaining time: 0:38:39
145
+ Epoch 1 | iter 1120 step 35 | loss train: 1.325, val: 1.379 | iter time: 359.95 ms (step) remaining time: 0:38:28
146
+ Epoch 1 | iter 1152 step 36 | loss train: 1.408, val: 1.379 | iter time: 359.67 ms (step) remaining time: 0:38:17
147
+ Epoch 1 | iter 1184 step 37 | loss train: 1.338, val: 1.379 | iter time: 359.97 ms (step) remaining time: 0:38:06
148
+ Epoch 1 | iter 1216 step 38 | loss train: 1.405, val: 1.379 | iter time: 358.23 ms (step) remaining time: 0:37:55
149
+ Epoch 1 | iter 1248 step 39 | loss train: 1.398, val: 1.379 | iter time: 360.06 ms (step) remaining time: 0:37:44
150
+ Epoch 1 | iter 1280 step 40 | loss train: 1.390, val: 1.379 | iter time: 360.98 ms (step) remaining time: 0:37:33
151
+ Epoch 1 | iter 1312 step 41 | loss train: 1.357, val: 1.379 | iter time: 358.70 ms (step) remaining time: 0:37:22
152
+ Epoch 1 | iter 1344 step 42 | loss train: 1.413, val: 1.379 | iter time: 359.47 ms (step) remaining time: 0:37:11
153
+ Epoch 1 | iter 1376 step 43 | loss train: 1.390, val: 1.379 | iter time: 359.25 ms (step) remaining time: 0:37:00
154
+ Epoch 1 | iter 1408 step 44 | loss train: 1.381, val: 1.379 | iter time: 358.82 ms (step) remaining time: 0:36:49
155
+ Epoch 1 | iter 1440 step 45 | loss train: 1.391, val: 1.379 | iter time: 359.25 ms (step) remaining time: 0:36:38
156
+ Epoch 1 | iter 1472 step 46 | loss train: 1.452, val: 1.379 | iter time: 359.81 ms (step) remaining time: 0:36:29
157
+ Epoch 1 | iter 1504 step 47 | loss train: 1.389, val: 1.379 | iter time: 360.08 ms (step) remaining time: 0:36:18
158
+ Epoch 1 | iter 1536 step 48 | loss train: 1.364, val: 1.379 | iter time: 359.99 ms (step) remaining time: 0:36:07
159
+ Epoch 1 | iter 1568 step 49 | loss train: 1.368, val: 1.379 | iter time: 364.93 ms (step) remaining time: 0:35:56
160
+ Epoch 1 | iter 1600 step 50 | loss train: 1.431, val: 1.379 | iter time: 359.03 ms (step) remaining time: 0:35:45
161
+ Validating ...
162
+ iter 1600: val loss 1.3427, val time: 21931.36 ms
163
+ Epoch 1 | iter 1632 step 51 | loss train: 1.318, val: 1.343 | iter time: 359.29 ms (step) remaining time: 0:36:58
164
+ Epoch 1 | iter 1664 step 52 | loss train: 1.498, val: 1.343 | iter time: 360.88 ms (step) remaining time: 0:36:45
165
+ Epoch 1 | iter 1696 step 53 | loss train: 1.443, val: 1.343 | iter time: 360.17 ms (step) remaining time: 0:36:32
166
+ Epoch 1 | iter 1728 step 54 | loss train: 1.368, val: 1.343 | iter time: 358.24 ms (step) remaining time: 0:36:19
167
+ Epoch 1 | iter 1760 step 55 | loss train: 1.428, val: 1.343 | iter time: 720.31 ms (step) remaining time: 0:36:08
168
+ Epoch 1 | iter 1792 step 56 | loss train: 1.382, val: 1.343 | iter time: 358.18 ms (step) remaining time: 0:35:55
169
+ Epoch 1 | iter 1824 step 57 | loss train: 1.369, val: 1.343 | iter time: 357.93 ms (step) remaining time: 0:35:42
170
+ Epoch 1 | iter 1856 step 58 | loss train: 1.337, val: 1.343 | iter time: 360.48 ms (step) remaining time: 0:35:30
171
+ Epoch 1 | iter 1888 step 59 | loss train: 1.411, val: 1.343 | iter time: 358.99 ms (step) remaining time: 0:35:17
172
+ Epoch 1 | iter 1920 step 60 | loss train: 1.419, val: 1.343 | iter time: 361.85 ms (step) remaining time: 0:35:05
173
+ Epoch 1 | iter 1952 step 61 | loss train: 1.348, val: 1.343 | iter time: 359.02 ms (step) remaining time: 0:34:52
174
+ Epoch 1 | iter 1984 step 62 | loss train: 1.420, val: 1.343 | iter time: 359.25 ms (step) remaining time: 0:34:40
175
+ Epoch 1 | iter 2016 step 63 | loss train: 1.403, val: 1.343 | iter time: 359.86 ms (step) remaining time: 0:34:27
176
+ Epoch 1 | iter 2048 step 64 | loss train: 1.392, val: 1.343 | iter time: 360.11 ms (step) remaining time: 0:34:15
177
+ Epoch 1 | iter 2080 step 65 | loss train: 1.390, val: 1.343 | iter time: 360.07 ms (step) remaining time: 0:34:03
178
+ Epoch 1 | iter 2112 step 66 | loss train: 1.367, val: 1.343 | iter time: 360.96 ms (step) remaining time: 0:33:51
179
+ Epoch 1 | iter 2144 step 67 | loss train: 1.403, val: 1.343 | iter time: 360.24 ms (step) remaining time: 0:33:38
180
+ Epoch 1 | iter 2176 step 68 | loss train: 1.337, val: 1.343 | iter time: 358.68 ms (step) remaining time: 0:33:26
181
+ Epoch 1 | iter 2208 step 69 | loss train: 1.410, val: 1.343 | iter time: 361.21 ms (step) remaining time: 0:33:14
182
+ Epoch 1 | iter 2240 step 70 | loss train: 1.352, val: 1.343 | iter time: 359.72 ms (step) remaining time: 0:33:02
183
+ Epoch 1 | iter 2272 step 71 | loss train: 1.358, val: 1.343 | iter time: 359.27 ms (step) remaining time: 0:32:50
184
+ Epoch 1 | iter 2304 step 72 | loss train: 1.328, val: 1.343 | iter time: 360.73 ms (step) remaining time: 0:32:38
185
+ Epoch 1 | iter 2336 step 73 | loss train: 1.409, val: 1.343 | iter time: 358.84 ms (step) remaining time: 0:32:26
186
+ Epoch 1 | iter 2368 step 74 | loss train: 1.349, val: 1.343 | iter time: 360.37 ms (step) remaining time: 0:32:14
187
+ Epoch 1 | iter 2400 step 75 | loss train: 1.409, val: 1.343 | iter time: 360.47 ms (step) remaining time: 0:32:02
188
+ Epoch 1 | iter 2432 step 76 | loss train: 1.433, val: 1.343 | iter time: 359.26 ms (step) remaining time: 0:31:50
189
+ Epoch 1 | iter 2464 step 77 | loss train: 1.365, val: 1.343 | iter time: 360.33 ms (step) remaining time: 0:31:39
190
+ Epoch 1 | iter 2496 step 78 | loss train: 1.414, val: 1.343 | iter time: 358.72 ms (step) remaining time: 0:31:27
191
+ Epoch 1 | iter 2528 step 79 | loss train: 1.334, val: 1.343 | iter time: 360.70 ms (step) remaining time: 0:31:15
192
+ Epoch 1 | iter 2560 step 80 | loss train: 1.348, val: 1.343 | iter time: 360.65 ms (step) remaining time: 0:31:03
193
+ Epoch 1 | iter 2592 step 81 | loss train: 1.435, val: 1.343 | iter time: 360.49 ms (step) remaining time: 0:30:51
194
+ Epoch 1 | iter 2624 step 82 | loss train: 1.382, val: 1.343 | iter time: 360.39 ms (step) remaining time: 0:30:40
195
+ Epoch 1 | iter 2656 step 83 | loss train: 1.479, val: 1.343 | iter time: 358.98 ms (step) remaining time: 0:30:28
196
+ Epoch 1 | iter 2688 step 84 | loss train: 1.392, val: 1.343 | iter time: 361.67 ms (step) remaining time: 0:30:16
197
+ Epoch 1 | iter 2720 step 85 | loss train: 1.437, val: 1.343 | iter time: 359.09 ms (step) remaining time: 0:30:05
198
+ Epoch 1 | iter 2752 step 86 | loss train: 1.411, val: 1.343 | iter time: 359.38 ms (step) remaining time: 0:29:53
199
+ Epoch 1 | iter 2784 step 87 | loss train: 1.325, val: 1.343 | iter time: 360.46 ms (step) remaining time: 0:29:41
200
+ Epoch 1 | iter 2816 step 88 | loss train: 1.329, val: 1.343 | iter time: 359.10 ms (step) remaining time: 0:29:30
201
+ Epoch 1 | iter 2848 step 89 | loss train: 1.355, val: 1.343 | iter time: 358.21 ms (step) remaining time: 0:29:18
202
+ Epoch 1 | iter 2880 step 90 | loss train: 1.348, val: 1.343 | iter time: 360.35 ms (step) remaining time: 0:29:07
203
+ Epoch 1 | iter 2912 step 91 | loss train: 1.291, val: 1.343 | iter time: 359.44 ms (step) remaining time: 0:28:55
204
+ Epoch 1 | iter 2944 step 92 | loss train: 1.345, val: 1.343 | iter time: 359.07 ms (step) remaining time: 0:28:43
205
+ Epoch 1 | iter 2976 step 93 | loss train: 1.396, val: 1.343 | iter time: 359.81 ms (step) remaining time: 0:28:32
206
+ Epoch 1 | iter 3008 step 94 | loss train: 1.364, val: 1.343 | iter time: 361.59 ms (step) remaining time: 0:28:20
207
+ Epoch 1 | iter 3040 step 95 | loss train: 1.356, val: 1.343 | iter time: 360.84 ms (step) remaining time: 0:28:09
208
+ Epoch 1 | iter 3072 step 96 | loss train: 1.395, val: 1.343 | iter time: 360.73 ms (step) remaining time: 0:27:57
209
+ Epoch 1 | iter 3104 step 97 | loss train: 1.398, val: 1.343 | iter time: 359.35 ms (step) remaining time: 0:27:46
210
+ Epoch 1 | iter 3136 step 98 | loss train: 1.362, val: 1.343 | iter time: 361.78 ms (step) remaining time: 0:27:34
211
+ Epoch 1 | iter 3168 step 99 | loss train: 1.406, val: 1.343 | iter time: 359.70 ms (step) remaining time: 0:27:23
212
+ Epoch 1 | iter 3200 step 100 | loss train: 1.339, val: 1.343 | iter time: 360.47 ms (step) remaining time: 0:27:11
213
+ Validating ...
214
+ iter 3200: val loss 1.3158, val time: 21942.52 ms
215
+ Saving checkpoint to 'out/pretrain/tinyllama/2410_lr4e-5/step-00000100/lit_model.pth'
216
+ Epoch 1 | iter 3232 step 101 | loss train: 1.381, val: 1.316 | iter time: 358.25 ms (step) remaining time: 0:27:55
217
+ Epoch 1 | iter 3264 step 102 | loss train: 1.368, val: 1.316 | iter time: 359.11 ms (step) remaining time: 0:27:43
218
+ Epoch 1 | iter 3296 step 103 | loss train: 1.344, val: 1.316 | iter time: 357.26 ms (step) remaining time: 0:27:30
219
+ Epoch 1 | iter 3328 step 104 | loss train: 1.366, val: 1.316 | iter time: 360.14 ms (step) remaining time: 0:27:18
220
+ Epoch 1 | iter 3360 step 105 | loss train: 1.404, val: 1.316 | iter time: 358.91 ms (step) remaining time: 0:27:06
221
+ Epoch 1 | iter 3392 step 106 | loss train: 1.349, val: 1.316 | iter time: 358.59 ms (step) remaining time: 0:26:53
222
+ Epoch 1 | iter 3424 step 107 | loss train: 1.283, val: 1.316 | iter time: 360.36 ms (step) remaining time: 0:26:41
223
+ Epoch 1 | iter 3456 step 108 | loss train: 1.374, val: 1.316 | iter time: 358.78 ms (step) remaining time: 0:26:29
224
+ Epoch 1 | iter 3488 step 109 | loss train: 1.427, val: 1.316 | iter time: 359.47 ms (step) remaining time: 0:26:17
225
+ Epoch 1 | iter 3520 step 110 | loss train: 1.350, val: 1.316 | iter time: 360.06 ms (step) remaining time: 0:26:05
226
+ Epoch 1 | iter 3552 step 111 | loss train: 1.363, val: 1.316 | iter time: 359.22 ms (step) remaining time: 0:25:53
227
+ Epoch 1 | iter 3584 step 112 | loss train: 1.376, val: 1.316 | iter time: 359.21 ms (step) remaining time: 0:25:40
228
+ Epoch 1 | iter 3616 step 113 | loss train: 1.390, val: 1.316 | iter time: 840.22 ms (step) remaining time: 0:25:29
229
+ Epoch 1 | iter 3648 step 114 | loss train: 1.346, val: 1.316 | iter time: 360.40 ms (step) remaining time: 0:25:17
230
+ Epoch 1 | iter 3680 step 115 | loss train: 1.420, val: 1.316 | iter time: 359.43 ms (step) remaining time: 0:25:05
231
+ Epoch 1 | iter 3712 step 116 | loss train: 1.386, val: 1.316 | iter time: 359.26 ms (step) remaining time: 0:24:53
232
+ Epoch 1 | iter 3744 step 117 | loss train: 1.410, val: 1.316 | iter time: 359.65 ms (step) remaining time: 0:24:41
233
+ Epoch 1 | iter 3776 step 118 | loss train: 1.330, val: 1.316 | iter time: 360.43 ms (step) remaining time: 0:24:29
234
+ Epoch 1 | iter 3808 step 119 | loss train: 1.406, val: 1.316 | iter time: 357.91 ms (step) remaining time: 0:24:17
235
+ Epoch 1 | iter 3840 step 120 | loss train: 1.424, val: 1.316 | iter time: 358.77 ms (step) remaining time: 0:24:05
236
+ Epoch 1 | iter 3872 step 121 | loss train: 1.317, val: 1.316 | iter time: 359.10 ms (step) remaining time: 0:23:53
237
+ Epoch 1 | iter 3904 step 122 | loss train: 1.387, val: 1.316 | iter time: 359.86 ms (step) remaining time: 0:23:41
238
+ Epoch 1 | iter 3936 step 123 | loss train: 1.346, val: 1.316 | iter time: 360.78 ms (step) remaining time: 0:23:29
239
+ Epoch 1 | iter 3968 step 124 | loss train: 1.341, val: 1.316 | iter time: 358.04 ms (step) remaining time: 0:23:17
240
+ Epoch 1 | iter 4000 step 125 | loss train: 1.385, val: 1.316 | iter time: 358.34 ms (step) remaining time: 0:23:06
241
+ Epoch 1 | iter 4032 step 126 | loss train: 1.359, val: 1.316 | iter time: 358.82 ms (step) remaining time: 0:22:54
242
+ Epoch 1 | iter 4064 step 127 | loss train: 1.450, val: 1.316 | iter time: 358.75 ms (step) remaining time: 0:22:43
243
+ Epoch 1 | iter 4096 step 128 | loss train: 1.368, val: 1.316 | iter time: 359.37 ms (step) remaining time: 0:22:31
244
+ Epoch 1 | iter 4128 step 129 | loss train: 1.357, val: 1.316 | iter time: 359.95 ms (step) remaining time: 0:22:19
245
+ Epoch 1 | iter 4160 step 130 | loss train: 1.385, val: 1.316 | iter time: 357.61 ms (step) remaining time: 0:22:07
246
+ Epoch 1 | iter 4192 step 131 | loss train: 1.397, val: 1.316 | iter time: 359.87 ms (step) remaining time: 0:21:56
247
+ Epoch 1 | iter 4224 step 132 | loss train: 1.335, val: 1.316 | iter time: 357.53 ms (step) remaining time: 0:21:44
248
+ Epoch 1 | iter 4256 step 133 | loss train: 1.376, val: 1.316 | iter time: 358.33 ms (step) remaining time: 0:21:32
249
+ Epoch 1 | iter 4288 step 134 | loss train: 1.363, val: 1.316 | iter time: 359.74 ms (step) remaining time: 0:21:20
250
+ Epoch 1 | iter 4320 step 135 | loss train: 1.286, val: 1.316 | iter time: 358.13 ms (step) remaining time: 0:21:09
251
+ Epoch 1 | iter 4352 step 136 | loss train: 1.322, val: 1.316 | iter time: 360.76 ms (step) remaining time: 0:20:57
252
+ Epoch 1 | iter 4384 step 137 | loss train: 1.327, val: 1.316 | iter time: 359.59 ms (step) remaining time: 0:20:45
253
+ Epoch 1 | iter 4416 step 138 | loss train: 1.345, val: 1.316 | iter time: 359.11 ms (step) remaining time: 0:20:34
254
+ Epoch 1 | iter 4448 step 139 | loss train: 1.369, val: 1.316 | iter time: 358.76 ms (step) remaining time: 0:20:22
255
+ Epoch 1 | iter 4480 step 140 | loss train: 1.340, val: 1.316 | iter time: 360.79 ms (step) remaining time: 0:20:10
256
+ Epoch 1 | iter 4512 step 141 | loss train: 1.381, val: 1.316 | iter time: 359.95 ms (step) remaining time: 0:19:59
257
+ Epoch 1 | iter 4544 step 142 | loss train: 1.411, val: 1.316 | iter time: 359.07 ms (step) remaining time: 0:19:47
258
+ Epoch 1 | iter 4576 step 143 | loss train: 1.394, val: 1.316 | iter time: 359.39 ms (step) remaining time: 0:19:35
259
+ Epoch 1 | iter 4608 step 144 | loss train: 1.304, val: 1.316 | iter time: 359.09 ms (step) remaining time: 0:19:24
260
+ Epoch 1 | iter 4640 step 145 | loss train: 1.331, val: 1.316 | iter time: 358.95 ms (step) remaining time: 0:19:12
261
+ Epoch 1 | iter 4672 step 146 | loss train: 1.285, val: 1.316 | iter time: 359.53 ms (step) remaining time: 0:19:00
262
+ Epoch 1 | iter 4704 step 147 | loss train: 1.324, val: 1.316 | iter time: 358.91 ms (step) remaining time: 0:18:49
263
+ Epoch 1 | iter 4736 step 148 | loss train: 1.313, val: 1.316 | iter time: 359.92 ms (step) remaining time: 0:18:37
264
+ Epoch 1 | iter 4768 step 149 | loss train: 1.374, val: 1.316 | iter time: 358.69 ms (step) remaining time: 0:18:26
265
+ Epoch 1 | iter 4800 step 150 | loss train: 1.386, val: 1.316 | iter time: 358.81 ms (step) remaining time: 0:18:14
266
+ Validating ...
267
+ iter 4800: val loss 1.3066, val time: 21939.04 ms
268
+ Epoch 1 | iter 4832 step 151 | loss train: 1.398, val: 1.307 | iter time: 357.83 ms (step) remaining time: 0:18:16
269
+ Epoch 1 | iter 4864 step 152 | loss train: 1.343, val: 1.307 | iter time: 358.51 ms (step) remaining time: 0:18:05
270
+ Epoch 1 | iter 4896 step 153 | loss train: 1.341, val: 1.307 | iter time: 359.38 ms (step) remaining time: 0:17:53
271
+ Epoch 1 | iter 4928 step 154 | loss train: 1.384, val: 1.307 | iter time: 361.23 ms (step) remaining time: 0:17:41
272
+ Epoch 1 | iter 4960 step 155 | loss train: 1.322, val: 1.307 | iter time: 360.36 ms (step) remaining time: 0:17:29
273
+ Epoch 1 | iter 4992 step 156 | loss train: 1.345, val: 1.307 | iter time: 359.18 ms (step) remaining time: 0:17:18
274
+ Epoch 1 | iter 5024 step 157 | loss train: 1.326, val: 1.307 | iter time: 358.44 ms (step) remaining time: 0:17:06
275
+ Epoch 1 | iter 5056 step 158 | loss train: 1.440, val: 1.307 | iter time: 360.00 ms (step) remaining time: 0:16:54
276
+ Epoch 1 | iter 5088 step 159 | loss train: 1.359, val: 1.307 | iter time: 358.61 ms (step) remaining time: 0:16:43
277
+ Epoch 1 | iter 5120 step 160 | loss train: 1.389, val: 1.307 | iter time: 358.95 ms (step) remaining time: 0:16:31
278
+ Epoch 1 | iter 5152 step 161 | loss train: 1.365, val: 1.307 | iter time: 361.49 ms (step) remaining time: 0:16:19
279
+ Epoch 1 | iter 5184 step 162 | loss train: 1.256, val: 1.307 | iter time: 358.73 ms (step) remaining time: 0:16:08
280
+ Epoch 1 | iter 5216 step 163 | loss train: 1.318, val: 1.307 | iter time: 359.95 ms (step) remaining time: 0:15:56
281
+ Epoch 1 | iter 5248 step 164 | loss train: 1.389, val: 1.307 | iter time: 359.48 ms (step) remaining time: 0:15:44
282
+ Epoch 1 | iter 5280 step 165 | loss train: 1.329, val: 1.307 | iter time: 359.49 ms (step) remaining time: 0:15:33
283
+ Epoch 1 | iter 5312 step 166 | loss train: 1.317, val: 1.307 | iter time: 359.08 ms (step) remaining time: 0:15:21
284
+ Epoch 1 | iter 5344 step 167 | loss train: 1.367, val: 1.307 | iter time: 359.26 ms (step) remaining time: 0:15:09
285
+ Epoch 1 | iter 5376 step 168 | loss train: 1.252, val: 1.307 | iter time: 359.39 ms (step) remaining time: 0:14:58
286
+ Epoch 1 | iter 5408 step 169 | loss train: 1.337, val: 1.307 | iter time: 359.35 ms (step) remaining time: 0:14:46
287
+ Epoch 1 | iter 5440 step 170 | loss train: 1.365, val: 1.307 | iter time: 360.12 ms (step) remaining time: 0:14:35
288
+ Epoch 1 | iter 5472 step 171 | loss train: 1.270, val: 1.307 | iter time: 360.55 ms (step) remaining time: 0:14:23
289
+ Epoch 1 | iter 5504 step 172 | loss train: 1.398, val: 1.307 | iter time: 360.00 ms (step) remaining time: 0:14:11
290
+ Epoch 1 | iter 5536 step 173 | loss train: 1.412, val: 1.307 | iter time: 359.76 ms (step) remaining time: 0:14:00
291
+ Epoch 1 | iter 5568 step 174 | loss train: 1.338, val: 1.307 | iter time: 360.71 ms (step) remaining time: 0:13:48
292
+ Epoch 1 | iter 5600 step 175 | loss train: 1.366, val: 1.307 | iter time: 360.22 ms (step) remaining time: 0:13:37
293
+ Epoch 1 | iter 5632 step 176 | loss train: 1.326, val: 1.307 | iter time: 359.66 ms (step) remaining time: 0:13:25
294
+ Epoch 1 | iter 5664 step 177 | loss train: 1.432, val: 1.307 | iter time: 359.91 ms (step) remaining time: 0:13:14
295
+ Epoch 1 | iter 5696 step 178 | loss train: 1.405, val: 1.307 | iter time: 360.93 ms (step) remaining time: 0:13:02
296
+ Epoch 1 | iter 5728 step 179 | loss train: 1.292, val: 1.307 | iter time: 358.00 ms (step) remaining time: 0:12:51
297
+ Epoch 1 | iter 5760 step 180 | loss train: 1.379, val: 1.307 | iter time: 359.26 ms (step) remaining time: 0:12:39
298
+ Epoch 1 | iter 5792 step 181 | loss train: 1.291, val: 1.307 | iter time: 359.61 ms (step) remaining time: 0:12:27
299
+ Epoch 1 | iter 5824 step 182 | loss train: 1.354, val: 1.307 | iter time: 358.58 ms (step) remaining time: 0:12:16
300
+ Epoch 1 | iter 5856 step 183 | loss train: 1.341, val: 1.307 | iter time: 358.16 ms (step) remaining time: 0:12:05
301
+ Epoch 1 | iter 5888 step 184 | loss train: 1.337, val: 1.307 | iter time: 361.95 ms (step) remaining time: 0:11:53
302
+ Epoch 1 | iter 5920 step 185 | loss train: 1.403, val: 1.307 | iter time: 359.13 ms (step) remaining time: 0:11:42
303
+ Epoch 1 | iter 5952 step 186 | loss train: 1.327, val: 1.307 | iter time: 360.21 ms (step) remaining time: 0:11:30
304
+ Epoch 1 | iter 5984 step 187 | loss train: 1.338, val: 1.307 | iter time: 357.91 ms (step) remaining time: 0:11:19
305
+ Epoch 1 | iter 6016 step 188 | loss train: 1.345, val: 1.307 | iter time: 359.86 ms (step) remaining time: 0:11:07
306
+ Epoch 1 | iter 6048 step 189 | loss train: 1.340, val: 1.307 | iter time: 358.22 ms (step) remaining time: 0:10:56
307
+ Epoch 1 | iter 6080 step 190 | loss train: 1.316, val: 1.307 | iter time: 358.41 ms (step) remaining time: 0:10:44
308
+ Epoch 1 | iter 6112 step 191 | loss train: 1.280, val: 1.307 | iter time: 359.95 ms (step) remaining time: 0:10:33
309
+ Epoch 1 | iter 6144 step 192 | loss train: 1.297, val: 1.307 | iter time: 361.44 ms (step) remaining time: 0:10:21
310
+ Epoch 1 | iter 6176 step 193 | loss train: 1.335, val: 1.307 | iter time: 359.54 ms (step) remaining time: 0:10:10
311
+ Epoch 1 | iter 6208 step 194 | loss train: 1.325, val: 1.307 | iter time: 601.91 ms (step) remaining time: 0:09:59
312
+ Epoch 1 | iter 6240 step 195 | loss train: 1.319, val: 1.307 | iter time: 360.10 ms (step) remaining time: 0:09:47
313
+ Epoch 1 | iter 6272 step 196 | loss train: 1.436, val: 1.307 | iter time: 361.06 ms (step) remaining time: 0:09:36
314
+ Epoch 1 | iter 6304 step 197 | loss train: 1.370, val: 1.307 | iter time: 359.46 ms (step) remaining time: 0:09:24
315
+ Epoch 1 | iter 6336 step 198 | loss train: 1.334, val: 1.307 | iter time: 359.54 ms (step) remaining time: 0:09:13
316
+ Epoch 1 | iter 6368 step 199 | loss train: 1.338, val: 1.307 | iter time: 360.19 ms (step) remaining time: 0:09:02
317
+ Epoch 1 | iter 6400 step 200 | loss train: 1.291, val: 1.307 | iter time: 358.99 ms (step) remaining time: 0:08:50
318
+ Validating ...
319
+ iter 6400: val loss 1.3077, val time: 21938.48 ms
320
+ Saving checkpoint to 'out/pretrain/tinyllama/2410_lr4e-5/step-00000200/lit_model.pth'
321
+ Epoch 1 | iter 6432 step 201 | loss train: 1.266, val: 1.308 | iter time: 358.15 ms (step) remaining time: 0:08:48
322
+ Epoch 1 | iter 6464 step 202 | loss train: 1.420, val: 1.308 | iter time: 358.33 ms (step) remaining time: 0:08:36
323
+ Epoch 1 | iter 6496 step 203 | loss train: 1.386, val: 1.308 | iter time: 357.83 ms (step) remaining time: 0:08:24
324
+ Epoch 1 | iter 6528 step 204 | loss train: 1.381, val: 1.308 | iter time: 358.65 ms (step) remaining time: 0:08:13
325
+ Epoch 1 | iter 6560 step 205 | loss train: 1.332, val: 1.308 | iter time: 358.33 ms (step) remaining time: 0:08:01
326
+ Epoch 1 | iter 6592 step 206 | loss train: 1.342, val: 1.308 | iter time: 357.48 ms (step) remaining time: 0:07:49
327
+ Epoch 1 | iter 6624 step 207 | loss train: 1.273, val: 1.308 | iter time: 359.42 ms (step) remaining time: 0:07:38
328
+ Epoch 1 | iter 6656 step 208 | loss train: 1.347, val: 1.308 | iter time: 359.79 ms (step) remaining time: 0:07:26
329
+ Epoch 1 | iter 6688 step 209 | loss train: 1.321, val: 1.308 | iter time: 359.60 ms (step) remaining time: 0:07:15
330
+ Epoch 1 | iter 6720 step 210 | loss train: 1.356, val: 1.308 | iter time: 357.11 ms (step) remaining time: 0:07:03
331
+ Epoch 1 | iter 6752 step 211 | loss train: 1.302, val: 1.308 | iter time: 357.99 ms (step) remaining time: 0:06:52
332
+ Epoch 1 | iter 6784 step 212 | loss train: 1.326, val: 1.308 | iter time: 358.70 ms (step) remaining time: 0:06:40
333
+ Epoch 1 | iter 6816 step 213 | loss train: 1.310, val: 1.308 | iter time: 358.43 ms (step) remaining time: 0:06:29
334
+ Epoch 1 | iter 6848 step 214 | loss train: 1.315, val: 1.308 | iter time: 359.56 ms (step) remaining time: 0:06:17
335
+ Epoch 1 | iter 6880 step 215 | loss train: 1.358, val: 1.308 | iter time: 359.18 ms (step) remaining time: 0:06:06
336
+ Epoch 1 | iter 6912 step 216 | loss train: 1.358, val: 1.308 | iter time: 360.38 ms (step) remaining time: 0:05:54
337
+ Epoch 1 | iter 6944 step 217 | loss train: 1.273, val: 1.308 | iter time: 360.52 ms (step) remaining time: 0:05:42
338
+ Epoch 1 | iter 6976 step 218 | loss train: 1.302, val: 1.308 | iter time: 359.96 ms (step) remaining time: 0:05:31
339
+ Epoch 1 | iter 7008 step 219 | loss train: 1.278, val: 1.308 | iter time: 360.58 ms (step) remaining time: 0:05:19
340
+ Epoch 1 | iter 7040 step 220 | loss train: 1.260, val: 1.308 | iter time: 359.10 ms (step) remaining time: 0:05:08
341
+ Epoch 1 | iter 7072 step 221 | loss train: 1.382, val: 1.308 | iter time: 359.99 ms (step) remaining time: 0:04:56
342
+ Epoch 1 | iter 7104 step 222 | loss train: 1.306, val: 1.308 | iter time: 359.81 ms (step) remaining time: 0:04:45
343
+ Epoch 1 | iter 7136 step 223 | loss train: 1.324, val: 1.308 | iter time: 359.68 ms (step) remaining time: 0:04:34
344
+ Epoch 1 | iter 7168 step 224 | loss train: 1.346, val: 1.308 | iter time: 359.18 ms (step) remaining time: 0:04:22
345
+ Epoch 1 | iter 7200 step 225 | loss train: 1.341, val: 1.308 | iter time: 358.65 ms (step) remaining time: 0:04:11
346
+ Epoch 1 | iter 7232 step 226 | loss train: 1.374, val: 1.308 | iter time: 359.97 ms (step) remaining time: 0:03:59
347
+ Epoch 1 | iter 7264 step 227 | loss train: 1.348, val: 1.308 | iter time: 359.89 ms (step) remaining time: 0:03:48
348
+ Epoch 1 | iter 7296 step 228 | loss train: 1.372, val: 1.308 | iter time: 358.29 ms (step) remaining time: 0:03:36
349
+ Epoch 1 | iter 7328 step 229 | loss train: 1.357, val: 1.308 | iter time: 358.90 ms (step) remaining time: 0:03:25
350
+ Epoch 1 | iter 7360 step 230 | loss train: 1.391, val: 1.308 | iter time: 358.95 ms (step) remaining time: 0:03:13
351
+ Epoch 1 | iter 7392 step 231 | loss train: 1.294, val: 1.308 | iter time: 358.25 ms (step) remaining time: 0:03:02
352
+ Epoch 1 | iter 7424 step 232 | loss train: 1.358, val: 1.308 | iter time: 360.24 ms (step) remaining time: 0:02:50
353
+ Epoch 1 | iter 7456 step 233 | loss train: 1.320, val: 1.308 | iter time: 358.27 ms (step) remaining time: 0:02:39
354
+ Epoch 1 | iter 7488 step 234 | loss train: 1.354, val: 1.308 | iter time: 360.41 ms (step) remaining time: 0:02:28
355
+ Epoch 1 | iter 7520 step 235 | loss train: 1.355, val: 1.308 | iter time: 361.19 ms (step) remaining time: 0:02:16
356
+ Epoch 1 | iter 7552 step 236 | loss train: 1.352, val: 1.308 | iter time: 359.60 ms (step) remaining time: 0:02:05
357
+ Epoch 1 | iter 7584 step 237 | loss train: 1.353, val: 1.308 | iter time: 359.13 ms (step) remaining time: 0:01:53
358
+ Epoch 1 | iter 7616 step 238 | loss train: 1.296, val: 1.308 | iter time: 361.19 ms (step) remaining time: 0:01:42
359
+ Epoch 1 | iter 7648 step 239 | loss train: 1.425, val: 1.308 | iter time: 359.71 ms (step) remaining time: 0:01:31
360
+ Epoch 1 | iter 7680 step 240 | loss train: 1.341, val: 1.308 | iter time: 360.70 ms (step) remaining time: 0:01:19
361
+ Epoch 1 | iter 7712 step 241 | loss train: 1.389, val: 1.308 | iter time: 358.63 ms (step) remaining time: 0:01:08
362
+ Epoch 1 | iter 7744 step 242 | loss train: 1.308, val: 1.308 | iter time: 359.20 ms (step) remaining time: 0:00:56
363
+ Epoch 1 | iter 7776 step 243 | loss train: 1.372, val: 1.308 | iter time: 360.04 ms (step) remaining time: 0:00:45
364
+ Epoch 1 | iter 7808 step 244 | loss train: 1.375, val: 1.308 | iter time: 360.05 ms (step) remaining time: 0:00:34
365
+ Epoch 1 | iter 7840 step 245 | loss train: 1.313, val: 1.308 | iter time: 360.37 ms (step) remaining time: 0:00:22
366
+ Epoch 1 | iter 7872 step 246 | loss train: 1.275, val: 1.308 | iter time: 363.17 ms (step) remaining time: 0:00:11
367
+ Epoch 2 | iter 7904 step 247 | loss train: 1.325, val: 1.308 | iter time: 358.22 ms (step) remaining time: 0:00:00
368
+ Validating ...
369
+ Final evaluation | val loss: 1.303 | val ppl: 3.680
370
+ Saving checkpoint to 'out/pretrain/tinyllama/2410_lr4e-5/final/lit_model.pth'
371
+ ----------------------------------------
372
+ | Performance
373
+ | - Total tokens : 258,998,272
374
+ | - Training Time : 2869.92 s
375
+ | - Tok/sec : 133.59 tok/s
376
+ | ----------------------------------------
377
+ | Memory Usage
378
+ | - Memory Used : 26.32 GB
379
+ ----------------------------------------
out/pretrain/tinyllama/teelogs/2411.txt ADDED
@@ -0,0 +1,360 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Initializing distributed: GLOBAL_RANK: 0, MEMBER: 1/4
2
+ Initializing distributed: GLOBAL_RANK: 3, MEMBER: 4/4
3
+ [rank: 3] Seed set to 42
4
+ Initializing distributed: GLOBAL_RANK: 2, MEMBER: 3/4
5
+ [rank: 2] Seed set to 42
6
+ Initializing distributed: GLOBAL_RANK: 1, MEMBER: 2/4
7
+ [rank: 1] Seed set to 42
8
+ ----------------------------------------------------------------------------------------------------
9
+ distributed_backend=nccl
10
+ All distributed processes registered. Starting with 4 processes
11
+ ----------------------------------------------------------------------------------------------------
12
+
13
+ [rank: 0] Seed set to 42
14
+ All GPUs are fully connected via NVLink.
15
+ {'data': {'batch_size': 1,
16
+ 'data_path': PosixPath('litgpt/data/arxiv'),
17
+ 'num_workers': 8,
18
+ 'seed': 42,
19
+ 'seq_length': 2048,
20
+ 'use_starcoder': True},
21
+ 'data_dir': PosixPath('litgpt/data/arxiv/2411'),
22
+ 'devices': 'auto',
23
+ 'eval': {'evaluate_example': 'first',
24
+ 'final_validation': True,
25
+ 'initial_validation': True,
26
+ 'interval': 50,
27
+ 'max_iters': 100,
28
+ 'max_new_tokens': None},
29
+ 'initial_checkpoint_dir': PosixPath('out/pretrain/2410/final'),
30
+ 'log': {'group': None, 'project': None, 'run': None},
31
+ 'logger_name': 'tensorboard',
32
+ 'model_config': {'attention_logit_softcapping': None,
33
+ 'attention_scores_scalar': None,
34
+ 'attn_bias': False,
35
+ 'bias': False,
36
+ 'block_size': 2048,
37
+ 'final_logit_softcapping': None,
38
+ 'gelu_approximate': 'none',
39
+ 'head_size': 64,
40
+ 'hf_config': {'name': 'TinyLlama-1.1B-intermediate-step-1431k-3T',
41
+ 'org': 'TinyLlama'},
42
+ 'intermediate_size': 5632,
43
+ 'lm_head_bias': False,
44
+ 'mlp_class_name': 'LLaMAMLP',
45
+ 'moe_intermediate_size': None,
46
+ 'n_embd': 2048,
47
+ 'n_expert': 0,
48
+ 'n_expert_per_token': 0,
49
+ 'n_head': 32,
50
+ 'n_layer': 22,
51
+ 'n_query_groups': 4,
52
+ 'name': 'tiny-llama-1.1b',
53
+ 'norm_1': True,
54
+ 'norm_2': True,
55
+ 'norm_class_name': 'RMSNorm',
56
+ 'norm_eps': 1e-05,
57
+ 'norm_qk': False,
58
+ 'norm_qk_type': 'default',
59
+ 'padded_vocab_size': 32000,
60
+ 'padding_multiple': 64,
61
+ 'parallel_residual': False,
62
+ 'post_attention_norm': False,
63
+ 'post_mlp_norm': False,
64
+ 'rope_adjustments': None,
65
+ 'rope_base': 10000,
66
+ 'rope_condense_ratio': 1,
67
+ 'rope_indices': None,
68
+ 'rope_local_base_freq': None,
69
+ 'rotary_percentage': 1.0,
70
+ 'scale_embeddings': False,
71
+ 'shared_attention_norm': False,
72
+ 'sliding_window_indices': None,
73
+ 'sliding_window_size': None,
74
+ 'vocab_size': 32000},
75
+ 'model_name': 'tiny-llama-1.1b',
76
+ 'num_nodes': 1,
77
+ 'optimizer': "{'class_path': 'torch.optim.AdamW', 'init_args': {'lr': 0.0004, "
78
+ "'weight_decay': 0.1, 'betas': [0.9, 0.95]}}",
79
+ 'out_dir': PosixPath('out/pretrain/2411'),
80
+ 'precision': 'bf16-mixed',
81
+ 'resume': False,
82
+ 'seed': 42,
83
+ 'tokenizer_dir': PosixPath('checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T'),
84
+ 'train': {'epochs': None,
85
+ 'global_batch_size': 512,
86
+ 'log_interval': 1,
87
+ 'lr_warmup_fraction': None,
88
+ 'lr_warmup_steps': 20,
89
+ 'max_norm': 1.0,
90
+ 'max_seq_length': 2048,
91
+ 'max_steps': None,
92
+ 'max_tokens': 247463936,
93
+ 'micro_batch_size': 4,
94
+ 'min_lr': 4e-05,
95
+ 'save_interval': 100,
96
+ 'tie_embeddings': None}}
97
+ Time to instantiate model: 0.02 seconds.
98
+ Total parameters: 1,100,048,384
99
+ [fix] out/pretrain/2410/final/lit_model.pth 是整包 state,提取 model 权重并原子覆盖...
100
+ [fix] 已覆盖为纯权重: out/pretrain/2410/final/lit_model.pth
101
+ Validating ...
102
+ Measured TFLOPs: 239.66
103
+ Epoch 1 | iter 32 step 1 | loss train: 1.384, val: 1.368 | iter time: 533.08 ms (step) remaining time: 0:46:31
104
+ Epoch 1 | iter 64 step 2 | loss train: 1.336, val: 1.368 | iter time: 358.01 ms (step) remaining time: 0:44:05
105
+ Epoch 1 | iter 96 step 3 | loss train: 1.355, val: 1.368 | iter time: 357.77 ms (step) remaining time: 0:43:12
106
+ Epoch 1 | iter 128 step 4 | loss train: 1.369, val: 1.368 | iter time: 354.34 ms (step) remaining time: 0:42:41
107
+ Epoch 1 | iter 160 step 5 | loss train: 1.488, val: 1.368 | iter time: 357.29 ms (step) remaining time: 0:42:19
108
+ Epoch 1 | iter 192 step 6 | loss train: 1.450, val: 1.368 | iter time: 359.57 ms (step) remaining time: 0:42:01
109
+ Epoch 1 | iter 224 step 7 | loss train: 1.324, val: 1.368 | iter time: 358.56 ms (step) remaining time: 0:41:46
110
+ Epoch 1 | iter 256 step 8 | loss train: 1.463, val: 1.368 | iter time: 358.41 ms (step) remaining time: 0:41:32
111
+ Epoch 1 | iter 288 step 9 | loss train: 1.444, val: 1.368 | iter time: 359.05 ms (step) remaining time: 0:41:18
112
+ Epoch 1 | iter 320 step 10 | loss train: 1.408, val: 1.368 | iter time: 361.15 ms (step) remaining time: 0:41:06
113
+ Epoch 1 | iter 352 step 11 | loss train: 1.416, val: 1.368 | iter time: 359.05 ms (step) remaining time: 0:40:53
114
+ Epoch 1 | iter 384 step 12 | loss train: 1.422, val: 1.368 | iter time: 358.24 ms (step) remaining time: 0:40:41
115
+ Epoch 1 | iter 416 step 13 | loss train: 1.382, val: 1.368 | iter time: 360.11 ms (step) remaining time: 0:40:29
116
+ Epoch 1 | iter 448 step 14 | loss train: 1.470, val: 1.368 | iter time: 358.91 ms (step) remaining time: 0:40:18
117
+ Epoch 1 | iter 480 step 15 | loss train: 1.357, val: 1.368 | iter time: 356.48 ms (step) remaining time: 0:40:06
118
+ Epoch 1 | iter 512 step 16 | loss train: 1.457, val: 1.368 | iter time: 360.12 ms (step) remaining time: 0:39:55
119
+ Epoch 1 | iter 544 step 17 | loss train: 1.420, val: 1.368 | iter time: 360.19 ms (step) remaining time: 0:39:43
120
+ Epoch 1 | iter 576 step 18 | loss train: 1.354, val: 1.368 | iter time: 358.56 ms (step) remaining time: 0:39:32
121
+ Epoch 1 | iter 608 step 19 | loss train: 1.424, val: 1.368 | iter time: 361.34 ms (step) remaining time: 0:39:21
122
+ Epoch 1 | iter 640 step 20 | loss train: 1.486, val: 1.368 | iter time: 359.60 ms (step) remaining time: 0:39:10
123
+ Epoch 1 | iter 672 step 21 | loss train: 1.439, val: 1.368 | iter time: 359.33 ms (step) remaining time: 0:38:59
124
+ Epoch 1 | iter 704 step 22 | loss train: 1.404, val: 1.368 | iter time: 359.51 ms (step) remaining time: 0:38:48
125
+ Epoch 1 | iter 736 step 23 | loss train: 1.415, val: 1.368 | iter time: 360.30 ms (step) remaining time: 0:38:37
126
+ Epoch 1 | iter 768 step 24 | loss train: 1.389, val: 1.368 | iter time: 359.13 ms (step) remaining time: 0:38:26
127
+ Epoch 1 | iter 800 step 25 | loss train: 1.428, val: 1.368 | iter time: 360.96 ms (step) remaining time: 0:38:15
128
+ Epoch 1 | iter 832 step 26 | loss train: 1.476, val: 1.368 | iter time: 359.89 ms (step) remaining time: 0:38:04
129
+ Epoch 1 | iter 864 step 27 | loss train: 1.432, val: 1.368 | iter time: 361.16 ms (step) remaining time: 0:37:53
130
+ Epoch 1 | iter 896 step 28 | loss train: 1.451, val: 1.368 | iter time: 359.82 ms (step) remaining time: 0:37:42
131
+ Epoch 1 | iter 928 step 29 | loss train: 1.438, val: 1.368 | iter time: 358.61 ms (step) remaining time: 0:37:31
132
+ Epoch 1 | iter 960 step 30 | loss train: 1.334, val: 1.368 | iter time: 358.93 ms (step) remaining time: 0:37:20
133
+ Epoch 1 | iter 992 step 31 | loss train: 1.447, val: 1.368 | iter time: 359.07 ms (step) remaining time: 0:37:09
134
+ Epoch 1 | iter 1024 step 32 | loss train: 1.439, val: 1.368 | iter time: 358.27 ms (step) remaining time: 0:36:59
135
+ Epoch 1 | iter 1056 step 33 | loss train: 1.355, val: 1.368 | iter time: 358.88 ms (step) remaining time: 0:36:48
136
+ Epoch 1 | iter 1088 step 34 | loss train: 1.349, val: 1.368 | iter time: 359.38 ms (step) remaining time: 0:36:37
137
+ Epoch 1 | iter 1120 step 35 | loss train: 1.491, val: 1.368 | iter time: 361.12 ms (step) remaining time: 0:36:26
138
+ Epoch 1 | iter 1152 step 36 | loss train: 1.405, val: 1.368 | iter time: 360.52 ms (step) remaining time: 0:36:15
139
+ Epoch 1 | iter 1184 step 37 | loss train: 1.444, val: 1.368 | iter time: 359.66 ms (step) remaining time: 0:36:04
140
+ Epoch 1 | iter 1216 step 38 | loss train: 1.419, val: 1.368 | iter time: 361.19 ms (step) remaining time: 0:35:53
141
+ Epoch 1 | iter 1248 step 39 | loss train: 1.442, val: 1.368 | iter time: 740.53 ms (step) remaining time: 0:35:44
142
+ Epoch 1 | iter 1280 step 40 | loss train: 1.410, val: 1.368 | iter time: 359.23 ms (step) remaining time: 0:35:33
143
+ Epoch 1 | iter 1312 step 41 | loss train: 1.447, val: 1.368 | iter time: 359.81 ms (step) remaining time: 0:35:22
144
+ Epoch 1 | iter 1344 step 42 | loss train: 1.471, val: 1.368 | iter time: 360.20 ms (step) remaining time: 0:35:11
145
+ Epoch 1 | iter 1376 step 43 | loss train: 1.330, val: 1.368 | iter time: 360.02 ms (step) remaining time: 0:35:00
146
+ Epoch 1 | iter 1408 step 44 | loss train: 1.475, val: 1.368 | iter time: 360.05 ms (step) remaining time: 0:34:50
147
+ Epoch 1 | iter 1440 step 45 | loss train: 1.403, val: 1.368 | iter time: 358.60 ms (step) remaining time: 0:34:39
148
+ Epoch 1 | iter 1472 step 46 | loss train: 1.459, val: 1.368 | iter time: 359.23 ms (step) remaining time: 0:34:28
149
+ Epoch 1 | iter 1504 step 47 | loss train: 1.436, val: 1.368 | iter time: 360.80 ms (step) remaining time: 0:34:17
150
+ Epoch 1 | iter 1536 step 48 | loss train: 1.388, val: 1.368 | iter time: 359.82 ms (step) remaining time: 0:34:06
151
+ Epoch 1 | iter 1568 step 49 | loss train: 1.371, val: 1.368 | iter time: 361.93 ms (step) remaining time: 0:33:55
152
+ Epoch 1 | iter 1600 step 50 | loss train: 1.383, val: 1.368 | iter time: 358.59 ms (step) remaining time: 0:33:44
153
+ Validating ...
154
+ iter 1600: val loss 1.4162, val time: 9248.40 ms
155
+ Epoch 1 | iter 1632 step 51 | loss train: 1.403, val: 1.416 | iter time: 359.61 ms (step) remaining time: 0:34:06
156
+ Epoch 1 | iter 1664 step 52 | loss train: 1.407, val: 1.416 | iter time: 358.95 ms (step) remaining time: 0:33:55
157
+ Epoch 1 | iter 1696 step 53 | loss train: 1.438, val: 1.416 | iter time: 360.01 ms (step) remaining time: 0:33:43
158
+ Epoch 1 | iter 1728 step 54 | loss train: 1.404, val: 1.416 | iter time: 360.49 ms (step) remaining time: 0:33:31
159
+ Epoch 1 | iter 1760 step 55 | loss train: 1.431, val: 1.416 | iter time: 359.98 ms (step) remaining time: 0:33:20
160
+ Epoch 1 | iter 1792 step 56 | loss train: 1.512, val: 1.416 | iter time: 361.11 ms (step) remaining time: 0:33:08
161
+ Epoch 1 | iter 1824 step 57 | loss train: 1.416, val: 1.416 | iter time: 360.76 ms (step) remaining time: 0:32:56
162
+ Epoch 1 | iter 1856 step 58 | loss train: 1.393, val: 1.416 | iter time: 360.52 ms (step) remaining time: 0:32:45
163
+ Epoch 1 | iter 1888 step 59 | loss train: 1.444, val: 1.416 | iter time: 360.60 ms (step) remaining time: 0:32:33
164
+ Epoch 1 | iter 1920 step 60 | loss train: 1.461, val: 1.416 | iter time: 360.63 ms (step) remaining time: 0:32:22
165
+ Epoch 1 | iter 1952 step 61 | loss train: 1.432, val: 1.416 | iter time: 360.83 ms (step) remaining time: 0:32:10
166
+ Epoch 1 | iter 1984 step 62 | loss train: 1.344, val: 1.416 | iter time: 359.97 ms (step) remaining time: 0:31:59
167
+ Epoch 1 | iter 2016 step 63 | loss train: 1.396, val: 1.416 | iter time: 359.48 ms (step) remaining time: 0:31:47
168
+ Epoch 1 | iter 2048 step 64 | loss train: 1.507, val: 1.416 | iter time: 358.87 ms (step) remaining time: 0:31:36
169
+ Epoch 1 | iter 2080 step 65 | loss train: 1.382, val: 1.416 | iter time: 361.07 ms (step) remaining time: 0:31:24
170
+ Epoch 1 | iter 2112 step 66 | loss train: 1.400, val: 1.416 | iter time: 359.70 ms (step) remaining time: 0:31:13
171
+ Epoch 1 | iter 2144 step 67 | loss train: 1.398, val: 1.416 | iter time: 360.04 ms (step) remaining time: 0:31:01
172
+ Epoch 1 | iter 2176 step 68 | loss train: 1.426, val: 1.416 | iter time: 361.01 ms (step) remaining time: 0:30:50
173
+ Epoch 1 | iter 2208 step 69 | loss train: 1.490, val: 1.416 | iter time: 359.98 ms (step) remaining time: 0:30:39
174
+ Epoch 1 | iter 2240 step 70 | loss train: 1.478, val: 1.416 | iter time: 360.47 ms (step) remaining time: 0:30:27
175
+ Epoch 1 | iter 2272 step 71 | loss train: 1.422, val: 1.416 | iter time: 359.49 ms (step) remaining time: 0:30:16
176
+ Epoch 1 | iter 2304 step 72 | loss train: 1.412, val: 1.416 | iter time: 360.69 ms (step) remaining time: 0:30:05
177
+ Epoch 1 | iter 2336 step 73 | loss train: 1.459, val: 1.416 | iter time: 359.38 ms (step) remaining time: 0:29:53
178
+ Epoch 1 | iter 2368 step 74 | loss train: 1.406, val: 1.416 | iter time: 359.04 ms (step) remaining time: 0:29:42
179
+ Epoch 1 | iter 2400 step 75 | loss train: 1.379, val: 1.416 | iter time: 359.77 ms (step) remaining time: 0:29:31
180
+ Epoch 1 | iter 2432 step 76 | loss train: 1.362, val: 1.416 | iter time: 360.03 ms (step) remaining time: 0:29:19
181
+ Epoch 1 | iter 2464 step 77 | loss train: 1.351, val: 1.416 | iter time: 358.80 ms (step) remaining time: 0:29:08
182
+ Epoch 1 | iter 2496 step 78 | loss train: 1.492, val: 1.416 | iter time: 359.87 ms (step) remaining time: 0:28:57
183
+ Epoch 1 | iter 2528 step 79 | loss train: 1.343, val: 1.416 | iter time: 359.52 ms (step) remaining time: 0:28:45
184
+ Epoch 1 | iter 2560 step 80 | loss train: 1.408, val: 1.416 | iter time: 361.03 ms (step) remaining time: 0:28:34
185
+ Epoch 1 | iter 2592 step 81 | loss train: 1.454, val: 1.416 | iter time: 358.81 ms (step) remaining time: 0:28:23
186
+ Epoch 1 | iter 2624 step 82 | loss train: 1.472, val: 1.416 | iter time: 359.54 ms (step) remaining time: 0:28:12
187
+ Epoch 1 | iter 2656 step 83 | loss train: 1.431, val: 1.416 | iter time: 361.34 ms (step) remaining time: 0:28:01
188
+ Epoch 1 | iter 2688 step 84 | loss train: 1.431, val: 1.416 | iter time: 360.81 ms (step) remaining time: 0:27:50
189
+ Epoch 1 | iter 2720 step 85 | loss train: 1.389, val: 1.416 | iter time: 361.35 ms (step) remaining time: 0:27:38
190
+ Epoch 1 | iter 2752 step 86 | loss train: 1.345, val: 1.416 | iter time: 361.94 ms (step) remaining time: 0:27:27
191
+ Epoch 1 | iter 2784 step 87 | loss train: 1.425, val: 1.416 | iter time: 359.44 ms (step) remaining time: 0:27:16
192
+ Epoch 1 | iter 2816 step 88 | loss train: 1.421, val: 1.416 | iter time: 362.02 ms (step) remaining time: 0:27:05
193
+ Epoch 1 | iter 2848 step 89 | loss train: 1.388, val: 1.416 | iter time: 360.05 ms (step) remaining time: 0:26:54
194
+ Epoch 1 | iter 2880 step 90 | loss train: 1.451, val: 1.416 | iter time: 361.13 ms (step) remaining time: 0:26:42
195
+ Epoch 1 | iter 2912 step 91 | loss train: 1.421, val: 1.416 | iter time: 358.93 ms (step) remaining time: 0:26:32
196
+ Epoch 1 | iter 2944 step 92 | loss train: 1.320, val: 1.416 | iter time: 360.94 ms (step) remaining time: 0:26:21
197
+ Epoch 1 | iter 2976 step 93 | loss train: 1.451, val: 1.416 | iter time: 358.59 ms (step) remaining time: 0:26:09
198
+ Epoch 1 | iter 3008 step 94 | loss train: 1.450, val: 1.416 | iter time: 359.83 ms (step) remaining time: 0:25:58
199
+ Epoch 1 | iter 3040 step 95 | loss train: 1.451, val: 1.416 | iter time: 360.40 ms (step) remaining time: 0:25:47
200
+ Epoch 1 | iter 3072 step 96 | loss train: 1.413, val: 1.416 | iter time: 360.42 ms (step) remaining time: 0:25:36
201
+ Epoch 1 | iter 3104 step 97 | loss train: 1.410, val: 1.416 | iter time: 360.21 ms (step) remaining time: 0:25:25
202
+ Epoch 1 | iter 3136 step 98 | loss train: 1.428, val: 1.416 | iter time: 360.18 ms (step) remaining time: 0:25:14
203
+ Epoch 1 | iter 3168 step 99 | loss train: 1.398, val: 1.416 | iter time: 358.78 ms (step) remaining time: 0:25:03
204
+ Epoch 1 | iter 3200 step 100 | loss train: 1.362, val: 1.416 | iter time: 359.73 ms (step) remaining time: 0:24:51
205
+ Validating ...
206
+ iter 3200: val loss 1.4050, val time: 9254.27 ms
207
+ Saving checkpoint to 'out/pretrain/2411/step-00000100/lit_model.pth'
208
+ Epoch 1 | iter 3232 step 101 | loss train: 1.439, val: 1.405 | iter time: 357.45 ms (step) remaining time: 0:25:15
209
+ Epoch 1 | iter 3264 step 102 | loss train: 1.414, val: 1.405 | iter time: 356.66 ms (step) remaining time: 0:25:03
210
+ Epoch 1 | iter 3296 step 103 | loss train: 1.393, val: 1.405 | iter time: 359.91 ms (step) remaining time: 0:24:52
211
+ Epoch 1 | iter 3328 step 104 | loss train: 1.375, val: 1.405 | iter time: 357.56 ms (step) remaining time: 0:24:40
212
+ Epoch 1 | iter 3360 step 105 | loss train: 1.360, val: 1.405 | iter time: 360.33 ms (step) remaining time: 0:24:28
213
+ Epoch 1 | iter 3392 step 106 | loss train: 1.416, val: 1.405 | iter time: 359.51 ms (step) remaining time: 0:24:17
214
+ Epoch 1 | iter 3424 step 107 | loss train: 1.431, val: 1.405 | iter time: 359.40 ms (step) remaining time: 0:24:05
215
+ Epoch 1 | iter 3456 step 108 | loss train: 1.393, val: 1.405 | iter time: 360.01 ms (step) remaining time: 0:23:53
216
+ Epoch 1 | iter 3488 step 109 | loss train: 1.456, val: 1.405 | iter time: 358.84 ms (step) remaining time: 0:23:42
217
+ Epoch 1 | iter 3520 step 110 | loss train: 1.387, val: 1.405 | iter time: 360.06 ms (step) remaining time: 0:23:30
218
+ Epoch 1 | iter 3552 step 111 | loss train: 1.333, val: 1.405 | iter time: 361.29 ms (step) remaining time: 0:23:19
219
+ Epoch 1 | iter 3584 step 112 | loss train: 1.362, val: 1.405 | iter time: 359.46 ms (step) remaining time: 0:23:07
220
+ Epoch 1 | iter 3616 step 113 | loss train: 1.385, val: 1.405 | iter time: 360.14 ms (step) remaining time: 0:22:56
221
+ Epoch 1 | iter 3648 step 114 | loss train: 1.411, val: 1.405 | iter time: 359.16 ms (step) remaining time: 0:22:44
222
+ Epoch 1 | iter 3680 step 115 | loss train: 1.415, val: 1.405 | iter time: 358.68 ms (step) remaining time: 0:22:32
223
+ Epoch 1 | iter 3712 step 116 | loss train: 1.399, val: 1.405 | iter time: 358.66 ms (step) remaining time: 0:22:21
224
+ Epoch 1 | iter 3744 step 117 | loss train: 1.375, val: 1.405 | iter time: 359.46 ms (step) remaining time: 0:22:09
225
+ Epoch 1 | iter 3776 step 118 | loss train: 1.354, val: 1.405 | iter time: 360.39 ms (step) remaining time: 0:21:58
226
+ Epoch 1 | iter 3808 step 119 | loss train: 1.386, val: 1.405 | iter time: 360.57 ms (step) remaining time: 0:21:47
227
+ Epoch 1 | iter 3840 step 120 | loss train: 1.388, val: 1.405 | iter time: 359.30 ms (step) remaining time: 0:21:35
228
+ Epoch 1 | iter 3872 step 121 | loss train: 1.384, val: 1.405 | iter time: 359.24 ms (step) remaining time: 0:21:24
229
+ Epoch 1 | iter 3904 step 122 | loss train: 1.404, val: 1.405 | iter time: 361.93 ms (step) remaining time: 0:21:12
230
+ Epoch 1 | iter 3936 step 123 | loss train: 1.447, val: 1.405 | iter time: 359.03 ms (step) remaining time: 0:21:01
231
+ Epoch 1 | iter 3968 step 124 | loss train: 1.338, val: 1.405 | iter time: 361.84 ms (step) remaining time: 0:20:49
232
+ Epoch 1 | iter 4000 step 125 | loss train: 1.384, val: 1.405 | iter time: 359.67 ms (step) remaining time: 0:20:38
233
+ Epoch 1 | iter 4032 step 126 | loss train: 1.360, val: 1.405 | iter time: 361.66 ms (step) remaining time: 0:20:26
234
+ Epoch 1 | iter 4064 step 127 | loss train: 1.383, val: 1.405 | iter time: 359.82 ms (step) remaining time: 0:20:15
235
+ Epoch 1 | iter 4096 step 128 | loss train: 1.347, val: 1.405 | iter time: 361.78 ms (step) remaining time: 0:20:04
236
+ Epoch 1 | iter 4128 step 129 | loss train: 1.381, val: 1.405 | iter time: 360.25 ms (step) remaining time: 0:19:52
237
+ Epoch 1 | iter 4160 step 130 | loss train: 1.412, val: 1.405 | iter time: 360.70 ms (step) remaining time: 0:19:41
238
+ Epoch 1 | iter 4192 step 131 | loss train: 1.339, val: 1.405 | iter time: 360.27 ms (step) remaining time: 0:19:30
239
+ Epoch 1 | iter 4224 step 132 | loss train: 1.359, val: 1.405 | iter time: 359.63 ms (step) remaining time: 0:19:18
240
+ Epoch 1 | iter 4256 step 133 | loss train: 1.337, val: 1.405 | iter time: 360.16 ms (step) remaining time: 0:19:07
241
+ Epoch 1 | iter 4288 step 134 | loss train: 1.426, val: 1.405 | iter time: 359.02 ms (step) remaining time: 0:18:56
242
+ Epoch 1 | iter 4320 step 135 | loss train: 1.286, val: 1.405 | iter time: 360.57 ms (step) remaining time: 0:18:44
243
+ Epoch 1 | iter 4352 step 136 | loss train: 1.323, val: 1.405 | iter time: 358.27 ms (step) remaining time: 0:18:33
244
+ Epoch 1 | iter 4384 step 137 | loss train: 1.329, val: 1.405 | iter time: 359.05 ms (step) remaining time: 0:18:22
245
+ Epoch 1 | iter 4416 step 138 | loss train: 1.333, val: 1.405 | iter time: 359.39 ms (step) remaining time: 0:18:10
246
+ Epoch 1 | iter 4448 step 139 | loss train: 1.379, val: 1.405 | iter time: 358.93 ms (step) remaining time: 0:17:59
247
+ Epoch 1 | iter 4480 step 140 | loss train: 1.415, val: 1.405 | iter time: 359.14 ms (step) remaining time: 0:17:48
248
+ Epoch 1 | iter 4512 step 141 | loss train: 1.342, val: 1.405 | iter time: 360.54 ms (step) remaining time: 0:17:36
249
+ Epoch 1 | iter 4544 step 142 | loss train: 1.388, val: 1.405 | iter time: 361.40 ms (step) remaining time: 0:17:25
250
+ Epoch 1 | iter 4576 step 143 | loss train: 1.349, val: 1.405 | iter time: 359.42 ms (step) remaining time: 0:17:14
251
+ Epoch 1 | iter 4608 step 144 | loss train: 1.347, val: 1.405 | iter time: 361.98 ms (step) remaining time: 0:17:03
252
+ Epoch 1 | iter 4640 step 145 | loss train: 1.391, val: 1.405 | iter time: 359.69 ms (step) remaining time: 0:16:52
253
+ Epoch 1 | iter 4672 step 146 | loss train: 1.371, val: 1.405 | iter time: 361.41 ms (step) remaining time: 0:16:40
254
+ Epoch 1 | iter 4704 step 147 | loss train: 1.404, val: 1.405 | iter time: 360.51 ms (step) remaining time: 0:16:29
255
+ Epoch 1 | iter 4736 step 148 | loss train: 1.366, val: 1.405 | iter time: 361.40 ms (step) remaining time: 0:16:18
256
+ Epoch 1 | iter 4768 step 149 | loss train: 1.365, val: 1.405 | iter time: 360.73 ms (step) remaining time: 0:16:06
257
+ Epoch 1 | iter 4800 step 150 | loss train: 1.466, val: 1.405 | iter time: 359.57 ms (step) remaining time: 0:15:55
258
+ Validating ...
259
+ iter 4800: val loss 1.3828, val time: 9254.19 ms
260
+ Epoch 1 | iter 4832 step 151 | loss train: 1.345, val: 1.383 | iter time: 360.62 ms (step) remaining time: 0:15:49
261
+ Epoch 1 | iter 4864 step 152 | loss train: 1.357, val: 1.383 | iter time: 358.48 ms (step) remaining time: 0:15:38
262
+ Epoch 1 | iter 4896 step 153 | loss train: 1.303, val: 1.383 | iter time: 359.80 ms (step) remaining time: 0:15:26
263
+ Epoch 1 | iter 4928 step 154 | loss train: 1.347, val: 1.383 | iter time: 360.01 ms (step) remaining time: 0:15:15
264
+ Epoch 1 | iter 4960 step 155 | loss train: 1.382, val: 1.383 | iter time: 358.21 ms (step) remaining time: 0:15:04
265
+ Epoch 1 | iter 4992 step 156 | loss train: 1.290, val: 1.383 | iter time: 361.00 ms (step) remaining time: 0:14:53
266
+ Epoch 1 | iter 5024 step 157 | loss train: 1.422, val: 1.383 | iter time: 360.10 ms (step) remaining time: 0:14:41
267
+ Epoch 1 | iter 5056 step 158 | loss train: 1.391, val: 1.383 | iter time: 361.31 ms (step) remaining time: 0:14:30
268
+ Epoch 1 | iter 5088 step 159 | loss train: 1.343, val: 1.383 | iter time: 359.19 ms (step) remaining time: 0:14:19
269
+ Epoch 1 | iter 5120 step 160 | loss train: 1.417, val: 1.383 | iter time: 357.94 ms (step) remaining time: 0:14:07
270
+ Epoch 1 | iter 5152 step 161 | loss train: 1.394, val: 1.383 | iter time: 358.74 ms (step) remaining time: 0:13:56
271
+ Epoch 1 | iter 5184 step 162 | loss train: 1.357, val: 1.383 | iter time: 358.74 ms (step) remaining time: 0:13:45
272
+ Epoch 1 | iter 5216 step 163 | loss train: 1.381, val: 1.383 | iter time: 359.33 ms (step) remaining time: 0:13:33
273
+ Epoch 1 | iter 5248 step 164 | loss train: 1.393, val: 1.383 | iter time: 359.34 ms (step) remaining time: 0:13:22
274
+ Epoch 1 | iter 5280 step 165 | loss train: 1.362, val: 1.383 | iter time: 360.92 ms (step) remaining time: 0:13:11
275
+ Epoch 1 | iter 5312 step 166 | loss train: 1.380, val: 1.383 | iter time: 359.72 ms (step) remaining time: 0:13:00
276
+ Epoch 1 | iter 5344 step 167 | loss train: 1.358, val: 1.383 | iter time: 359.68 ms (step) remaining time: 0:12:48
277
+ Epoch 1 | iter 5376 step 168 | loss train: 1.351, val: 1.383 | iter time: 361.31 ms (step) remaining time: 0:12:37
278
+ Epoch 1 | iter 5408 step 169 | loss train: 1.357, val: 1.383 | iter time: 359.44 ms (step) remaining time: 0:12:26
279
+ Epoch 1 | iter 5440 step 170 | loss train: 1.396, val: 1.383 | iter time: 358.26 ms (step) remaining time: 0:12:15
280
+ Epoch 1 | iter 5472 step 171 | loss train: 1.394, val: 1.383 | iter time: 359.21 ms (step) remaining time: 0:12:03
281
+ Epoch 1 | iter 5504 step 172 | loss train: 1.385, val: 1.383 | iter time: 360.31 ms (step) remaining time: 0:11:52
282
+ Epoch 1 | iter 5536 step 173 | loss train: 1.365, val: 1.383 | iter time: 361.09 ms (step) remaining time: 0:11:41
283
+ Epoch 1 | iter 5568 step 174 | loss train: 1.345, val: 1.383 | iter time: 359.66 ms (step) remaining time: 0:11:30
284
+ Epoch 1 | iter 5600 step 175 | loss train: 1.288, val: 1.383 | iter time: 360.00 ms (step) remaining time: 0:11:18
285
+ Epoch 1 | iter 5632 step 176 | loss train: 1.343, val: 1.383 | iter time: 360.03 ms (step) remaining time: 0:11:07
286
+ Epoch 1 | iter 5664 step 177 | loss train: 1.352, val: 1.383 | iter time: 359.72 ms (step) remaining time: 0:10:56
287
+ Epoch 1 | iter 5696 step 178 | loss train: 1.371, val: 1.383 | iter time: 361.38 ms (step) remaining time: 0:10:45
288
+ Epoch 1 | iter 5728 step 179 | loss train: 1.364, val: 1.383 | iter time: 358.27 ms (step) remaining time: 0:10:34
289
+ Epoch 1 | iter 5760 step 180 | loss train: 1.341, val: 1.383 | iter time: 359.06 ms (step) remaining time: 0:10:22
290
+ Epoch 1 | iter 5792 step 181 | loss train: 1.406, val: 1.383 | iter time: 357.99 ms (step) remaining time: 0:10:11
291
+ Epoch 1 | iter 5824 step 182 | loss train: 1.419, val: 1.383 | iter time: 358.49 ms (step) remaining time: 0:10:00
292
+ Epoch 1 | iter 5856 step 183 | loss train: 1.361, val: 1.383 | iter time: 360.05 ms (step) remaining time: 0:09:49
293
+ Epoch 1 | iter 5888 step 184 | loss train: 1.406, val: 1.383 | iter time: 359.90 ms (step) remaining time: 0:09:38
294
+ Epoch 1 | iter 5920 step 185 | loss train: 1.304, val: 1.383 | iter time: 359.92 ms (step) remaining time: 0:09:26
295
+ Epoch 1 | iter 5952 step 186 | loss train: 1.379, val: 1.383 | iter time: 359.36 ms (step) remaining time: 0:09:15
296
+ Epoch 1 | iter 5984 step 187 | loss train: 1.330, val: 1.383 | iter time: 360.07 ms (step) remaining time: 0:09:04
297
+ Epoch 1 | iter 6016 step 188 | loss train: 1.376, val: 1.383 | iter time: 360.45 ms (step) remaining time: 0:08:53
298
+ Epoch 1 | iter 6048 step 189 | loss train: 1.328, val: 1.383 | iter time: 360.58 ms (step) remaining time: 0:08:42
299
+ Epoch 1 | iter 6080 step 190 | loss train: 1.340, val: 1.383 | iter time: 359.85 ms (step) remaining time: 0:08:31
300
+ Epoch 1 | iter 6112 step 191 | loss train: 1.291, val: 1.383 | iter time: 360.37 ms (step) remaining time: 0:08:19
301
+ Epoch 1 | iter 6144 step 192 | loss train: 1.360, val: 1.383 | iter time: 359.17 ms (step) remaining time: 0:08:08
302
+ Epoch 1 | iter 6176 step 193 | loss train: 1.341, val: 1.383 | iter time: 357.75 ms (step) remaining time: 0:07:57
303
+ Epoch 1 | iter 6208 step 194 | loss train: 1.360, val: 1.383 | iter time: 360.25 ms (step) remaining time: 0:07:46
304
+ Epoch 1 | iter 6240 step 195 | loss train: 1.347, val: 1.383 | iter time: 358.13 ms (step) remaining time: 0:07:35
305
+ Epoch 1 | iter 6272 step 196 | loss train: 1.321, val: 1.383 | iter time: 360.99 ms (step) remaining time: 0:07:24
306
+ Epoch 1 | iter 6304 step 197 | loss train: 1.358, val: 1.383 | iter time: 359.30 ms (step) remaining time: 0:07:13
307
+ Epoch 1 | iter 6336 step 198 | loss train: 1.309, val: 1.383 | iter time: 359.97 ms (step) remaining time: 0:07:01
308
+ Epoch 1 | iter 6368 step 199 | loss train: 1.285, val: 1.383 | iter time: 362.01 ms (step) remaining time: 0:06:50
309
+ Epoch 1 | iter 6400 step 200 | loss train: 1.358, val: 1.383 | iter time: 360.60 ms (step) remaining time: 0:06:39
310
+ Validating ...
311
+ iter 6400: val loss 1.3643, val time: 9276.42 ms
312
+ Saving checkpoint to 'out/pretrain/2411/step-00000200/lit_model.pth'
313
+ Epoch 1 | iter 6432 step 201 | loss train: 1.301, val: 1.364 | iter time: 357.65 ms (step) remaining time: 0:06:33
314
+ Epoch 1 | iter 6464 step 202 | loss train: 1.383, val: 1.364 | iter time: 357.12 ms (step) remaining time: 0:06:21
315
+ Epoch 1 | iter 6496 step 203 | loss train: 1.345, val: 1.364 | iter time: 357.83 ms (step) remaining time: 0:06:10
316
+ Epoch 1 | iter 6528 step 204 | loss train: 1.373, val: 1.364 | iter time: 359.57 ms (step) remaining time: 0:05:59
317
+ Epoch 1 | iter 6560 step 205 | loss train: 1.372, val: 1.364 | iter time: 359.81 ms (step) remaining time: 0:05:47
318
+ Epoch 1 | iter 6592 step 206 | loss train: 1.298, val: 1.364 | iter time: 358.90 ms (step) remaining time: 0:05:36
319
+ Epoch 1 | iter 6624 step 207 | loss train: 1.378, val: 1.364 | iter time: 361.14 ms (step) remaining time: 0:05:25
320
+ Epoch 1 | iter 6656 step 208 | loss train: 1.347, val: 1.364 | iter time: 360.97 ms (step) remaining time: 0:05:14
321
+ Epoch 1 | iter 6688 step 209 | loss train: 1.376, val: 1.364 | iter time: 360.26 ms (step) remaining time: 0:05:02
322
+ Epoch 1 | iter 6720 step 210 | loss train: 1.289, val: 1.364 | iter time: 360.51 ms (step) remaining time: 0:04:51
323
+ Epoch 1 | iter 6752 step 211 | loss train: 1.302, val: 1.364 | iter time: 359.02 ms (step) remaining time: 0:04:40
324
+ Epoch 1 | iter 6784 step 212 | loss train: 1.287, val: 1.364 | iter time: 358.60 ms (step) remaining time: 0:04:29
325
+ Epoch 1 | iter 6816 step 213 | loss train: 1.340, val: 1.364 | iter time: 361.35 ms (step) remaining time: 0:04:17
326
+ Epoch 1 | iter 6848 step 214 | loss train: 1.362, val: 1.364 | iter time: 360.09 ms (step) remaining time: 0:04:06
327
+ Epoch 1 | iter 6880 step 215 | loss train: 1.334, val: 1.364 | iter time: 359.07 ms (step) remaining time: 0:03:55
328
+ Epoch 1 | iter 6912 step 216 | loss train: 1.293, val: 1.364 | iter time: 358.31 ms (step) remaining time: 0:03:44
329
+ Epoch 1 | iter 6944 step 217 | loss train: 1.274, val: 1.364 | iter time: 359.84 ms (step) remaining time: 0:03:32
330
+ Epoch 1 | iter 6976 step 218 | loss train: 1.331, val: 1.364 | iter time: 360.24 ms (step) remaining time: 0:03:21
331
+ Epoch 1 | iter 7008 step 219 | loss train: 1.396, val: 1.364 | iter time: 359.51 ms (step) remaining time: 0:03:10
332
+ Epoch 1 | iter 7040 step 220 | loss train: 1.305, val: 1.364 | iter time: 359.02 ms (step) remaining time: 0:02:59
333
+ Epoch 1 | iter 7072 step 221 | loss train: 1.274, val: 1.364 | iter time: 360.91 ms (step) remaining time: 0:02:47
334
+ Epoch 1 | iter 7104 step 222 | loss train: 1.315, val: 1.364 | iter time: 360.72 ms (step) remaining time: 0:02:36
335
+ Epoch 1 | iter 7136 step 223 | loss train: 1.319, val: 1.364 | iter time: 359.08 ms (step) remaining time: 0:02:25
336
+ Epoch 1 | iter 7168 step 224 | loss train: 1.274, val: 1.364 | iter time: 358.49 ms (step) remaining time: 0:02:14
337
+ Epoch 1 | iter 7200 step 225 | loss train: 1.383, val: 1.364 | iter time: 358.98 ms (step) remaining time: 0:02:03
338
+ Epoch 1 | iter 7232 step 226 | loss train: 1.332, val: 1.364 | iter time: 361.54 ms (step) remaining time: 0:01:51
339
+ Epoch 1 | iter 7264 step 227 | loss train: 1.358, val: 1.364 | iter time: 359.08 ms (step) remaining time: 0:01:40
340
+ Epoch 1 | iter 7296 step 228 | loss train: 1.330, val: 1.364 | iter time: 359.41 ms (step) remaining time: 0:01:29
341
+ Epoch 1 | iter 7328 step 229 | loss train: 1.388, val: 1.364 | iter time: 360.89 ms (step) remaining time: 0:01:18
342
+ Epoch 1 | iter 7360 step 230 | loss train: 1.324, val: 1.364 | iter time: 359.52 ms (step) remaining time: 0:01:07
343
+ Epoch 1 | iter 7392 step 231 | loss train: 1.350, val: 1.364 | iter time: 360.05 ms (step) remaining time: 0:00:55
344
+ Epoch 1 | iter 7424 step 232 | loss train: 1.282, val: 1.364 | iter time: 360.44 ms (step) remaining time: 0:00:44
345
+ Epoch 1 | iter 7456 step 233 | loss train: 1.389, val: 1.364 | iter time: 358.88 ms (step) remaining time: 0:00:33
346
+ Epoch 1 | iter 7488 step 234 | loss train: 1.346, val: 1.364 | iter time: 358.37 ms (step) remaining time: 0:00:22
347
+ Epoch 1 | iter 7520 step 235 | loss train: 1.326, val: 1.364 | iter time: 359.94 ms (step) remaining time: 0:00:11
348
+ Epoch 1 | iter 7552 step 236 | loss train: 1.333, val: 1.364 | iter time: 360.17 ms (step) remaining time: 0:00:00
349
+ Validating ...
350
+ Final evaluation | val loss: 1.357 | val ppl: 3.883
351
+ Saving checkpoint to 'out/pretrain/2411/final/lit_model.pth'
352
+ ----------------------------------------
353
+ | Performance
354
+ | - Total tokens : 247,463,936
355
+ | - Training Time : 2675.12 s
356
+ | - Tok/sec : 221.25 tok/s
357
+ | ----------------------------------------
358
+ | Memory Usage
359
+ | - Memory Used : 26.32 GB
360
+ ----------------------------------------
out/pretrain/tinyllama/teelogs/2411_full.txt ADDED
@@ -0,0 +1,372 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Initializing distributed: GLOBAL_RANK: 0, MEMBER: 1/4
2
+ Initializing distributed: GLOBAL_RANK: 2, MEMBER: 3/4
3
+ [rank: 2] Seed set to 42
4
+ Initializing distributed: GLOBAL_RANK: 3, MEMBER: 4/4
5
+ [rank: 3] Seed set to 42
6
+ Initializing distributed: GLOBAL_RANK: 1, MEMBER: 2/4
7
+ ----------------------------------------------------------------------------------------------------
8
+ distributed_backend=nccl
9
+ All distributed processes registered. Starting with 4 processes
10
+ ----------------------------------------------------------------------------------------------------
11
+
12
+ [rank: 1] Seed set to 42
13
+ [rank: 0] Seed set to 42
14
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
15
+ train_dataloader = data.train_dataloader()
16
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
17
+ train_dataloader = data.train_dataloader()
18
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
19
+ train_dataloader = data.train_dataloader()
20
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
21
+ train_dataloader = data.train_dataloader()
22
+ All GPUs are fully connected via NVLink.
23
+ {'data': {'batch_size': 1,
24
+ 'data_path': PosixPath('litgpt/data/arxiv'),
25
+ 'num_workers': 8,
26
+ 'seed': 42,
27
+ 'seq_length': 2048,
28
+ 'use_starcoder': True},
29
+ 'data_dir': PosixPath('litgpt/data/arxiv/2411'),
30
+ 'devices': 'auto',
31
+ 'eval': {'evaluate_example': 'first',
32
+ 'final_validation': True,
33
+ 'initial_validation': True,
34
+ 'interval': 50,
35
+ 'max_iters': 200,
36
+ 'max_new_tokens': None},
37
+ 'initial_checkpoint_dir': PosixPath('out/pretrain/2410_full/final'),
38
+ 'log': {'group': None, 'project': None, 'run': None},
39
+ 'logger_name': 'tensorboard',
40
+ 'model_config': {'attention_logit_softcapping': None,
41
+ 'attention_scores_scalar': None,
42
+ 'attn_bias': False,
43
+ 'bias': False,
44
+ 'block_size': 2048,
45
+ 'final_logit_softcapping': None,
46
+ 'gelu_approximate': 'none',
47
+ 'head_size': 64,
48
+ 'hf_config': {'name': 'TinyLlama-1.1B-intermediate-step-1431k-3T',
49
+ 'org': 'TinyLlama'},
50
+ 'intermediate_size': 5632,
51
+ 'lm_head_bias': False,
52
+ 'mlp_class_name': 'LLaMAMLP',
53
+ 'moe_intermediate_size': None,
54
+ 'n_embd': 2048,
55
+ 'n_expert': 0,
56
+ 'n_expert_per_token': 0,
57
+ 'n_head': 32,
58
+ 'n_layer': 22,
59
+ 'n_query_groups': 4,
60
+ 'name': 'tiny-llama-1.1b',
61
+ 'norm_1': True,
62
+ 'norm_2': True,
63
+ 'norm_class_name': 'RMSNorm',
64
+ 'norm_eps': 1e-05,
65
+ 'norm_qk': False,
66
+ 'norm_qk_type': 'default',
67
+ 'padded_vocab_size': 32000,
68
+ 'padding_multiple': 64,
69
+ 'parallel_residual': False,
70
+ 'post_attention_norm': False,
71
+ 'post_mlp_norm': False,
72
+ 'rope_adjustments': None,
73
+ 'rope_base': 10000,
74
+ 'rope_condense_ratio': 1,
75
+ 'rope_indices': None,
76
+ 'rope_local_base_freq': None,
77
+ 'rotary_percentage': 1.0,
78
+ 'scale_embeddings': False,
79
+ 'shared_attention_norm': False,
80
+ 'sliding_window_indices': None,
81
+ 'sliding_window_size': None,
82
+ 'vocab_size': 32000},
83
+ 'model_name': 'tiny-llama-1.1b',
84
+ 'num_nodes': 1,
85
+ 'optimizer': "{'class_path': 'torch.optim.AdamW', 'init_args': {'lr': 0.0004, "
86
+ "'weight_decay': 0.1, 'betas': [0.9, 0.95]}}",
87
+ 'out_dir': PosixPath('out/pretrain/2411_full'),
88
+ 'ppl': False,
89
+ 'precision': 'bf16-mixed',
90
+ 'resume': False,
91
+ 'seed': 42,
92
+ 'tokenizer_dir': PosixPath('checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T'),
93
+ 'train': {'epochs': None,
94
+ 'global_batch_size': 512,
95
+ 'log_interval': 1,
96
+ 'lr_warmup_fraction': None,
97
+ 'lr_warmup_steps': 20,
98
+ 'max_norm': 1.0,
99
+ 'max_seq_length': 2048,
100
+ 'max_steps': None,
101
+ 'max_tokens': 250609664,
102
+ 'micro_batch_size': 4,
103
+ 'min_lr': 4e-05,
104
+ 'save_interval': 100,
105
+ 'tie_embeddings': None}}
106
+ Time to instantiate model: 0.04 seconds.
107
+ Total parameters: 1,100,048,384
108
+ [fix] out/pretrain/2410_full/final/lit_model.pth 是整包 state,提取 model 权重并原子覆盖...
109
+ [fix] 已覆盖为纯权重: out/pretrain/2410_full/final/lit_model.pth
110
+ Validating ...
111
+ Measured TFLOPs: 239.66
112
+ Epoch 1 | iter 32 step 1 | loss train: 1.385, val: 1.312 | iter time: 536.05 ms (step) remaining time: 0:47:19
113
+ Epoch 1 | iter 64 step 2 | loss train: 1.337, val: 1.312 | iter time: 358.46 ms (step) remaining time: 0:44:50
114
+ Epoch 1 | iter 96 step 3 | loss train: 1.356, val: 1.312 | iter time: 358.56 ms (step) remaining time: 0:43:54
115
+ Epoch 1 | iter 128 step 4 | loss train: 1.369, val: 1.312 | iter time: 357.84 ms (step) remaining time: 0:43:21
116
+ Epoch 1 | iter 160 step 5 | loss train: 1.488, val: 1.312 | iter time: 357.15 ms (step) remaining time: 0:42:57
117
+ Epoch 1 | iter 192 step 6 | loss train: 1.451, val: 1.312 | iter time: 357.91 ms (step) remaining time: 0:42:38
118
+ Epoch 1 | iter 224 step 7 | loss train: 1.325, val: 1.312 | iter time: 360.05 ms (step) remaining time: 0:42:23
119
+ Epoch 1 | iter 256 step 8 | loss train: 1.463, val: 1.312 | iter time: 359.24 ms (step) remaining time: 0:42:08
120
+ Epoch 1 | iter 288 step 9 | loss train: 1.442, val: 1.312 | iter time: 360.24 ms (step) remaining time: 0:41:55
121
+ Epoch 1 | iter 320 step 10 | loss train: 1.406, val: 1.312 | iter time: 359.20 ms (step) remaining time: 0:41:42
122
+ Epoch 1 | iter 352 step 11 | loss train: 1.415, val: 1.312 | iter time: 360.21 ms (step) remaining time: 0:41:29
123
+ Epoch 1 | iter 384 step 12 | loss train: 1.419, val: 1.312 | iter time: 359.54 ms (step) remaining time: 0:41:17
124
+ Epoch 1 | iter 416 step 13 | loss train: 1.382, val: 1.312 | iter time: 361.15 ms (step) remaining time: 0:41:05
125
+ Epoch 1 | iter 448 step 14 | loss train: 1.473, val: 1.312 | iter time: 360.67 ms (step) remaining time: 0:40:53
126
+ Epoch 1 | iter 480 step 15 | loss train: 1.360, val: 1.312 | iter time: 359.17 ms (step) remaining time: 0:40:41
127
+ Epoch 1 | iter 512 step 16 | loss train: 1.460, val: 1.312 | iter time: 361.14 ms (step) remaining time: 0:40:30
128
+ Epoch 1 | iter 544 step 17 | loss train: 1.422, val: 1.312 | iter time: 358.70 ms (step) remaining time: 0:40:18
129
+ Epoch 1 | iter 576 step 18 | loss train: 1.353, val: 1.312 | iter time: 359.36 ms (step) remaining time: 0:40:07
130
+ Epoch 1 | iter 608 step 19 | loss train: 1.420, val: 1.312 | iter time: 360.97 ms (step) remaining time: 0:39:56
131
+ Epoch 1 | iter 640 step 20 | loss train: 1.477, val: 1.312 | iter time: 359.06 ms (step) remaining time: 0:39:44
132
+ Epoch 1 | iter 672 step 21 | loss train: 1.434, val: 1.312 | iter time: 359.01 ms (step) remaining time: 0:39:33
133
+ Epoch 1 | iter 704 step 22 | loss train: 1.402, val: 1.312 | iter time: 361.25 ms (step) remaining time: 0:39:22
134
+ Epoch 1 | iter 736 step 23 | loss train: 1.421, val: 1.312 | iter time: 357.72 ms (step) remaining time: 0:39:11
135
+ Epoch 1 | iter 768 step 24 | loss train: 1.394, val: 1.312 | iter time: 359.50 ms (step) remaining time: 0:39:00
136
+ Epoch 1 | iter 800 step 25 | loss train: 1.428, val: 1.312 | iter time: 359.04 ms (step) remaining time: 0:38:49
137
+ Epoch 1 | iter 832 step 26 | loss train: 1.477, val: 1.312 | iter time: 358.50 ms (step) remaining time: 0:38:37
138
+ Epoch 1 | iter 864 step 27 | loss train: 1.431, val: 1.312 | iter time: 360.47 ms (step) remaining time: 0:38:26
139
+ Epoch 1 | iter 896 step 28 | loss train: 1.452, val: 1.312 | iter time: 360.67 ms (step) remaining time: 0:38:15
140
+ Epoch 1 | iter 928 step 29 | loss train: 1.439, val: 1.312 | iter time: 359.16 ms (step) remaining time: 0:38:04
141
+ Epoch 1 | iter 960 step 30 | loss train: 1.332, val: 1.312 | iter time: 359.50 ms (step) remaining time: 0:37:53
142
+ Epoch 1 | iter 992 step 31 | loss train: 1.444, val: 1.312 | iter time: 360.83 ms (step) remaining time: 0:37:42
143
+ Epoch 1 | iter 1024 step 32 | loss train: 1.436, val: 1.312 | iter time: 359.66 ms (step) remaining time: 0:37:31
144
+ Epoch 1 | iter 1056 step 33 | loss train: 1.353, val: 1.312 | iter time: 359.79 ms (step) remaining time: 0:37:20
145
+ Epoch 1 | iter 1088 step 34 | loss train: 1.349, val: 1.312 | iter time: 358.40 ms (step) remaining time: 0:37:09
146
+ Epoch 1 | iter 1120 step 35 | loss train: 1.492, val: 1.312 | iter time: 361.49 ms (step) remaining time: 0:36:58
147
+ Epoch 1 | iter 1152 step 36 | loss train: 1.406, val: 1.312 | iter time: 360.74 ms (step) remaining time: 0:36:48
148
+ Epoch 1 | iter 1184 step 37 | loss train: 1.444, val: 1.312 | iter time: 359.37 ms (step) remaining time: 0:36:37
149
+ Epoch 1 | iter 1216 step 38 | loss train: 1.421, val: 1.312 | iter time: 359.50 ms (step) remaining time: 0:36:26
150
+ Epoch 1 | iter 1248 step 39 | loss train: 1.443, val: 1.312 | iter time: 361.23 ms (step) remaining time: 0:36:15
151
+ Epoch 1 | iter 1280 step 40 | loss train: 1.412, val: 1.312 | iter time: 360.27 ms (step) remaining time: 0:36:04
152
+ Epoch 1 | iter 1312 step 41 | loss train: 1.448, val: 1.312 | iter time: 357.78 ms (step) remaining time: 0:35:53
153
+ Epoch 1 | iter 1344 step 42 | loss train: 1.471, val: 1.312 | iter time: 359.73 ms (step) remaining time: 0:35:43
154
+ Epoch 1 | iter 1376 step 43 | loss train: 1.332, val: 1.312 | iter time: 360.19 ms (step) remaining time: 0:35:32
155
+ Epoch 1 | iter 1408 step 44 | loss train: 1.476, val: 1.312 | iter time: 360.41 ms (step) remaining time: 0:35:21
156
+ Epoch 1 | iter 1440 step 45 | loss train: 1.401, val: 1.312 | iter time: 358.95 ms (step) remaining time: 0:35:10
157
+ Epoch 1 | iter 1472 step 46 | loss train: 1.456, val: 1.312 | iter time: 358.22 ms (step) remaining time: 0:34:59
158
+ Epoch 1 | iter 1504 step 47 | loss train: 1.435, val: 1.312 | iter time: 357.37 ms (step) remaining time: 0:34:48
159
+ Epoch 1 | iter 1536 step 48 | loss train: 1.385, val: 1.312 | iter time: 362.32 ms (step) remaining time: 0:34:37
160
+ Epoch 1 | iter 1568 step 49 | loss train: 1.370, val: 1.312 | iter time: 360.46 ms (step) remaining time: 0:34:26
161
+ Epoch 1 | iter 1600 step 50 | loss train: 1.382, val: 1.312 | iter time: 359.46 ms (step) remaining time: 0:34:15
162
+ Validating ...
163
+ iter 1600: val loss 1.2939, val time: 22370.16 ms
164
+ Epoch 1 | iter 1632 step 51 | loss train: 1.405, val: 1.294 | iter time: 360.53 ms (step) remaining time: 0:35:29
165
+ Epoch 1 | iter 1664 step 52 | loss train: 1.407, val: 1.294 | iter time: 359.34 ms (step) remaining time: 0:35:16
166
+ Epoch 1 | iter 1696 step 53 | loss train: 1.439, val: 1.294 | iter time: 358.07 ms (step) remaining time: 0:35:03
167
+ Epoch 1 | iter 1728 step 54 | loss train: 1.402, val: 1.294 | iter time: 359.39 ms (step) remaining time: 0:34:50
168
+ Epoch 1 | iter 1760 step 55 | loss train: 1.430, val: 1.294 | iter time: 361.28 ms (step) remaining time: 0:34:37
169
+ Epoch 1 | iter 1792 step 56 | loss train: 1.509, val: 1.294 | iter time: 358.62 ms (step) remaining time: 0:34:25
170
+ Epoch 1 | iter 1824 step 57 | loss train: 1.416, val: 1.294 | iter time: 358.34 ms (step) remaining time: 0:34:12
171
+ Epoch 1 | iter 1856 step 58 | loss train: 1.394, val: 1.294 | iter time: 358.73 ms (step) remaining time: 0:33:59
172
+ Epoch 1 | iter 1888 step 59 | loss train: 1.443, val: 1.294 | iter time: 360.27 ms (step) remaining time: 0:33:47
173
+ Epoch 1 | iter 1920 step 60 | loss train: 1.460, val: 1.294 | iter time: 360.60 ms (step) remaining time: 0:33:34
174
+ Epoch 1 | iter 1952 step 61 | loss train: 1.432, val: 1.294 | iter time: 359.26 ms (step) remaining time: 0:33:22
175
+ Epoch 1 | iter 1984 step 62 | loss train: 1.343, val: 1.294 | iter time: 358.09 ms (step) remaining time: 0:33:10
176
+ Epoch 1 | iter 2016 step 63 | loss train: 1.394, val: 1.294 | iter time: 358.02 ms (step) remaining time: 0:32:57
177
+ Epoch 1 | iter 2048 step 64 | loss train: 1.506, val: 1.294 | iter time: 359.22 ms (step) remaining time: 0:32:45
178
+ Epoch 1 | iter 2080 step 65 | loss train: 1.380, val: 1.294 | iter time: 359.89 ms (step) remaining time: 0:32:33
179
+ Epoch 1 | iter 2112 step 66 | loss train: 1.400, val: 1.294 | iter time: 360.79 ms (step) remaining time: 0:32:20
180
+ Epoch 1 | iter 2144 step 67 | loss train: 1.399, val: 1.294 | iter time: 358.23 ms (step) remaining time: 0:32:08
181
+ Epoch 1 | iter 2176 step 68 | loss train: 1.425, val: 1.294 | iter time: 360.82 ms (step) remaining time: 0:31:56
182
+ Epoch 1 | iter 2208 step 69 | loss train: 1.490, val: 1.294 | iter time: 358.42 ms (step) remaining time: 0:31:44
183
+ Epoch 1 | iter 2240 step 70 | loss train: 1.476, val: 1.294 | iter time: 359.49 ms (step) remaining time: 0:31:32
184
+ Epoch 1 | iter 2272 step 71 | loss train: 1.421, val: 1.294 | iter time: 360.85 ms (step) remaining time: 0:31:20
185
+ Epoch 1 | iter 2304 step 72 | loss train: 1.413, val: 1.294 | iter time: 360.18 ms (step) remaining time: 0:31:08
186
+ Epoch 1 | iter 2336 step 73 | loss train: 1.459, val: 1.294 | iter time: 360.26 ms (step) remaining time: 0:30:56
187
+ Epoch 1 | iter 2368 step 74 | loss train: 1.406, val: 1.294 | iter time: 360.15 ms (step) remaining time: 0:30:44
188
+ Epoch 1 | iter 2400 step 75 | loss train: 1.379, val: 1.294 | iter time: 358.32 ms (step) remaining time: 0:30:32
189
+ Epoch 1 | iter 2432 step 76 | loss train: 1.362, val: 1.294 | iter time: 359.50 ms (step) remaining time: 0:30:21
190
+ Epoch 1 | iter 2464 step 77 | loss train: 1.351, val: 1.294 | iter time: 359.54 ms (step) remaining time: 0:30:09
191
+ Epoch 1 | iter 2496 step 78 | loss train: 1.492, val: 1.294 | iter time: 359.40 ms (step) remaining time: 0:29:57
192
+ Epoch 1 | iter 2528 step 79 | loss train: 1.343, val: 1.294 | iter time: 359.06 ms (step) remaining time: 0:29:45
193
+ Epoch 1 | iter 2560 step 80 | loss train: 1.406, val: 1.294 | iter time: 358.65 ms (step) remaining time: 0:29:33
194
+ Epoch 1 | iter 2592 step 81 | loss train: 1.453, val: 1.294 | iter time: 358.60 ms (step) remaining time: 0:29:22
195
+ Epoch 1 | iter 2624 step 82 | loss train: 1.470, val: 1.294 | iter time: 358.12 ms (step) remaining time: 0:29:10
196
+ Epoch 1 | iter 2656 step 83 | loss train: 1.431, val: 1.294 | iter time: 358.82 ms (step) remaining time: 0:28:58
197
+ Epoch 1 | iter 2688 step 84 | loss train: 1.432, val: 1.294 | iter time: 359.61 ms (step) remaining time: 0:28:47
198
+ Epoch 1 | iter 2720 step 85 | loss train: 1.388, val: 1.294 | iter time: 361.41 ms (step) remaining time: 0:28:35
199
+ Epoch 1 | iter 2752 step 86 | loss train: 1.344, val: 1.294 | iter time: 360.90 ms (step) remaining time: 0:28:23
200
+ Epoch 1 | iter 2784 step 87 | loss train: 1.425, val: 1.294 | iter time: 359.16 ms (step) remaining time: 0:28:12
201
+ Epoch 1 | iter 2816 step 88 | loss train: 1.422, val: 1.294 | iter time: 359.69 ms (step) remaining time: 0:28:00
202
+ Epoch 1 | iter 2848 step 89 | loss train: 1.388, val: 1.294 | iter time: 359.55 ms (step) remaining time: 0:27:48
203
+ Epoch 1 | iter 2880 step 90 | loss train: 1.452, val: 1.294 | iter time: 359.84 ms (step) remaining time: 0:27:37
204
+ Epoch 1 | iter 2912 step 91 | loss train: 1.422, val: 1.294 | iter time: 360.12 ms (step) remaining time: 0:27:25
205
+ Epoch 1 | iter 2944 step 92 | loss train: 1.318, val: 1.294 | iter time: 360.34 ms (step) remaining time: 0:27:14
206
+ Epoch 1 | iter 2976 step 93 | loss train: 1.452, val: 1.294 | iter time: 359.09 ms (step) remaining time: 0:27:02
207
+ Epoch 1 | iter 3008 step 94 | loss train: 1.451, val: 1.294 | iter time: 359.24 ms (step) remaining time: 0:26:51
208
+ Epoch 1 | iter 3040 step 95 | loss train: 1.450, val: 1.294 | iter time: 361.68 ms (step) remaining time: 0:26:39
209
+ Epoch 1 | iter 3072 step 96 | loss train: 1.411, val: 1.294 | iter time: 358.21 ms (step) remaining time: 0:26:28
210
+ Epoch 1 | iter 3104 step 97 | loss train: 1.410, val: 1.294 | iter time: 359.25 ms (step) remaining time: 0:26:16
211
+ Epoch 1 | iter 3136 step 98 | loss train: 1.428, val: 1.294 | iter time: 357.92 ms (step) remaining time: 0:26:05
212
+ Epoch 1 | iter 3168 step 99 | loss train: 1.397, val: 1.294 | iter time: 362.02 ms (step) remaining time: 0:25:53
213
+ Epoch 1 | iter 3200 step 100 | loss train: 1.359, val: 1.294 | iter time: 359.59 ms (step) remaining time: 0:25:42
214
+ Validating ...
215
+ iter 3200: val loss 1.2439, val time: 22378.01 ms
216
+ Saving checkpoint to 'out/pretrain/2411_full/step-00000100/lit_model.pth'
217
+ Epoch 1 | iter 3232 step 101 | loss train: 1.438, val: 1.244 | iter time: 358.59 ms (step) remaining time: 0:26:24
218
+ Epoch 1 | iter 3264 step 102 | loss train: 1.414, val: 1.244 | iter time: 359.90 ms (step) remaining time: 0:26:12
219
+ Epoch 1 | iter 3296 step 103 | loss train: 1.393, val: 1.244 | iter time: 358.44 ms (step) remaining time: 0:25:59
220
+ Epoch 1 | iter 3328 step 104 | loss train: 1.375, val: 1.244 | iter time: 360.59 ms (step) remaining time: 0:25:47
221
+ Epoch 1 | iter 3360 step 105 | loss train: 1.360, val: 1.244 | iter time: 359.67 ms (step) remaining time: 0:25:35
222
+ Epoch 1 | iter 3392 step 106 | loss train: 1.415, val: 1.244 | iter time: 358.96 ms (step) remaining time: 0:25:23
223
+ Epoch 1 | iter 3424 step 107 | loss train: 1.429, val: 1.244 | iter time: 359.56 ms (step) remaining time: 0:25:11
224
+ Epoch 1 | iter 3456 step 108 | loss train: 1.394, val: 1.244 | iter time: 358.84 ms (step) remaining time: 0:24:58
225
+ Epoch 1 | iter 3488 step 109 | loss train: 1.455, val: 1.244 | iter time: 359.08 ms (step) remaining time: 0:24:46
226
+ Epoch 1 | iter 3520 step 110 | loss train: 1.388, val: 1.244 | iter time: 359.00 ms (step) remaining time: 0:24:34
227
+ Epoch 1 | iter 3552 step 111 | loss train: 1.333, val: 1.244 | iter time: 358.53 ms (step) remaining time: 0:24:22
228
+ Epoch 1 | iter 3584 step 112 | loss train: 1.359, val: 1.244 | iter time: 360.46 ms (step) remaining time: 0:24:10
229
+ Epoch 1 | iter 3616 step 113 | loss train: 1.384, val: 1.244 | iter time: 359.24 ms (step) remaining time: 0:23:58
230
+ Epoch 1 | iter 3648 step 114 | loss train: 1.412, val: 1.244 | iter time: 360.66 ms (step) remaining time: 0:23:46
231
+ Epoch 1 | iter 3680 step 115 | loss train: 1.415, val: 1.244 | iter time: 361.23 ms (step) remaining time: 0:23:34
232
+ Epoch 1 | iter 3712 step 116 | loss train: 1.406, val: 1.244 | iter time: 359.68 ms (step) remaining time: 0:23:22
233
+ Epoch 1 | iter 3744 step 117 | loss train: 1.377, val: 1.244 | iter time: 361.09 ms (step) remaining time: 0:23:10
234
+ Epoch 1 | iter 3776 step 118 | loss train: 1.355, val: 1.244 | iter time: 359.84 ms (step) remaining time: 0:22:59
235
+ Epoch 1 | iter 3808 step 119 | loss train: 1.388, val: 1.244 | iter time: 358.49 ms (step) remaining time: 0:22:47
236
+ Epoch 1 | iter 3840 step 120 | loss train: 1.387, val: 1.244 | iter time: 359.27 ms (step) remaining time: 0:22:35
237
+ Epoch 1 | iter 3872 step 121 | loss train: 1.385, val: 1.244 | iter time: 359.35 ms (step) remaining time: 0:22:23
238
+ Epoch 1 | iter 3904 step 122 | loss train: 1.405, val: 1.244 | iter time: 360.18 ms (step) remaining time: 0:22:11
239
+ Epoch 1 | iter 3936 step 123 | loss train: 1.448, val: 1.244 | iter time: 358.64 ms (step) remaining time: 0:21:59
240
+ Epoch 1 | iter 3968 step 124 | loss train: 1.338, val: 1.244 | iter time: 360.89 ms (step) remaining time: 0:21:47
241
+ Epoch 1 | iter 4000 step 125 | loss train: 1.384, val: 1.244 | iter time: 358.47 ms (step) remaining time: 0:21:35
242
+ Epoch 1 | iter 4032 step 126 | loss train: 1.360, val: 1.244 | iter time: 358.89 ms (step) remaining time: 0:21:23
243
+ Epoch 1 | iter 4064 step 127 | loss train: 1.383, val: 1.244 | iter time: 361.35 ms (step) remaining time: 0:21:12
244
+ Epoch 1 | iter 4096 step 128 | loss train: 1.347, val: 1.244 | iter time: 361.50 ms (step) remaining time: 0:21:00
245
+ Epoch 1 | iter 4128 step 129 | loss train: 1.382, val: 1.244 | iter time: 358.77 ms (step) remaining time: 0:20:48
246
+ Epoch 1 | iter 4160 step 130 | loss train: 1.412, val: 1.244 | iter time: 358.04 ms (step) remaining time: 0:20:36
247
+ Epoch 1 | iter 4192 step 131 | loss train: 1.340, val: 1.244 | iter time: 359.86 ms (step) remaining time: 0:20:25
248
+ Epoch 1 | iter 4224 step 132 | loss train: 1.360, val: 1.244 | iter time: 358.69 ms (step) remaining time: 0:20:13
249
+ Epoch 1 | iter 4256 step 133 | loss train: 1.337, val: 1.244 | iter time: 359.84 ms (step) remaining time: 0:20:01
250
+ Epoch 1 | iter 4288 step 134 | loss train: 1.427, val: 1.244 | iter time: 357.63 ms (step) remaining time: 0:19:49
251
+ Epoch 1 | iter 4320 step 135 | loss train: 1.287, val: 1.244 | iter time: 358.42 ms (step) remaining time: 0:19:38
252
+ Epoch 1 | iter 4352 step 136 | loss train: 1.322, val: 1.244 | iter time: 360.74 ms (step) remaining time: 0:19:26
253
+ Epoch 1 | iter 4384 step 137 | loss train: 1.330, val: 1.244 | iter time: 360.09 ms (step) remaining time: 0:19:14
254
+ Epoch 1 | iter 4416 step 138 | loss train: 1.333, val: 1.244 | iter time: 359.74 ms (step) remaining time: 0:19:03
255
+ Epoch 1 | iter 4448 step 139 | loss train: 1.379, val: 1.244 | iter time: 358.36 ms (step) remaining time: 0:18:51
256
+ Epoch 1 | iter 4480 step 140 | loss train: 1.417, val: 1.244 | iter time: 360.26 ms (step) remaining time: 0:18:39
257
+ Epoch 1 | iter 4512 step 141 | loss train: 1.342, val: 1.244 | iter time: 359.52 ms (step) remaining time: 0:18:28
258
+ Epoch 1 | iter 4544 step 142 | loss train: 1.389, val: 1.244 | iter time: 360.43 ms (step) remaining time: 0:18:16
259
+ Epoch 1 | iter 4576 step 143 | loss train: 1.347, val: 1.244 | iter time: 360.40 ms (step) remaining time: 0:18:04
260
+ Epoch 1 | iter 4608 step 144 | loss train: 1.347, val: 1.244 | iter time: 360.14 ms (step) remaining time: 0:17:53
261
+ Epoch 1 | iter 4640 step 145 | loss train: 1.392, val: 1.244 | iter time: 358.88 ms (step) remaining time: 0:17:41
262
+ Epoch 1 | iter 4672 step 146 | loss train: 1.371, val: 1.244 | iter time: 361.56 ms (step) remaining time: 0:17:30
263
+ Epoch 1 | iter 4704 step 147 | loss train: 1.404, val: 1.244 | iter time: 360.01 ms (step) remaining time: 0:17:18
264
+ Epoch 1 | iter 4736 step 148 | loss train: 1.366, val: 1.244 | iter time: 358.92 ms (step) remaining time: 0:17:07
265
+ Epoch 1 | iter 4768 step 149 | loss train: 1.365, val: 1.244 | iter time: 359.66 ms (step) remaining time: 0:16:55
266
+ Epoch 1 | iter 4800 step 150 | loss train: 1.466, val: 1.244 | iter time: 358.44 ms (step) remaining time: 0:16:44
267
+ Validating ...
268
+ iter 4800: val loss 1.1958, val time: 22369.24 ms
269
+ Epoch 1 | iter 4832 step 151 | loss train: 1.345, val: 1.196 | iter time: 359.84 ms (step) remaining time: 0:16:45
270
+ Epoch 1 | iter 4864 step 152 | loss train: 1.358, val: 1.196 | iter time: 359.76 ms (step) remaining time: 0:16:33
271
+ Epoch 1 | iter 4896 step 153 | loss train: 1.304, val: 1.196 | iter time: 358.12 ms (step) remaining time: 0:16:22
272
+ Epoch 1 | iter 4928 step 154 | loss train: 1.348, val: 1.196 | iter time: 358.95 ms (step) remaining time: 0:16:10
273
+ Epoch 1 | iter 4960 step 155 | loss train: 1.382, val: 1.196 | iter time: 359.76 ms (step) remaining time: 0:15:58
274
+ Epoch 1 | iter 4992 step 156 | loss train: 1.290, val: 1.196 | iter time: 360.15 ms (step) remaining time: 0:15:46
275
+ Epoch 1 | iter 5024 step 157 | loss train: 1.422, val: 1.196 | iter time: 358.70 ms (step) remaining time: 0:15:35
276
+ Epoch 1 | iter 5056 step 158 | loss train: 1.391, val: 1.196 | iter time: 359.50 ms (step) remaining time: 0:15:23
277
+ Epoch 1 | iter 5088 step 159 | loss train: 1.343, val: 1.196 | iter time: 358.65 ms (step) remaining time: 0:15:11
278
+ Epoch 1 | iter 5120 step 160 | loss train: 1.419, val: 1.196 | iter time: 360.49 ms (step) remaining time: 0:15:00
279
+ Epoch 1 | iter 5152 step 161 | loss train: 1.394, val: 1.196 | iter time: 359.69 ms (step) remaining time: 0:14:48
280
+ Epoch 1 | iter 5184 step 162 | loss train: 1.358, val: 1.196 | iter time: 360.02 ms (step) remaining time: 0:14:36
281
+ Epoch 1 | iter 5216 step 163 | loss train: 1.382, val: 1.196 | iter time: 360.12 ms (step) remaining time: 0:14:25
282
+ Epoch 1 | iter 5248 step 164 | loss train: 1.395, val: 1.196 | iter time: 358.85 ms (step) remaining time: 0:14:13
283
+ Epoch 1 | iter 5280 step 165 | loss train: 1.364, val: 1.196 | iter time: 359.12 ms (step) remaining time: 0:14:01
284
+ Epoch 1 | iter 5312 step 166 | loss train: 1.379, val: 1.196 | iter time: 360.38 ms (step) remaining time: 0:13:50
285
+ Epoch 1 | iter 5344 step 167 | loss train: 1.358, val: 1.196 | iter time: 358.86 ms (step) remaining time: 0:13:38
286
+ Epoch 1 | iter 5376 step 168 | loss train: 1.351, val: 1.196 | iter time: 360.70 ms (step) remaining time: 0:13:27
287
+ Epoch 1 | iter 5408 step 169 | loss train: 1.358, val: 1.196 | iter time: 359.88 ms (step) remaining time: 0:13:15
288
+ Epoch 1 | iter 5440 step 170 | loss train: 1.396, val: 1.196 | iter time: 358.87 ms (step) remaining time: 0:13:04
289
+ Epoch 1 | iter 5472 step 171 | loss train: 1.395, val: 1.196 | iter time: 358.15 ms (step) remaining time: 0:12:52
290
+ Epoch 1 | iter 5504 step 172 | loss train: 1.385, val: 1.196 | iter time: 358.52 ms (step) remaining time: 0:12:41
291
+ Epoch 1 | iter 5536 step 173 | loss train: 1.366, val: 1.196 | iter time: 358.65 ms (step) remaining time: 0:12:29
292
+ Epoch 1 | iter 5568 step 174 | loss train: 1.345, val: 1.196 | iter time: 359.53 ms (step) remaining time: 0:12:17
293
+ Epoch 1 | iter 5600 step 175 | loss train: 1.289, val: 1.196 | iter time: 358.53 ms (step) remaining time: 0:12:06
294
+ Epoch 1 | iter 5632 step 176 | loss train: 1.343, val: 1.196 | iter time: 359.26 ms (step) remaining time: 0:11:54
295
+ Epoch 1 | iter 5664 step 177 | loss train: 1.352, val: 1.196 | iter time: 361.21 ms (step) remaining time: 0:11:43
296
+ Epoch 1 | iter 5696 step 178 | loss train: 1.371, val: 1.196 | iter time: 358.45 ms (step) remaining time: 0:11:31
297
+ Epoch 1 | iter 5728 step 179 | loss train: 1.365, val: 1.196 | iter time: 579.10 ms (step) remaining time: 0:11:20
298
+ Epoch 1 | iter 5760 step 180 | loss train: 1.342, val: 1.196 | iter time: 358.31 ms (step) remaining time: 0:11:08
299
+ Epoch 1 | iter 5792 step 181 | loss train: 1.407, val: 1.196 | iter time: 358.42 ms (step) remaining time: 0:10:57
300
+ Epoch 1 | iter 5824 step 182 | loss train: 1.420, val: 1.196 | iter time: 359.56 ms (step) remaining time: 0:10:45
301
+ Epoch 1 | iter 5856 step 183 | loss train: 1.362, val: 1.196 | iter time: 359.50 ms (step) remaining time: 0:10:34
302
+ Epoch 1 | iter 5888 step 184 | loss train: 1.406, val: 1.196 | iter time: 359.14 ms (step) remaining time: 0:10:23
303
+ Epoch 1 | iter 5920 step 185 | loss train: 1.303, val: 1.196 | iter time: 359.29 ms (step) remaining time: 0:10:11
304
+ Epoch 1 | iter 5952 step 186 | loss train: 1.381, val: 1.196 | iter time: 360.81 ms (step) remaining time: 0:10:00
305
+ Epoch 1 | iter 5984 step 187 | loss train: 1.330, val: 1.196 | iter time: 358.61 ms (step) remaining time: 0:09:48
306
+ Epoch 1 | iter 6016 step 188 | loss train: 1.376, val: 1.196 | iter time: 359.75 ms (step) remaining time: 0:09:37
307
+ Epoch 1 | iter 6048 step 189 | loss train: 1.328, val: 1.196 | iter time: 359.61 ms (step) remaining time: 0:09:25
308
+ Epoch 1 | iter 6080 step 190 | loss train: 1.341, val: 1.196 | iter time: 359.27 ms (step) remaining time: 0:09:14
309
+ Epoch 1 | iter 6112 step 191 | loss train: 1.291, val: 1.196 | iter time: 360.10 ms (step) remaining time: 0:09:02
310
+ Epoch 1 | iter 6144 step 192 | loss train: 1.360, val: 1.196 | iter time: 357.85 ms (step) remaining time: 0:08:51
311
+ Epoch 1 | iter 6176 step 193 | loss train: 1.342, val: 1.196 | iter time: 358.16 ms (step) remaining time: 0:08:40
312
+ Epoch 1 | iter 6208 step 194 | loss train: 1.360, val: 1.196 | iter time: 357.86 ms (step) remaining time: 0:08:28
313
+ Epoch 1 | iter 6240 step 195 | loss train: 1.348, val: 1.196 | iter time: 359.78 ms (step) remaining time: 0:08:17
314
+ Epoch 1 | iter 6272 step 196 | loss train: 1.321, val: 1.196 | iter time: 358.91 ms (step) remaining time: 0:08:05
315
+ Epoch 1 | iter 6304 step 197 | loss train: 1.359, val: 1.196 | iter time: 359.99 ms (step) remaining time: 0:07:54
316
+ Epoch 1 | iter 6336 step 198 | loss train: 1.310, val: 1.196 | iter time: 359.90 ms (step) remaining time: 0:07:43
317
+ Epoch 1 | iter 6368 step 199 | loss train: 1.284, val: 1.196 | iter time: 359.14 ms (step) remaining time: 0:07:31
318
+ Epoch 1 | iter 6400 step 200 | loss train: 1.357, val: 1.196 | iter time: 360.49 ms (step) remaining time: 0:07:20
319
+ Validating ...
320
+ iter 6400: val loss 1.1577, val time: 22364.37 ms
321
+ Saving checkpoint to 'out/pretrain/2411_full/step-00000200/lit_model.pth'
322
+ Epoch 1 | iter 6432 step 201 | loss train: 1.301, val: 1.158 | iter time: 355.83 ms (step) remaining time: 0:07:16
323
+ Epoch 1 | iter 6464 step 202 | loss train: 1.383, val: 1.158 | iter time: 355.74 ms (step) remaining time: 0:07:04
324
+ Epoch 1 | iter 6496 step 203 | loss train: 1.345, val: 1.158 | iter time: 359.40 ms (step) remaining time: 0:06:53
325
+ Epoch 1 | iter 6528 step 204 | loss train: 1.374, val: 1.158 | iter time: 360.00 ms (step) remaining time: 0:06:41
326
+ Epoch 1 | iter 6560 step 205 | loss train: 1.372, val: 1.158 | iter time: 357.77 ms (step) remaining time: 0:06:29
327
+ Epoch 1 | iter 6592 step 206 | loss train: 1.299, val: 1.158 | iter time: 360.10 ms (step) remaining time: 0:06:18
328
+ Epoch 1 | iter 6624 step 207 | loss train: 1.378, val: 1.158 | iter time: 357.40 ms (step) remaining time: 0:06:06
329
+ Epoch 1 | iter 6656 step 208 | loss train: 1.348, val: 1.158 | iter time: 359.02 ms (step) remaining time: 0:05:55
330
+ Epoch 1 | iter 6688 step 209 | loss train: 1.377, val: 1.158 | iter time: 358.85 ms (step) remaining time: 0:05:43
331
+ Epoch 1 | iter 6720 step 210 | loss train: 1.288, val: 1.158 | iter time: 359.82 ms (step) remaining time: 0:05:32
332
+ Epoch 1 | iter 6752 step 211 | loss train: 1.302, val: 1.158 | iter time: 357.84 ms (step) remaining time: 0:05:20
333
+ Epoch 1 | iter 6784 step 212 | loss train: 1.287, val: 1.158 | iter time: 359.81 ms (step) remaining time: 0:05:09
334
+ Epoch 1 | iter 6816 step 213 | loss train: 1.339, val: 1.158 | iter time: 360.20 ms (step) remaining time: 0:04:57
335
+ Epoch 1 | iter 6848 step 214 | loss train: 1.361, val: 1.158 | iter time: 361.63 ms (step) remaining time: 0:04:46
336
+ Epoch 1 | iter 6880 step 215 | loss train: 1.334, val: 1.158 | iter time: 360.10 ms (step) remaining time: 0:04:34
337
+ Epoch 1 | iter 6912 step 216 | loss train: 1.294, val: 1.158 | iter time: 359.26 ms (step) remaining time: 0:04:23
338
+ Epoch 1 | iter 6944 step 217 | loss train: 1.272, val: 1.158 | iter time: 359.47 ms (step) remaining time: 0:04:11
339
+ Epoch 1 | iter 6976 step 218 | loss train: 1.331, val: 1.158 | iter time: 359.73 ms (step) remaining time: 0:04:00
340
+ Epoch 1 | iter 7008 step 219 | loss train: 1.395, val: 1.158 | iter time: 359.78 ms (step) remaining time: 0:03:48
341
+ Epoch 1 | iter 7040 step 220 | loss train: 1.306, val: 1.158 | iter time: 360.37 ms (step) remaining time: 0:03:37
342
+ Epoch 1 | iter 7072 step 221 | loss train: 1.274, val: 1.158 | iter time: 358.43 ms (step) remaining time: 0:03:25
343
+ Epoch 1 | iter 7104 step 222 | loss train: 1.315, val: 1.158 | iter time: 359.68 ms (step) remaining time: 0:03:14
344
+ Epoch 1 | iter 7136 step 223 | loss train: 1.319, val: 1.158 | iter time: 358.74 ms (step) remaining time: 0:03:02
345
+ Epoch 1 | iter 7168 step 224 | loss train: 1.274, val: 1.158 | iter time: 358.32 ms (step) remaining time: 0:02:51
346
+ Epoch 1 | iter 7200 step 225 | loss train: 1.382, val: 1.158 | iter time: 360.61 ms (step) remaining time: 0:02:39
347
+ Epoch 1 | iter 7232 step 226 | loss train: 1.332, val: 1.158 | iter time: 360.36 ms (step) remaining time: 0:02:28
348
+ Epoch 1 | iter 7264 step 227 | loss train: 1.357, val: 1.158 | iter time: 360.40 ms (step) remaining time: 0:02:16
349
+ Epoch 1 | iter 7296 step 228 | loss train: 1.331, val: 1.158 | iter time: 358.67 ms (step) remaining time: 0:02:05
350
+ Epoch 1 | iter 7328 step 229 | loss train: 1.388, val: 1.158 | iter time: 358.12 ms (step) remaining time: 0:01:54
351
+ Epoch 1 | iter 7360 step 230 | loss train: 1.323, val: 1.158 | iter time: 359.22 ms (step) remaining time: 0:01:42
352
+ Epoch 1 | iter 7392 step 231 | loss train: 1.350, val: 1.158 | iter time: 360.58 ms (step) remaining time: 0:01:31
353
+ Epoch 1 | iter 7424 step 232 | loss train: 1.281, val: 1.158 | iter time: 360.66 ms (step) remaining time: 0:01:19
354
+ Epoch 1 | iter 7456 step 233 | loss train: 1.389, val: 1.158 | iter time: 358.50 ms (step) remaining time: 0:01:08
355
+ Epoch 1 | iter 7488 step 234 | loss train: 1.345, val: 1.158 | iter time: 359.13 ms (step) remaining time: 0:00:56
356
+ Epoch 1 | iter 7520 step 235 | loss train: 1.326, val: 1.158 | iter time: 357.69 ms (step) remaining time: 0:00:45
357
+ Epoch 1 | iter 7552 step 236 | loss train: 1.332, val: 1.158 | iter time: 359.30 ms (step) remaining time: 0:00:34
358
+ Epoch 1 | iter 7584 step 237 | loss train: 1.343, val: 1.158 | iter time: 359.36 ms (step) remaining time: 0:00:22
359
+ Epoch 1 | iter 7616 step 238 | loss train: 1.297, val: 1.158 | iter time: 359.37 ms (step) remaining time: 0:00:11
360
+ Epoch 2 | iter 7648 step 239 | loss train: 1.201, val: 1.158 | iter time: 359.80 ms (step) remaining time: 0:00:00
361
+ Validating ...
362
+ Final evaluation | val loss: 1.132 | val ppl: 3.101
363
+ Saving checkpoint to 'out/pretrain/2411_full/final/lit_model.pth'
364
+ ----------------------------------------
365
+ | Performance
366
+ | - Total tokens : 250,609,664
367
+ | - Training Time : 2784.71 s
368
+ | - Tok/sec : 135.07 tok/s
369
+ | ----------------------------------------
370
+ | Memory Usage
371
+ | - Memory Used : 26.32 GB
372
+ ----------------------------------------
out/pretrain/tinyllama/teelogs/2411_lr4e-5.txt ADDED
@@ -0,0 +1,371 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Initializing distributed: GLOBAL_RANK: 0, MEMBER: 1/4
2
+ Initializing distributed: GLOBAL_RANK: 1, MEMBER: 2/4
3
+ [rank: 1] Seed set to 42
4
+ Initializing distributed: GLOBAL_RANK: 2, MEMBER: 3/4
5
+ [rank: 2] Seed set to 42
6
+ Initializing distributed: GLOBAL_RANK: 3, MEMBER: 4/4
7
+ [rank: 3] Seed set to 42
8
+ ----------------------------------------------------------------------------------------------------
9
+ distributed_backend=nccl
10
+ All distributed processes registered. Starting with 4 processes
11
+ ----------------------------------------------------------------------------------------------------
12
+
13
+ [rank: 0] Seed set to 42
14
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
15
+ train_dataloader = data.train_dataloader()
16
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
17
+ train_dataloader = data.train_dataloader()
18
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
19
+ train_dataloader = data.train_dataloader()
20
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
21
+ train_dataloader = data.train_dataloader()
22
+ All GPUs are fully connected via NVLink.
23
+ {'data': {'batch_size': 1,
24
+ 'data_path': PosixPath('litgpt/data/arxiv'),
25
+ 'num_workers': 0,
26
+ 'seed': 42,
27
+ 'seq_length': 2048,
28
+ 'use_starcoder': True},
29
+ 'data_dir': PosixPath('litgpt/data/arxiv/2411'),
30
+ 'devices': 'auto',
31
+ 'eval': {'evaluate_example': 'first',
32
+ 'final_validation': True,
33
+ 'initial_validation': True,
34
+ 'interval': 50,
35
+ 'max_iters': 200,
36
+ 'max_new_tokens': None},
37
+ 'initial_checkpoint_dir': PosixPath('out/pretrain/tinyllama/2410_full/final'),
38
+ 'log': {'group': None, 'project': None, 'run': None},
39
+ 'logger_name': 'tensorboard',
40
+ 'model_config': {'attention_logit_softcapping': None,
41
+ 'attention_scores_scalar': None,
42
+ 'attn_bias': False,
43
+ 'bias': False,
44
+ 'block_size': 2048,
45
+ 'final_logit_softcapping': None,
46
+ 'gelu_approximate': 'none',
47
+ 'head_size': 64,
48
+ 'hf_config': {'name': 'TinyLlama-1.1B-intermediate-step-1431k-3T',
49
+ 'org': 'TinyLlama'},
50
+ 'intermediate_size': 5632,
51
+ 'lm_head_bias': False,
52
+ 'mlp_class_name': 'LLaMAMLP',
53
+ 'moe_intermediate_size': None,
54
+ 'n_embd': 2048,
55
+ 'n_expert': 0,
56
+ 'n_expert_per_token': 0,
57
+ 'n_head': 32,
58
+ 'n_layer': 22,
59
+ 'n_query_groups': 4,
60
+ 'name': 'tiny-llama-1.1b',
61
+ 'norm_1': True,
62
+ 'norm_2': True,
63
+ 'norm_class_name': 'RMSNorm',
64
+ 'norm_eps': 1e-05,
65
+ 'norm_qk': False,
66
+ 'norm_qk_type': 'default',
67
+ 'padded_vocab_size': 32000,
68
+ 'padding_multiple': 64,
69
+ 'parallel_residual': False,
70
+ 'post_attention_norm': False,
71
+ 'post_mlp_norm': False,
72
+ 'rope_adjustments': None,
73
+ 'rope_base': 10000,
74
+ 'rope_condense_ratio': 1,
75
+ 'rope_indices': None,
76
+ 'rope_local_base_freq': None,
77
+ 'rotary_percentage': 1.0,
78
+ 'scale_embeddings': False,
79
+ 'shared_attention_norm': False,
80
+ 'sliding_window_indices': None,
81
+ 'sliding_window_size': None,
82
+ 'vocab_size': 32000},
83
+ 'model_name': 'tiny-llama-1.1b',
84
+ 'num_nodes': 1,
85
+ 'optimizer': "{'class_path': 'torch.optim.AdamW', 'init_args': {'lr': 4e-05, "
86
+ "'weight_decay': 0.1, 'betas': [0.9, 0.95]}}",
87
+ 'out_dir': PosixPath('out/pretrain/tinyllama/2411_lr4e-5'),
88
+ 'ppl': False,
89
+ 'precision': 'bf16-mixed',
90
+ 'resume': False,
91
+ 'seed': 42,
92
+ 'tokenizer_dir': PosixPath('checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T'),
93
+ 'train': {'epochs': None,
94
+ 'global_batch_size': 512,
95
+ 'log_interval': 1,
96
+ 'lr_warmup_fraction': None,
97
+ 'lr_warmup_steps': 20,
98
+ 'max_norm': 1.0,
99
+ 'max_seq_length': 2048,
100
+ 'max_steps': None,
101
+ 'max_tokens': 250609664,
102
+ 'micro_batch_size': 4,
103
+ 'min_lr': 4e-05,
104
+ 'save_interval': 100,
105
+ 'tie_embeddings': None}}
106
+ Time to instantiate model: 0.04 seconds.
107
+ Total parameters: 1,100,048,384
108
+ [ok] out/pretrain/tinyllama/2410_full/final/lit_model.pth 已是纯权重
109
+ Validating ...
110
+ Measured TFLOPs: 239.66
111
+ Epoch 1 | iter 32 step 1 | loss train: 1.425, val: 1.370 | iter time: 557.26 ms (step) remaining time: 0:46:27
112
+ Epoch 1 | iter 64 step 2 | loss train: 1.357, val: 1.370 | iter time: 358.88 ms (step) remaining time: 0:44:24
113
+ Epoch 1 | iter 96 step 3 | loss train: 1.439, val: 1.370 | iter time: 357.24 ms (step) remaining time: 0:43:38
114
+ Epoch 1 | iter 128 step 4 | loss train: 1.387, val: 1.370 | iter time: 359.44 ms (step) remaining time: 0:43:10
115
+ Epoch 1 | iter 160 step 5 | loss train: 1.403, val: 1.370 | iter time: 358.31 ms (step) remaining time: 0:42:49
116
+ Epoch 1 | iter 192 step 6 | loss train: 1.406, val: 1.370 | iter time: 359.10 ms (step) remaining time: 0:42:32
117
+ Epoch 1 | iter 224 step 7 | loss train: 1.423, val: 1.370 | iter time: 361.44 ms (step) remaining time: 0:42:17
118
+ Epoch 1 | iter 256 step 8 | loss train: 1.407, val: 1.370 | iter time: 358.88 ms (step) remaining time: 0:42:04
119
+ Epoch 1 | iter 288 step 9 | loss train: 1.441, val: 1.370 | iter time: 359.81 ms (step) remaining time: 0:41:50
120
+ Epoch 1 | iter 320 step 10 | loss train: 1.432, val: 1.370 | iter time: 357.69 ms (step) remaining time: 0:41:38
121
+ Epoch 1 | iter 352 step 11 | loss train: 1.422, val: 1.370 | iter time: 359.13 ms (step) remaining time: 0:41:26
122
+ Epoch 1 | iter 384 step 12 | loss train: 1.411, val: 1.370 | iter time: 361.31 ms (step) remaining time: 0:41:14
123
+ Epoch 1 | iter 416 step 13 | loss train: 1.398, val: 1.370 | iter time: 359.58 ms (step) remaining time: 0:41:02
124
+ Epoch 1 | iter 448 step 14 | loss train: 1.334, val: 1.370 | iter time: 358.30 ms (step) remaining time: 0:40:51
125
+ Epoch 1 | iter 480 step 15 | loss train: 1.360, val: 1.370 | iter time: 361.08 ms (step) remaining time: 0:40:40
126
+ Epoch 1 | iter 512 step 16 | loss train: 1.343, val: 1.370 | iter time: 359.32 ms (step) remaining time: 0:40:28
127
+ Epoch 1 | iter 544 step 17 | loss train: 1.461, val: 1.370 | iter time: 359.93 ms (step) remaining time: 0:40:17
128
+ Epoch 1 | iter 576 step 18 | loss train: 1.396, val: 1.370 | iter time: 359.52 ms (step) remaining time: 0:40:06
129
+ Epoch 1 | iter 608 step 19 | loss train: 1.358, val: 1.370 | iter time: 359.94 ms (step) remaining time: 0:39:55
130
+ Epoch 1 | iter 640 step 20 | loss train: 1.376, val: 1.370 | iter time: 360.33 ms (step) remaining time: 0:39:44
131
+ Epoch 1 | iter 672 step 21 | loss train: 1.387, val: 1.370 | iter time: 358.21 ms (step) remaining time: 0:39:32
132
+ Epoch 1 | iter 704 step 22 | loss train: 1.390, val: 1.370 | iter time: 358.39 ms (step) remaining time: 0:39:21
133
+ Epoch 1 | iter 736 step 23 | loss train: 1.424, val: 1.370 | iter time: 359.39 ms (step) remaining time: 0:39:10
134
+ Epoch 1 | iter 768 step 24 | loss train: 1.454, val: 1.370 | iter time: 360.41 ms (step) remaining time: 0:38:59
135
+ Epoch 1 | iter 800 step 25 | loss train: 1.416, val: 1.370 | iter time: 361.36 ms (step) remaining time: 0:38:48
136
+ Epoch 1 | iter 832 step 26 | loss train: 1.419, val: 1.370 | iter time: 359.95 ms (step) remaining time: 0:38:37
137
+ Epoch 1 | iter 864 step 27 | loss train: 1.362, val: 1.370 | iter time: 359.55 ms (step) remaining time: 0:38:26
138
+ Epoch 1 | iter 896 step 28 | loss train: 1.386, val: 1.370 | iter time: 359.90 ms (step) remaining time: 0:38:15
139
+ Epoch 1 | iter 928 step 29 | loss train: 1.414, val: 1.370 | iter time: 360.66 ms (step) remaining time: 0:38:04
140
+ Epoch 1 | iter 960 step 30 | loss train: 1.452, val: 1.370 | iter time: 358.72 ms (step) remaining time: 0:37:53
141
+ Epoch 1 | iter 992 step 31 | loss train: 1.395, val: 1.370 | iter time: 361.33 ms (step) remaining time: 0:37:42
142
+ Epoch 1 | iter 1024 step 32 | loss train: 1.423, val: 1.370 | iter time: 360.06 ms (step) remaining time: 0:37:31
143
+ Epoch 1 | iter 1056 step 33 | loss train: 1.322, val: 1.370 | iter time: 360.32 ms (step) remaining time: 0:37:20
144
+ Epoch 1 | iter 1088 step 34 | loss train: 1.382, val: 1.370 | iter time: 360.10 ms (step) remaining time: 0:37:09
145
+ Epoch 1 | iter 1120 step 35 | loss train: 1.402, val: 1.370 | iter time: 359.58 ms (step) remaining time: 0:36:58
146
+ Epoch 1 | iter 1152 step 36 | loss train: 1.390, val: 1.370 | iter time: 358.14 ms (step) remaining time: 0:36:47
147
+ Epoch 1 | iter 1184 step 37 | loss train: 1.391, val: 1.370 | iter time: 359.54 ms (step) remaining time: 0:36:37
148
+ Epoch 1 | iter 1216 step 38 | loss train: 1.392, val: 1.370 | iter time: 361.43 ms (step) remaining time: 0:36:26
149
+ Epoch 1 | iter 1248 step 39 | loss train: 1.404, val: 1.370 | iter time: 358.98 ms (step) remaining time: 0:36:15
150
+ Epoch 1 | iter 1280 step 40 | loss train: 1.417, val: 1.370 | iter time: 359.41 ms (step) remaining time: 0:36:04
151
+ Epoch 1 | iter 1312 step 41 | loss train: 1.420, val: 1.370 | iter time: 361.02 ms (step) remaining time: 0:35:53
152
+ Epoch 1 | iter 1344 step 42 | loss train: 1.390, val: 1.370 | iter time: 360.54 ms (step) remaining time: 0:35:42
153
+ Epoch 1 | iter 1376 step 43 | loss train: 1.390, val: 1.370 | iter time: 359.14 ms (step) remaining time: 0:35:31
154
+ Epoch 1 | iter 1408 step 44 | loss train: 1.418, val: 1.370 | iter time: 358.42 ms (step) remaining time: 0:35:20
155
+ Epoch 1 | iter 1440 step 45 | loss train: 1.355, val: 1.370 | iter time: 359.41 ms (step) remaining time: 0:35:09
156
+ Epoch 1 | iter 1472 step 46 | loss train: 1.331, val: 1.370 | iter time: 358.29 ms (step) remaining time: 0:34:59
157
+ Epoch 1 | iter 1504 step 47 | loss train: 1.356, val: 1.370 | iter time: 358.49 ms (step) remaining time: 0:34:48
158
+ Epoch 1 | iter 1536 step 48 | loss train: 1.359, val: 1.370 | iter time: 360.59 ms (step) remaining time: 0:34:37
159
+ Epoch 1 | iter 1568 step 49 | loss train: 1.383, val: 1.370 | iter time: 361.03 ms (step) remaining time: 0:34:27
160
+ Epoch 1 | iter 1600 step 50 | loss train: 1.459, val: 1.370 | iter time: 359.73 ms (step) remaining time: 0:34:16
161
+ Validating ...
162
+ iter 1600: val loss 1.3383, val time: 21900.94 ms
163
+ Epoch 1 | iter 1632 step 51 | loss train: 1.385, val: 1.338 | iter time: 358.90 ms (step) remaining time: 0:35:25
164
+ Epoch 1 | iter 1664 step 52 | loss train: 1.389, val: 1.338 | iter time: 359.75 ms (step) remaining time: 0:35:12
165
+ Epoch 1 | iter 1696 step 53 | loss train: 1.427, val: 1.338 | iter time: 358.72 ms (step) remaining time: 0:35:00
166
+ Epoch 1 | iter 1728 step 54 | loss train: 1.336, val: 1.338 | iter time: 359.96 ms (step) remaining time: 0:34:47
167
+ Epoch 1 | iter 1760 step 55 | loss train: 1.348, val: 1.338 | iter time: 635.02 ms (step) remaining time: 0:34:35
168
+ Epoch 1 | iter 1792 step 56 | loss train: 1.421, val: 1.338 | iter time: 358.31 ms (step) remaining time: 0:34:22
169
+ Epoch 1 | iter 1824 step 57 | loss train: 1.421, val: 1.338 | iter time: 358.69 ms (step) remaining time: 0:34:10
170
+ Epoch 1 | iter 1856 step 58 | loss train: 1.427, val: 1.338 | iter time: 361.00 ms (step) remaining time: 0:33:57
171
+ Epoch 1 | iter 1888 step 59 | loss train: 1.435, val: 1.338 | iter time: 358.88 ms (step) remaining time: 0:33:45
172
+ Epoch 1 | iter 1920 step 60 | loss train: 1.418, val: 1.338 | iter time: 360.23 ms (step) remaining time: 0:33:32
173
+ Epoch 1 | iter 1952 step 61 | loss train: 1.373, val: 1.338 | iter time: 359.76 ms (step) remaining time: 0:33:20
174
+ Epoch 1 | iter 1984 step 62 | loss train: 1.432, val: 1.338 | iter time: 358.65 ms (step) remaining time: 0:33:08
175
+ Epoch 1 | iter 2016 step 63 | loss train: 1.288, val: 1.338 | iter time: 358.24 ms (step) remaining time: 0:32:55
176
+ Epoch 1 | iter 2048 step 64 | loss train: 1.354, val: 1.338 | iter time: 359.77 ms (step) remaining time: 0:32:43
177
+ Epoch 1 | iter 2080 step 65 | loss train: 1.350, val: 1.338 | iter time: 359.76 ms (step) remaining time: 0:32:31
178
+ Epoch 1 | iter 2112 step 66 | loss train: 1.341, val: 1.338 | iter time: 358.35 ms (step) remaining time: 0:32:19
179
+ Epoch 1 | iter 2144 step 67 | loss train: 1.413, val: 1.338 | iter time: 361.61 ms (step) remaining time: 0:32:07
180
+ Epoch 1 | iter 2176 step 68 | loss train: 1.380, val: 1.338 | iter time: 358.20 ms (step) remaining time: 0:31:55
181
+ Epoch 1 | iter 2208 step 69 | loss train: 1.374, val: 1.338 | iter time: 358.46 ms (step) remaining time: 0:31:43
182
+ Epoch 1 | iter 2240 step 70 | loss train: 1.466, val: 1.338 | iter time: 360.78 ms (step) remaining time: 0:31:31
183
+ Epoch 1 | iter 2272 step 71 | loss train: 1.385, val: 1.338 | iter time: 358.25 ms (step) remaining time: 0:31:19
184
+ Epoch 1 | iter 2304 step 72 | loss train: 1.379, val: 1.338 | iter time: 359.42 ms (step) remaining time: 0:31:07
185
+ Epoch 1 | iter 2336 step 73 | loss train: 1.410, val: 1.338 | iter time: 358.83 ms (step) remaining time: 0:30:55
186
+ Epoch 1 | iter 2368 step 74 | loss train: 1.296, val: 1.338 | iter time: 359.04 ms (step) remaining time: 0:30:43
187
+ Epoch 1 | iter 2400 step 75 | loss train: 1.381, val: 1.338 | iter time: 359.99 ms (step) remaining time: 0:30:31
188
+ Epoch 1 | iter 2432 step 76 | loss train: 1.411, val: 1.338 | iter time: 358.24 ms (step) remaining time: 0:30:19
189
+ Epoch 1 | iter 2464 step 77 | loss train: 1.360, val: 1.338 | iter time: 360.91 ms (step) remaining time: 0:30:07
190
+ Epoch 1 | iter 2496 step 78 | loss train: 1.420, val: 1.338 | iter time: 358.21 ms (step) remaining time: 0:29:56
191
+ Epoch 1 | iter 2528 step 79 | loss train: 1.315, val: 1.338 | iter time: 361.07 ms (step) remaining time: 0:29:44
192
+ Epoch 1 | iter 2560 step 80 | loss train: 1.351, val: 1.338 | iter time: 358.36 ms (step) remaining time: 0:29:32
193
+ Epoch 1 | iter 2592 step 81 | loss train: 1.430, val: 1.338 | iter time: 357.41 ms (step) remaining time: 0:29:20
194
+ Epoch 1 | iter 2624 step 82 | loss train: 1.388, val: 1.338 | iter time: 360.66 ms (step) remaining time: 0:29:09
195
+ Epoch 1 | iter 2656 step 83 | loss train: 1.356, val: 1.338 | iter time: 361.16 ms (step) remaining time: 0:28:57
196
+ Epoch 1 | iter 2688 step 84 | loss train: 1.398, val: 1.338 | iter time: 358.34 ms (step) remaining time: 0:28:45
197
+ Epoch 1 | iter 2720 step 85 | loss train: 1.425, val: 1.338 | iter time: 360.00 ms (step) remaining time: 0:28:34
198
+ Epoch 1 | iter 2752 step 86 | loss train: 1.379, val: 1.338 | iter time: 361.11 ms (step) remaining time: 0:28:22
199
+ Epoch 1 | iter 2784 step 87 | loss train: 1.336, val: 1.338 | iter time: 360.53 ms (step) remaining time: 0:28:10
200
+ Epoch 1 | iter 2816 step 88 | loss train: 1.273, val: 1.338 | iter time: 360.65 ms (step) remaining time: 0:27:59
201
+ Epoch 1 | iter 2848 step 89 | loss train: 1.362, val: 1.338 | iter time: 359.41 ms (step) remaining time: 0:27:47
202
+ Epoch 1 | iter 2880 step 90 | loss train: 1.361, val: 1.338 | iter time: 360.27 ms (step) remaining time: 0:27:36
203
+ Epoch 1 | iter 2912 step 91 | loss train: 1.392, val: 1.338 | iter time: 359.95 ms (step) remaining time: 0:27:24
204
+ Epoch 1 | iter 2944 step 92 | loss train: 1.367, val: 1.338 | iter time: 358.76 ms (step) remaining time: 0:27:13
205
+ Epoch 1 | iter 2976 step 93 | loss train: 1.412, val: 1.338 | iter time: 359.34 ms (step) remaining time: 0:27:01
206
+ Epoch 1 | iter 3008 step 94 | loss train: 1.381, val: 1.338 | iter time: 359.27 ms (step) remaining time: 0:26:50
207
+ Epoch 1 | iter 3040 step 95 | loss train: 1.361, val: 1.338 | iter time: 359.58 ms (step) remaining time: 0:26:38
208
+ Epoch 1 | iter 3072 step 96 | loss train: 1.366, val: 1.338 | iter time: 357.97 ms (step) remaining time: 0:26:27
209
+ Epoch 1 | iter 3104 step 97 | loss train: 1.349, val: 1.338 | iter time: 358.62 ms (step) remaining time: 0:26:15
210
+ Epoch 1 | iter 3136 step 98 | loss train: 1.381, val: 1.338 | iter time: 360.10 ms (step) remaining time: 0:26:04
211
+ Epoch 1 | iter 3168 step 99 | loss train: 1.358, val: 1.338 | iter time: 358.33 ms (step) remaining time: 0:25:52
212
+ Epoch 1 | iter 3200 step 100 | loss train: 1.398, val: 1.338 | iter time: 359.29 ms (step) remaining time: 0:25:41
213
+ Validating ...
214
+ iter 3200: val loss 1.3139, val time: 21910.85 ms
215
+ Saving checkpoint to 'out/pretrain/tinyllama/2411_lr4e-5/step-00000100/lit_model.pth'
216
+ Epoch 1 | iter 3232 step 101 | loss train: 1.311, val: 1.314 | iter time: 357.56 ms (step) remaining time: 0:26:21
217
+ Epoch 1 | iter 3264 step 102 | loss train: 1.323, val: 1.314 | iter time: 356.72 ms (step) remaining time: 0:26:09
218
+ Epoch 1 | iter 3296 step 103 | loss train: 1.438, val: 1.314 | iter time: 359.89 ms (step) remaining time: 0:25:56
219
+ Epoch 1 | iter 3328 step 104 | loss train: 1.375, val: 1.314 | iter time: 358.43 ms (step) remaining time: 0:25:44
220
+ Epoch 1 | iter 3360 step 105 | loss train: 1.390, val: 1.314 | iter time: 359.49 ms (step) remaining time: 0:25:32
221
+ Epoch 1 | iter 3392 step 106 | loss train: 1.341, val: 1.314 | iter time: 359.77 ms (step) remaining time: 0:25:20
222
+ Epoch 1 | iter 3424 step 107 | loss train: 1.385, val: 1.314 | iter time: 359.95 ms (step) remaining time: 0:25:08
223
+ Epoch 1 | iter 3456 step 108 | loss train: 1.376, val: 1.314 | iter time: 358.45 ms (step) remaining time: 0:24:56
224
+ Epoch 1 | iter 3488 step 109 | loss train: 1.362, val: 1.314 | iter time: 358.93 ms (step) remaining time: 0:24:44
225
+ Epoch 1 | iter 3520 step 110 | loss train: 1.355, val: 1.314 | iter time: 358.55 ms (step) remaining time: 0:24:32
226
+ Epoch 1 | iter 3552 step 111 | loss train: 1.465, val: 1.314 | iter time: 360.07 ms (step) remaining time: 0:24:20
227
+ Epoch 1 | iter 3584 step 112 | loss train: 1.377, val: 1.314 | iter time: 359.74 ms (step) remaining time: 0:24:08
228
+ Epoch 1 | iter 3616 step 113 | loss train: 1.381, val: 1.314 | iter time: 584.82 ms (step) remaining time: 0:23:56
229
+ Epoch 1 | iter 3648 step 114 | loss train: 1.431, val: 1.314 | iter time: 359.99 ms (step) remaining time: 0:23:44
230
+ Epoch 1 | iter 3680 step 115 | loss train: 1.325, val: 1.314 | iter time: 361.19 ms (step) remaining time: 0:23:32
231
+ Epoch 1 | iter 3712 step 116 | loss train: 1.322, val: 1.314 | iter time: 358.81 ms (step) remaining time: 0:23:20
232
+ Epoch 1 | iter 3744 step 117 | loss train: 1.369, val: 1.314 | iter time: 360.20 ms (step) remaining time: 0:23:08
233
+ Epoch 1 | iter 3776 step 118 | loss train: 1.396, val: 1.314 | iter time: 359.36 ms (step) remaining time: 0:22:56
234
+ Epoch 1 | iter 3808 step 119 | loss train: 1.327, val: 1.314 | iter time: 359.73 ms (step) remaining time: 0:22:44
235
+ Epoch 1 | iter 3840 step 120 | loss train: 1.401, val: 1.314 | iter time: 360.06 ms (step) remaining time: 0:22:33
236
+ Epoch 1 | iter 3872 step 121 | loss train: 1.395, val: 1.314 | iter time: 357.15 ms (step) remaining time: 0:22:21
237
+ Epoch 1 | iter 3904 step 122 | loss train: 1.335, val: 1.314 | iter time: 361.27 ms (step) remaining time: 0:22:09
238
+ Epoch 1 | iter 3936 step 123 | loss train: 1.401, val: 1.314 | iter time: 358.41 ms (step) remaining time: 0:21:57
239
+ Epoch 1 | iter 3968 step 124 | loss train: 1.397, val: 1.314 | iter time: 358.62 ms (step) remaining time: 0:21:45
240
+ Epoch 1 | iter 4000 step 125 | loss train: 1.360, val: 1.314 | iter time: 357.98 ms (step) remaining time: 0:21:34
241
+ Epoch 1 | iter 4032 step 126 | loss train: 1.334, val: 1.314 | iter time: 359.19 ms (step) remaining time: 0:21:22
242
+ Epoch 1 | iter 4064 step 127 | loss train: 1.383, val: 1.314 | iter time: 358.96 ms (step) remaining time: 0:21:10
243
+ Epoch 1 | iter 4096 step 128 | loss train: 1.360, val: 1.314 | iter time: 359.49 ms (step) remaining time: 0:20:58
244
+ Epoch 1 | iter 4128 step 129 | loss train: 1.442, val: 1.314 | iter time: 359.41 ms (step) remaining time: 0:20:47
245
+ Epoch 1 | iter 4160 step 130 | loss train: 1.379, val: 1.314 | iter time: 359.43 ms (step) remaining time: 0:20:35
246
+ Epoch 1 | iter 4192 step 131 | loss train: 1.338, val: 1.314 | iter time: 360.82 ms (step) remaining time: 0:20:23
247
+ Epoch 1 | iter 4224 step 132 | loss train: 1.317, val: 1.314 | iter time: 358.28 ms (step) remaining time: 0:20:12
248
+ Epoch 1 | iter 4256 step 133 | loss train: 1.375, val: 1.314 | iter time: 359.99 ms (step) remaining time: 0:20:00
249
+ Epoch 1 | iter 4288 step 134 | loss train: 1.378, val: 1.314 | iter time: 360.52 ms (step) remaining time: 0:19:48
250
+ Epoch 1 | iter 4320 step 135 | loss train: 1.407, val: 1.314 | iter time: 358.86 ms (step) remaining time: 0:19:36
251
+ Epoch 1 | iter 4352 step 136 | loss train: 1.369, val: 1.314 | iter time: 358.37 ms (step) remaining time: 0:19:25
252
+ Epoch 1 | iter 4384 step 137 | loss train: 1.383, val: 1.314 | iter time: 359.87 ms (step) remaining time: 0:19:13
253
+ Epoch 1 | iter 4416 step 138 | loss train: 1.391, val: 1.314 | iter time: 360.27 ms (step) remaining time: 0:19:02
254
+ Epoch 1 | iter 4448 step 139 | loss train: 1.351, val: 1.314 | iter time: 359.80 ms (step) remaining time: 0:18:50
255
+ Epoch 1 | iter 4480 step 140 | loss train: 1.364, val: 1.314 | iter time: 359.44 ms (step) remaining time: 0:18:38
256
+ Epoch 1 | iter 4512 step 141 | loss train: 1.379, val: 1.314 | iter time: 358.38 ms (step) remaining time: 0:18:27
257
+ Epoch 1 | iter 4544 step 142 | loss train: 1.359, val: 1.314 | iter time: 358.63 ms (step) remaining time: 0:18:15
258
+ Epoch 1 | iter 4576 step 143 | loss train: 1.373, val: 1.314 | iter time: 359.62 ms (step) remaining time: 0:18:03
259
+ Epoch 1 | iter 4608 step 144 | loss train: 1.381, val: 1.314 | iter time: 357.44 ms (step) remaining time: 0:17:52
260
+ Epoch 1 | iter 4640 step 145 | loss train: 1.403, val: 1.314 | iter time: 358.92 ms (step) remaining time: 0:17:40
261
+ Epoch 1 | iter 4672 step 146 | loss train: 1.279, val: 1.314 | iter time: 358.93 ms (step) remaining time: 0:17:29
262
+ Epoch 1 | iter 4704 step 147 | loss train: 1.358, val: 1.314 | iter time: 358.71 ms (step) remaining time: 0:17:17
263
+ Epoch 1 | iter 4736 step 148 | loss train: 1.381, val: 1.314 | iter time: 358.61 ms (step) remaining time: 0:17:06
264
+ Epoch 1 | iter 4768 step 149 | loss train: 1.395, val: 1.314 | iter time: 359.43 ms (step) remaining time: 0:16:54
265
+ Epoch 1 | iter 4800 step 150 | loss train: 1.352, val: 1.314 | iter time: 357.49 ms (step) remaining time: 0:16:43
266
+ Validating ...
267
+ iter 4800: val loss 1.3058, val time: 21890.86 ms
268
+ Epoch 1 | iter 4832 step 151 | loss train: 1.349, val: 1.306 | iter time: 357.46 ms (step) remaining time: 0:16:44
269
+ Epoch 1 | iter 4864 step 152 | loss train: 1.342, val: 1.306 | iter time: 359.41 ms (step) remaining time: 0:16:32
270
+ Epoch 1 | iter 4896 step 153 | loss train: 1.351, val: 1.306 | iter time: 359.81 ms (step) remaining time: 0:16:20
271
+ Epoch 1 | iter 4928 step 154 | loss train: 1.299, val: 1.306 | iter time: 358.87 ms (step) remaining time: 0:16:09
272
+ Epoch 1 | iter 4960 step 155 | loss train: 1.376, val: 1.306 | iter time: 358.86 ms (step) remaining time: 0:15:57
273
+ Epoch 1 | iter 4992 step 156 | loss train: 1.328, val: 1.306 | iter time: 360.34 ms (step) remaining time: 0:15:45
274
+ Epoch 1 | iter 5024 step 157 | loss train: 1.400, val: 1.306 | iter time: 358.51 ms (step) remaining time: 0:15:34
275
+ Epoch 1 | iter 5056 step 158 | loss train: 1.365, val: 1.306 | iter time: 358.29 ms (step) remaining time: 0:15:22
276
+ Epoch 1 | iter 5088 step 159 | loss train: 1.407, val: 1.306 | iter time: 360.57 ms (step) remaining time: 0:15:10
277
+ Epoch 1 | iter 5120 step 160 | loss train: 1.391, val: 1.306 | iter time: 358.20 ms (step) remaining time: 0:14:59
278
+ Epoch 1 | iter 5152 step 161 | loss train: 1.380, val: 1.306 | iter time: 357.64 ms (step) remaining time: 0:14:47
279
+ Epoch 1 | iter 5184 step 162 | loss train: 1.345, val: 1.306 | iter time: 358.64 ms (step) remaining time: 0:14:35
280
+ Epoch 1 | iter 5216 step 163 | loss train: 1.341, val: 1.306 | iter time: 359.35 ms (step) remaining time: 0:14:24
281
+ Epoch 1 | iter 5248 step 164 | loss train: 1.271, val: 1.306 | iter time: 359.06 ms (step) remaining time: 0:14:12
282
+ Epoch 1 | iter 5280 step 165 | loss train: 1.330, val: 1.306 | iter time: 358.38 ms (step) remaining time: 0:14:01
283
+ Epoch 1 | iter 5312 step 166 | loss train: 1.346, val: 1.306 | iter time: 360.27 ms (step) remaining time: 0:13:49
284
+ Epoch 1 | iter 5344 step 167 | loss train: 1.325, val: 1.306 | iter time: 358.67 ms (step) remaining time: 0:13:37
285
+ Epoch 1 | iter 5376 step 168 | loss train: 1.397, val: 1.306 | iter time: 360.59 ms (step) remaining time: 0:13:26
286
+ Epoch 1 | iter 5408 step 169 | loss train: 1.329, val: 1.306 | iter time: 359.50 ms (step) remaining time: 0:13:14
287
+ Epoch 1 | iter 5440 step 170 | loss train: 1.366, val: 1.306 | iter time: 360.68 ms (step) remaining time: 0:13:03
288
+ Epoch 1 | iter 5472 step 171 | loss train: 1.365, val: 1.306 | iter time: 358.40 ms (step) remaining time: 0:12:51
289
+ Epoch 1 | iter 5504 step 172 | loss train: 1.384, val: 1.306 | iter time: 358.19 ms (step) remaining time: 0:12:40
290
+ Epoch 1 | iter 5536 step 173 | loss train: 1.356, val: 1.306 | iter time: 358.01 ms (step) remaining time: 0:12:28
291
+ Epoch 1 | iter 5568 step 174 | loss train: 1.381, val: 1.306 | iter time: 358.96 ms (step) remaining time: 0:12:17
292
+ Epoch 1 | iter 5600 step 175 | loss train: 1.273, val: 1.306 | iter time: 358.58 ms (step) remaining time: 0:12:05
293
+ Epoch 1 | iter 5632 step 176 | loss train: 1.348, val: 1.306 | iter time: 361.06 ms (step) remaining time: 0:11:54
294
+ Epoch 1 | iter 5664 step 177 | loss train: 1.291, val: 1.306 | iter time: 357.70 ms (step) remaining time: 0:11:42
295
+ Epoch 1 | iter 5696 step 178 | loss train: 1.348, val: 1.306 | iter time: 359.83 ms (step) remaining time: 0:11:31
296
+ Epoch 1 | iter 5728 step 179 | loss train: 1.367, val: 1.306 | iter time: 358.49 ms (step) remaining time: 0:11:19
297
+ Epoch 1 | iter 5760 step 180 | loss train: 1.295, val: 1.306 | iter time: 357.70 ms (step) remaining time: 0:11:08
298
+ Epoch 1 | iter 5792 step 181 | loss train: 1.335, val: 1.306 | iter time: 359.37 ms (step) remaining time: 0:10:56
299
+ Epoch 1 | iter 5824 step 182 | loss train: 1.340, val: 1.306 | iter time: 358.24 ms (step) remaining time: 0:10:45
300
+ Epoch 1 | iter 5856 step 183 | loss train: 1.318, val: 1.306 | iter time: 355.62 ms (step) remaining time: 0:10:33
301
+ Epoch 1 | iter 5888 step 184 | loss train: 1.372, val: 1.306 | iter time: 358.89 ms (step) remaining time: 0:10:22
302
+ Epoch 1 | iter 5920 step 185 | loss train: 1.386, val: 1.306 | iter time: 357.10 ms (step) remaining time: 0:10:10
303
+ Epoch 1 | iter 5952 step 186 | loss train: 1.386, val: 1.306 | iter time: 360.36 ms (step) remaining time: 0:09:59
304
+ Epoch 1 | iter 5984 step 187 | loss train: 1.323, val: 1.306 | iter time: 360.42 ms (step) remaining time: 0:09:47
305
+ Epoch 1 | iter 6016 step 188 | loss train: 1.316, val: 1.306 | iter time: 359.29 ms (step) remaining time: 0:09:36
306
+ Epoch 1 | iter 6048 step 189 | loss train: 1.380, val: 1.306 | iter time: 357.98 ms (step) remaining time: 0:09:25
307
+ Epoch 1 | iter 6080 step 190 | loss train: 1.337, val: 1.306 | iter time: 360.39 ms (step) remaining time: 0:09:13
308
+ Epoch 1 | iter 6112 step 191 | loss train: 1.375, val: 1.306 | iter time: 359.23 ms (step) remaining time: 0:09:02
309
+ Epoch 1 | iter 6144 step 192 | loss train: 1.444, val: 1.306 | iter time: 358.45 ms (step) remaining time: 0:08:50
310
+ Epoch 1 | iter 6176 step 193 | loss train: 1.324, val: 1.306 | iter time: 359.90 ms (step) remaining time: 0:08:39
311
+ Epoch 1 | iter 6208 step 194 | loss train: 1.323, val: 1.306 | iter time: 612.97 ms (step) remaining time: 0:08:28
312
+ Epoch 1 | iter 6240 step 195 | loss train: 1.316, val: 1.306 | iter time: 360.08 ms (step) remaining time: 0:08:16
313
+ Epoch 1 | iter 6272 step 196 | loss train: 1.370, val: 1.306 | iter time: 356.98 ms (step) remaining time: 0:08:05
314
+ Epoch 1 | iter 6304 step 197 | loss train: 1.388, val: 1.306 | iter time: 361.03 ms (step) remaining time: 0:07:54
315
+ Epoch 1 | iter 6336 step 198 | loss train: 1.365, val: 1.306 | iter time: 358.46 ms (step) remaining time: 0:07:42
316
+ Epoch 1 | iter 6368 step 199 | loss train: 1.332, val: 1.306 | iter time: 360.29 ms (step) remaining time: 0:07:31
317
+ Epoch 1 | iter 6400 step 200 | loss train: 1.451, val: 1.306 | iter time: 359.46 ms (step) remaining time: 0:07:19
318
+ Validating ...
319
+ iter 6400: val loss 1.3072, val time: 21899.37 ms
320
+ Saving checkpoint to 'out/pretrain/tinyllama/2411_lr4e-5/step-00000200/lit_model.pth'
321
+ Epoch 1 | iter 6432 step 201 | loss train: 1.341, val: 1.307 | iter time: 358.12 ms (step) remaining time: 0:07:15
322
+ Epoch 1 | iter 6464 step 202 | loss train: 1.344, val: 1.307 | iter time: 359.29 ms (step) remaining time: 0:07:04
323
+ Epoch 1 | iter 6496 step 203 | loss train: 1.339, val: 1.307 | iter time: 360.33 ms (step) remaining time: 0:06:52
324
+ Epoch 1 | iter 6528 step 204 | loss train: 1.402, val: 1.307 | iter time: 360.76 ms (step) remaining time: 0:06:41
325
+ Epoch 1 | iter 6560 step 205 | loss train: 1.297, val: 1.307 | iter time: 359.61 ms (step) remaining time: 0:06:29
326
+ Epoch 1 | iter 6592 step 206 | loss train: 1.375, val: 1.307 | iter time: 360.07 ms (step) remaining time: 0:06:17
327
+ Epoch 1 | iter 6624 step 207 | loss train: 1.332, val: 1.307 | iter time: 361.61 ms (step) remaining time: 0:06:06
328
+ Epoch 1 | iter 6656 step 208 | loss train: 1.331, val: 1.307 | iter time: 359.24 ms (step) remaining time: 0:05:54
329
+ Epoch 1 | iter 6688 step 209 | loss train: 1.379, val: 1.307 | iter time: 359.51 ms (step) remaining time: 0:05:43
330
+ Epoch 1 | iter 6720 step 210 | loss train: 1.335, val: 1.307 | iter time: 360.13 ms (step) remaining time: 0:05:31
331
+ Epoch 1 | iter 6752 step 211 | loss train: 1.298, val: 1.307 | iter time: 358.39 ms (step) remaining time: 0:05:20
332
+ Epoch 1 | iter 6784 step 212 | loss train: 1.303, val: 1.307 | iter time: 359.29 ms (step) remaining time: 0:05:08
333
+ Epoch 1 | iter 6816 step 213 | loss train: 1.384, val: 1.307 | iter time: 360.50 ms (step) remaining time: 0:04:57
334
+ Epoch 1 | iter 6848 step 214 | loss train: 1.401, val: 1.307 | iter time: 359.59 ms (step) remaining time: 0:04:45
335
+ Epoch 1 | iter 6880 step 215 | loss train: 1.283, val: 1.307 | iter time: 358.61 ms (step) remaining time: 0:04:34
336
+ Epoch 1 | iter 6912 step 216 | loss train: 1.342, val: 1.307 | iter time: 360.32 ms (step) remaining time: 0:04:22
337
+ Epoch 1 | iter 6944 step 217 | loss train: 1.273, val: 1.307 | iter time: 359.43 ms (step) remaining time: 0:04:11
338
+ Epoch 1 | iter 6976 step 218 | loss train: 1.313, val: 1.307 | iter time: 360.51 ms (step) remaining time: 0:03:59
339
+ Epoch 1 | iter 7008 step 219 | loss train: 1.334, val: 1.307 | iter time: 359.06 ms (step) remaining time: 0:03:48
340
+ Epoch 1 | iter 7040 step 220 | loss train: 1.355, val: 1.307 | iter time: 359.24 ms (step) remaining time: 0:03:36
341
+ Epoch 1 | iter 7072 step 221 | loss train: 1.355, val: 1.307 | iter time: 359.64 ms (step) remaining time: 0:03:25
342
+ Epoch 1 | iter 7104 step 222 | loss train: 1.301, val: 1.307 | iter time: 358.80 ms (step) remaining time: 0:03:13
343
+ Epoch 1 | iter 7136 step 223 | loss train: 1.348, val: 1.307 | iter time: 358.49 ms (step) remaining time: 0:03:02
344
+ Epoch 1 | iter 7168 step 224 | loss train: 1.378, val: 1.307 | iter time: 358.56 ms (step) remaining time: 0:02:51
345
+ Epoch 1 | iter 7200 step 225 | loss train: 1.429, val: 1.307 | iter time: 358.82 ms (step) remaining time: 0:02:39
346
+ Epoch 1 | iter 7232 step 226 | loss train: 1.401, val: 1.307 | iter time: 359.42 ms (step) remaining time: 0:02:28
347
+ Epoch 1 | iter 7264 step 227 | loss train: 1.388, val: 1.307 | iter time: 360.22 ms (step) remaining time: 0:02:16
348
+ Epoch 1 | iter 7296 step 228 | loss train: 1.397, val: 1.307 | iter time: 359.12 ms (step) remaining time: 0:02:05
349
+ Epoch 1 | iter 7328 step 229 | loss train: 1.382, val: 1.307 | iter time: 358.52 ms (step) remaining time: 0:01:53
350
+ Epoch 1 | iter 7360 step 230 | loss train: 1.304, val: 1.307 | iter time: 358.21 ms (step) remaining time: 0:01:42
351
+ Epoch 1 | iter 7392 step 231 | loss train: 1.306, val: 1.307 | iter time: 358.57 ms (step) remaining time: 0:01:31
352
+ Epoch 1 | iter 7424 step 232 | loss train: 1.297, val: 1.307 | iter time: 359.78 ms (step) remaining time: 0:01:19
353
+ Epoch 1 | iter 7456 step 233 | loss train: 1.375, val: 1.307 | iter time: 360.36 ms (step) remaining time: 0:01:08
354
+ Epoch 1 | iter 7488 step 234 | loss train: 1.353, val: 1.307 | iter time: 360.17 ms (step) remaining time: 0:00:56
355
+ Epoch 1 | iter 7520 step 235 | loss train: 1.319, val: 1.307 | iter time: 359.95 ms (step) remaining time: 0:00:45
356
+ Epoch 1 | iter 7552 step 236 | loss train: 1.331, val: 1.307 | iter time: 359.22 ms (step) remaining time: 0:00:34
357
+ Epoch 1 | iter 7584 step 237 | loss train: 1.325, val: 1.307 | iter time: 360.10 ms (step) remaining time: 0:00:22
358
+ Epoch 1 | iter 7616 step 238 | loss train: 1.324, val: 1.307 | iter time: 360.19 ms (step) remaining time: 0:00:11
359
+ Epoch 2 | iter 7648 step 239 | loss train: 1.201, val: 1.307 | iter time: 359.45 ms (step) remaining time: 0:00:00
360
+ Validating ...
361
+ Final evaluation | val loss: 1.301 | val ppl: 3.672
362
+ Saving checkpoint to 'out/pretrain/tinyllama/2411_lr4e-5/final/lit_model.pth'
363
+ ----------------------------------------
364
+ | Performance
365
+ | - Total tokens : 250,609,664
366
+ | - Training Time : 2780.44 s
367
+ | - Tok/sec : 129.06 tok/s
368
+ | ----------------------------------------
369
+ | Memory Usage
370
+ | - Memory Used : 26.32 GB
371
+ ----------------------------------------
out/pretrain/tinyllama/teelogs/2412.txt ADDED
@@ -0,0 +1,414 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Initializing distributed: GLOBAL_RANK: 0, MEMBER: 1/4
2
+ Initializing distributed: GLOBAL_RANK: 2, MEMBER: 3/4
3
+ [rank: 2] Seed set to 42
4
+ Initializing distributed: GLOBAL_RANK: 3, MEMBER: 4/4
5
+ Initializing distributed: GLOBAL_RANK: 1, MEMBER: 2/4
6
+ ----------------------------------------------------------------------------------------------------
7
+ distributed_backend=nccl
8
+ All distributed processes registered. Starting with 4 processes
9
+ ----------------------------------------------------------------------------------------------------
10
+
11
+ [rank: 1] Seed set to 42
12
+ [rank: 3] Seed set to 42
13
+ [rank: 0] Seed set to 42
14
+ All GPUs are fully connected via NVLink.
15
+ {'data': {'batch_size': 1,
16
+ 'data_path': PosixPath('litgpt/data/arxiv'),
17
+ 'num_workers': 8,
18
+ 'seed': 42,
19
+ 'seq_length': 2048,
20
+ 'use_starcoder': True},
21
+ 'data_dir': PosixPath('litgpt/data/arxiv/2412'),
22
+ 'devices': 'auto',
23
+ 'eval': {'evaluate_example': 'first',
24
+ 'final_validation': True,
25
+ 'initial_validation': True,
26
+ 'interval': 50,
27
+ 'max_iters': 100,
28
+ 'max_new_tokens': None},
29
+ 'initial_checkpoint_dir': PosixPath('out/pretrain/2411/final'),
30
+ 'log': {'group': None, 'project': None, 'run': None},
31
+ 'logger_name': 'tensorboard',
32
+ 'model_config': {'attention_logit_softcapping': None,
33
+ 'attention_scores_scalar': None,
34
+ 'attn_bias': False,
35
+ 'bias': False,
36
+ 'block_size': 2048,
37
+ 'final_logit_softcapping': None,
38
+ 'gelu_approximate': 'none',
39
+ 'head_size': 64,
40
+ 'hf_config': {'name': 'TinyLlama-1.1B-intermediate-step-1431k-3T',
41
+ 'org': 'TinyLlama'},
42
+ 'intermediate_size': 5632,
43
+ 'lm_head_bias': False,
44
+ 'mlp_class_name': 'LLaMAMLP',
45
+ 'moe_intermediate_size': None,
46
+ 'n_embd': 2048,
47
+ 'n_expert': 0,
48
+ 'n_expert_per_token': 0,
49
+ 'n_head': 32,
50
+ 'n_layer': 22,
51
+ 'n_query_groups': 4,
52
+ 'name': 'tiny-llama-1.1b',
53
+ 'norm_1': True,
54
+ 'norm_2': True,
55
+ 'norm_class_name': 'RMSNorm',
56
+ 'norm_eps': 1e-05,
57
+ 'norm_qk': False,
58
+ 'norm_qk_type': 'default',
59
+ 'padded_vocab_size': 32000,
60
+ 'padding_multiple': 64,
61
+ 'parallel_residual': False,
62
+ 'post_attention_norm': False,
63
+ 'post_mlp_norm': False,
64
+ 'rope_adjustments': None,
65
+ 'rope_base': 10000,
66
+ 'rope_condense_ratio': 1,
67
+ 'rope_indices': None,
68
+ 'rope_local_base_freq': None,
69
+ 'rotary_percentage': 1.0,
70
+ 'scale_embeddings': False,
71
+ 'shared_attention_norm': False,
72
+ 'sliding_window_indices': None,
73
+ 'sliding_window_size': None,
74
+ 'vocab_size': 32000},
75
+ 'model_name': 'tiny-llama-1.1b',
76
+ 'num_nodes': 1,
77
+ 'optimizer': "{'class_path': 'torch.optim.AdamW', 'init_args': {'lr': 0.0004, "
78
+ "'weight_decay': 0.1, 'betas': [0.9, 0.95]}}",
79
+ 'out_dir': PosixPath('out/pretrain/2412'),
80
+ 'precision': 'bf16-mixed',
81
+ 'resume': False,
82
+ 'seed': 42,
83
+ 'tokenizer_dir': PosixPath('checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T'),
84
+ 'train': {'epochs': None,
85
+ 'global_batch_size': 512,
86
+ 'log_interval': 1,
87
+ 'lr_warmup_fraction': None,
88
+ 'lr_warmup_steps': 20,
89
+ 'max_norm': 1.0,
90
+ 'max_seq_length': 2048,
91
+ 'max_steps': None,
92
+ 'max_tokens': 301989888,
93
+ 'micro_batch_size': 4,
94
+ 'min_lr': 4e-05,
95
+ 'save_interval': 100,
96
+ 'tie_embeddings': None}}
97
+ Time to instantiate model: 0.04 seconds.
98
+ Total parameters: 1,100,048,384
99
+ [fix] out/pretrain/2411/final/lit_model.pth 是整包 state,提取 model 权重并原子覆盖...
100
+ [fix] 已覆盖为纯权重: out/pretrain/2411/final/lit_model.pth
101
+ Validating ...
102
+ Measured TFLOPs: 239.66
103
+ Epoch 1 | iter 32 step 1 | loss train: 1.340, val: 1.515 | iter time: 550.46 ms (step) remaining time: 0:56:58
104
+ Epoch 1 | iter 64 step 2 | loss train: 1.361, val: 1.515 | iter time: 357.72 ms (step) remaining time: 0:54:01
105
+ Epoch 1 | iter 96 step 3 | loss train: 1.446, val: 1.515 | iter time: 357.30 ms (step) remaining time: 0:52:56
106
+ Epoch 1 | iter 128 step 4 | loss train: 1.434, val: 1.515 | iter time: 358.09 ms (step) remaining time: 0:52:19
107
+ Epoch 1 | iter 160 step 5 | loss train: 1.324, val: 1.515 | iter time: 358.59 ms (step) remaining time: 0:51:54
108
+ Epoch 1 | iter 192 step 6 | loss train: 1.345, val: 1.515 | iter time: 357.57 ms (step) remaining time: 0:51:34
109
+ Epoch 1 | iter 224 step 7 | loss train: 1.359, val: 1.515 | iter time: 357.80 ms (step) remaining time: 0:51:17
110
+ Epoch 1 | iter 256 step 8 | loss train: 1.362, val: 1.515 | iter time: 357.59 ms (step) remaining time: 0:51:02
111
+ Epoch 1 | iter 288 step 9 | loss train: 1.357, val: 1.515 | iter time: 360.39 ms (step) remaining time: 0:50:48
112
+ Epoch 1 | iter 320 step 10 | loss train: 1.415, val: 1.515 | iter time: 357.40 ms (step) remaining time: 0:50:34
113
+ Epoch 1 | iter 352 step 11 | loss train: 1.381, val: 1.515 | iter time: 358.29 ms (step) remaining time: 0:50:21
114
+ Epoch 1 | iter 384 step 12 | loss train: 1.444, val: 1.515 | iter time: 360.64 ms (step) remaining time: 0:50:09
115
+ Epoch 1 | iter 416 step 13 | loss train: 1.399, val: 1.515 | iter time: 358.73 ms (step) remaining time: 0:49:57
116
+ Epoch 1 | iter 448 step 14 | loss train: 1.363, val: 1.515 | iter time: 360.75 ms (step) remaining time: 0:49:45
117
+ Epoch 1 | iter 480 step 15 | loss train: 1.498, val: 1.515 | iter time: 359.49 ms (step) remaining time: 0:49:33
118
+ Epoch 1 | iter 512 step 16 | loss train: 1.433, val: 1.515 | iter time: 358.86 ms (step) remaining time: 0:49:21
119
+ Epoch 1 | iter 544 step 17 | loss train: 1.432, val: 1.515 | iter time: 358.03 ms (step) remaining time: 0:49:09
120
+ Epoch 1 | iter 576 step 18 | loss train: 1.403, val: 1.515 | iter time: 360.03 ms (step) remaining time: 0:48:58
121
+ Epoch 1 | iter 608 step 19 | loss train: 1.403, val: 1.515 | iter time: 358.78 ms (step) remaining time: 0:48:46
122
+ Epoch 1 | iter 640 step 20 | loss train: 1.457, val: 1.515 | iter time: 359.36 ms (step) remaining time: 0:48:35
123
+ Epoch 1 | iter 672 step 21 | loss train: 1.389, val: 1.515 | iter time: 358.37 ms (step) remaining time: 0:48:24
124
+ Epoch 1 | iter 704 step 22 | loss train: 1.465, val: 1.515 | iter time: 359.08 ms (step) remaining time: 0:48:13
125
+ Epoch 1 | iter 736 step 23 | loss train: 1.411, val: 1.515 | iter time: 358.30 ms (step) remaining time: 0:48:02
126
+ Epoch 1 | iter 768 step 24 | loss train: 1.469, val: 1.515 | iter time: 359.87 ms (step) remaining time: 0:47:51
127
+ Epoch 1 | iter 800 step 25 | loss train: 1.427, val: 1.515 | iter time: 358.92 ms (step) remaining time: 0:47:40
128
+ Epoch 1 | iter 832 step 26 | loss train: 1.383, val: 1.515 | iter time: 360.24 ms (step) remaining time: 0:47:29
129
+ Epoch 1 | iter 864 step 27 | loss train: 1.419, val: 1.515 | iter time: 360.07 ms (step) remaining time: 0:47:18
130
+ Epoch 1 | iter 896 step 28 | loss train: 1.441, val: 1.515 | iter time: 359.76 ms (step) remaining time: 0:47:07
131
+ Epoch 1 | iter 928 step 29 | loss train: 1.427, val: 1.515 | iter time: 358.88 ms (step) remaining time: 0:46:56
132
+ Epoch 1 | iter 960 step 30 | loss train: 1.467, val: 1.515 | iter time: 359.25 ms (step) remaining time: 0:46:45
133
+ Epoch 1 | iter 992 step 31 | loss train: 1.430, val: 1.515 | iter time: 359.50 ms (step) remaining time: 0:46:34
134
+ Epoch 1 | iter 1024 step 32 | loss train: 1.436, val: 1.515 | iter time: 359.80 ms (step) remaining time: 0:46:24
135
+ Epoch 1 | iter 1056 step 33 | loss train: 1.396, val: 1.515 | iter time: 360.07 ms (step) remaining time: 0:46:13
136
+ Epoch 1 | iter 1088 step 34 | loss train: 1.380, val: 1.515 | iter time: 359.33 ms (step) remaining time: 0:46:02
137
+ Epoch 1 | iter 1120 step 35 | loss train: 1.468, val: 1.515 | iter time: 358.38 ms (step) remaining time: 0:45:51
138
+ Epoch 1 | iter 1152 step 36 | loss train: 1.416, val: 1.515 | iter time: 359.15 ms (step) remaining time: 0:45:40
139
+ Epoch 1 | iter 1184 step 37 | loss train: 1.398, val: 1.515 | iter time: 359.93 ms (step) remaining time: 0:45:29
140
+ Epoch 1 | iter 1216 step 38 | loss train: 1.409, val: 1.515 | iter time: 357.99 ms (step) remaining time: 0:45:18
141
+ Epoch 1 | iter 1248 step 39 | loss train: 1.389, val: 1.515 | iter time: 736.23 ms (step) remaining time: 0:45:10
142
+ Epoch 1 | iter 1280 step 40 | loss train: 1.404, val: 1.515 | iter time: 360.93 ms (step) remaining time: 0:44:59
143
+ Epoch 1 | iter 1312 step 41 | loss train: 1.405, val: 1.515 | iter time: 361.51 ms (step) remaining time: 0:44:48
144
+ Epoch 1 | iter 1344 step 42 | loss train: 1.451, val: 1.515 | iter time: 360.00 ms (step) remaining time: 0:44:37
145
+ Epoch 1 | iter 1376 step 43 | loss train: 1.460, val: 1.515 | iter time: 359.28 ms (step) remaining time: 0:44:26
146
+ Epoch 1 | iter 1408 step 44 | loss train: 1.472, val: 1.515 | iter time: 358.33 ms (step) remaining time: 0:44:14
147
+ Epoch 1 | iter 1440 step 45 | loss train: 1.415, val: 1.515 | iter time: 361.15 ms (step) remaining time: 0:44:03
148
+ Epoch 1 | iter 1472 step 46 | loss train: 1.460, val: 1.515 | iter time: 359.96 ms (step) remaining time: 0:43:52
149
+ Epoch 1 | iter 1504 step 47 | loss train: 1.489, val: 1.515 | iter time: 359.96 ms (step) remaining time: 0:43:41
150
+ Epoch 1 | iter 1536 step 48 | loss train: 1.385, val: 1.515 | iter time: 357.90 ms (step) remaining time: 0:43:30
151
+ Epoch 1 | iter 1568 step 49 | loss train: 1.511, val: 1.515 | iter time: 359.73 ms (step) remaining time: 0:43:19
152
+ Epoch 1 | iter 1600 step 50 | loss train: 1.442, val: 1.515 | iter time: 357.52 ms (step) remaining time: 0:43:08
153
+ Validating ...
154
+ iter 1600: val loss 1.6045, val time: 11279.09 ms
155
+ Epoch 1 | iter 1632 step 51 | loss train: 1.426, val: 1.604 | iter time: 359.26 ms (step) remaining time: 0:43:50
156
+ Epoch 1 | iter 1664 step 52 | loss train: 1.486, val: 1.604 | iter time: 360.42 ms (step) remaining time: 0:43:38
157
+ Epoch 1 | iter 1696 step 53 | loss train: 1.418, val: 1.604 | iter time: 359.57 ms (step) remaining time: 0:43:25
158
+ Epoch 1 | iter 1728 step 54 | loss train: 1.393, val: 1.604 | iter time: 360.09 ms (step) remaining time: 0:43:13
159
+ Epoch 1 | iter 1760 step 55 | loss train: 1.424, val: 1.604 | iter time: 357.67 ms (step) remaining time: 0:43:01
160
+ Epoch 1 | iter 1792 step 56 | loss train: 1.438, val: 1.604 | iter time: 361.34 ms (step) remaining time: 0:42:49
161
+ Epoch 1 | iter 1824 step 57 | loss train: 1.448, val: 1.604 | iter time: 358.53 ms (step) remaining time: 0:42:37
162
+ Epoch 1 | iter 1856 step 58 | loss train: 1.469, val: 1.604 | iter time: 360.29 ms (step) remaining time: 0:42:25
163
+ Epoch 1 | iter 1888 step 59 | loss train: 1.441, val: 1.604 | iter time: 359.49 ms (step) remaining time: 0:42:13
164
+ Epoch 1 | iter 1920 step 60 | loss train: 1.403, val: 1.604 | iter time: 358.95 ms (step) remaining time: 0:42:02
165
+ Epoch 1 | iter 1952 step 61 | loss train: 1.517, val: 1.604 | iter time: 359.77 ms (step) remaining time: 0:41:50
166
+ Epoch 1 | iter 1984 step 62 | loss train: 1.388, val: 1.604 | iter time: 360.10 ms (step) remaining time: 0:41:38
167
+ Epoch 1 | iter 2016 step 63 | loss train: 1.500, val: 1.604 | iter time: 358.88 ms (step) remaining time: 0:41:26
168
+ Epoch 1 | iter 2048 step 64 | loss train: 1.350, val: 1.604 | iter time: 359.68 ms (step) remaining time: 0:41:14
169
+ Epoch 1 | iter 2080 step 65 | loss train: 1.486, val: 1.604 | iter time: 360.35 ms (step) remaining time: 0:41:03
170
+ Epoch 1 | iter 2112 step 66 | loss train: 1.392, val: 1.604 | iter time: 358.01 ms (step) remaining time: 0:40:51
171
+ Epoch 1 | iter 2144 step 67 | loss train: 1.399, val: 1.604 | iter time: 359.99 ms (step) remaining time: 0:40:39
172
+ Epoch 1 | iter 2176 step 68 | loss train: 1.407, val: 1.604 | iter time: 358.84 ms (step) remaining time: 0:40:28
173
+ Epoch 1 | iter 2208 step 69 | loss train: 1.447, val: 1.604 | iter time: 360.85 ms (step) remaining time: 0:40:16
174
+ Epoch 1 | iter 2240 step 70 | loss train: 1.491, val: 1.604 | iter time: 358.11 ms (step) remaining time: 0:40:05
175
+ Epoch 1 | iter 2272 step 71 | loss train: 1.385, val: 1.604 | iter time: 360.35 ms (step) remaining time: 0:39:53
176
+ Epoch 1 | iter 2304 step 72 | loss train: 1.379, val: 1.604 | iter time: 359.89 ms (step) remaining time: 0:39:41
177
+ Epoch 1 | iter 2336 step 73 | loss train: 1.394, val: 1.604 | iter time: 359.85 ms (step) remaining time: 0:39:30
178
+ Epoch 1 | iter 2368 step 74 | loss train: 1.458, val: 1.604 | iter time: 360.69 ms (step) remaining time: 0:39:18
179
+ Epoch 1 | iter 2400 step 75 | loss train: 1.439, val: 1.604 | iter time: 358.82 ms (step) remaining time: 0:39:07
180
+ Epoch 1 | iter 2432 step 76 | loss train: 1.379, val: 1.604 | iter time: 357.71 ms (step) remaining time: 0:38:55
181
+ Epoch 1 | iter 2464 step 77 | loss train: 1.413, val: 1.604 | iter time: 360.82 ms (step) remaining time: 0:38:44
182
+ Epoch 1 | iter 2496 step 78 | loss train: 1.356, val: 1.604 | iter time: 359.71 ms (step) remaining time: 0:38:32
183
+ Epoch 1 | iter 2528 step 79 | loss train: 1.404, val: 1.604 | iter time: 359.23 ms (step) remaining time: 0:38:21
184
+ Epoch 1 | iter 2560 step 80 | loss train: 1.459, val: 1.604 | iter time: 359.05 ms (step) remaining time: 0:38:10
185
+ Epoch 1 | iter 2592 step 81 | loss train: 1.402, val: 1.604 | iter time: 358.61 ms (step) remaining time: 0:37:58
186
+ Epoch 1 | iter 2624 step 82 | loss train: 1.393, val: 1.604 | iter time: 360.20 ms (step) remaining time: 0:37:47
187
+ Epoch 1 | iter 2656 step 83 | loss train: 1.351, val: 1.604 | iter time: 359.56 ms (step) remaining time: 0:37:36
188
+ Epoch 1 | iter 2688 step 84 | loss train: 1.379, val: 1.604 | iter time: 359.61 ms (step) remaining time: 0:37:24
189
+ Epoch 1 | iter 2720 step 85 | loss train: 1.482, val: 1.604 | iter time: 360.17 ms (step) remaining time: 0:37:13
190
+ Epoch 1 | iter 2752 step 86 | loss train: 1.417, val: 1.604 | iter time: 358.16 ms (step) remaining time: 0:37:02
191
+ Epoch 1 | iter 2784 step 87 | loss train: 1.367, val: 1.604 | iter time: 359.33 ms (step) remaining time: 0:36:50
192
+ Epoch 1 | iter 2816 step 88 | loss train: 1.409, val: 1.604 | iter time: 357.66 ms (step) remaining time: 0:36:39
193
+ Epoch 1 | iter 2848 step 89 | loss train: 1.381, val: 1.604 | iter time: 359.77 ms (step) remaining time: 0:36:28
194
+ Epoch 1 | iter 2880 step 90 | loss train: 1.395, val: 1.604 | iter time: 357.83 ms (step) remaining time: 0:36:17
195
+ Epoch 1 | iter 2912 step 91 | loss train: 1.470, val: 1.604 | iter time: 358.83 ms (step) remaining time: 0:36:06
196
+ Epoch 1 | iter 2944 step 92 | loss train: 1.402, val: 1.604 | iter time: 359.46 ms (step) remaining time: 0:35:55
197
+ Epoch 1 | iter 2976 step 93 | loss train: 1.374, val: 1.604 | iter time: 360.89 ms (step) remaining time: 0:35:43
198
+ Epoch 1 | iter 3008 step 94 | loss train: 1.374, val: 1.604 | iter time: 359.28 ms (step) remaining time: 0:35:32
199
+ Epoch 1 | iter 3040 step 95 | loss train: 1.398, val: 1.604 | iter time: 359.69 ms (step) remaining time: 0:35:21
200
+ Epoch 1 | iter 3072 step 96 | loss train: 1.415, val: 1.604 | iter time: 358.53 ms (step) remaining time: 0:35:10
201
+ Epoch 1 | iter 3104 step 97 | loss train: 1.340, val: 1.604 | iter time: 359.64 ms (step) remaining time: 0:34:58
202
+ Epoch 1 | iter 3136 step 98 | loss train: 1.433, val: 1.604 | iter time: 359.08 ms (step) remaining time: 0:34:47
203
+ Epoch 1 | iter 3168 step 99 | loss train: 1.429, val: 1.604 | iter time: 361.53 ms (step) remaining time: 0:34:36
204
+ Epoch 1 | iter 3200 step 100 | loss train: 1.471, val: 1.604 | iter time: 359.28 ms (step) remaining time: 0:34:25
205
+ Validating ...
206
+ iter 3200: val loss 1.5566, val time: 11290.77 ms
207
+ Saving checkpoint to 'out/pretrain/2412/step-00000100/lit_model.pth'
208
+ Epoch 1 | iter 3232 step 101 | loss train: 1.348, val: 1.557 | iter time: 355.70 ms (step) remaining time: 0:35:05
209
+ Epoch 1 | iter 3264 step 102 | loss train: 1.418, val: 1.557 | iter time: 359.22 ms (step) remaining time: 0:34:53
210
+ Epoch 1 | iter 3296 step 103 | loss train: 1.400, val: 1.557 | iter time: 360.08 ms (step) remaining time: 0:34:41
211
+ Epoch 1 | iter 3328 step 104 | loss train: 1.394, val: 1.557 | iter time: 360.18 ms (step) remaining time: 0:34:29
212
+ Epoch 1 | iter 3360 step 105 | loss train: 1.484, val: 1.557 | iter time: 359.43 ms (step) remaining time: 0:34:17
213
+ Epoch 1 | iter 3392 step 106 | loss train: 1.481, val: 1.557 | iter time: 359.06 ms (step) remaining time: 0:34:05
214
+ Epoch 1 | iter 3424 step 107 | loss train: 1.436, val: 1.557 | iter time: 358.11 ms (step) remaining time: 0:33:53
215
+ Epoch 1 | iter 3456 step 108 | loss train: 1.422, val: 1.557 | iter time: 357.37 ms (step) remaining time: 0:33:41
216
+ Epoch 1 | iter 3488 step 109 | loss train: 1.392, val: 1.557 | iter time: 362.25 ms (step) remaining time: 0:33:30
217
+ Epoch 1 | iter 3520 step 110 | loss train: 1.430, val: 1.557 | iter time: 358.26 ms (step) remaining time: 0:33:18
218
+ Epoch 1 | iter 3552 step 111 | loss train: 1.469, val: 1.557 | iter time: 358.26 ms (step) remaining time: 0:33:06
219
+ Epoch 1 | iter 3584 step 112 | loss train: 1.414, val: 1.557 | iter time: 359.00 ms (step) remaining time: 0:32:54
220
+ Epoch 1 | iter 3616 step 113 | loss train: 1.365, val: 1.557 | iter time: 359.17 ms (step) remaining time: 0:32:42
221
+ Epoch 1 | iter 3648 step 114 | loss train: 1.386, val: 1.557 | iter time: 358.07 ms (step) remaining time: 0:32:31
222
+ Epoch 1 | iter 3680 step 115 | loss train: 1.385, val: 1.557 | iter time: 360.50 ms (step) remaining time: 0:32:19
223
+ Epoch 1 | iter 3712 step 116 | loss train: 1.406, val: 1.557 | iter time: 360.24 ms (step) remaining time: 0:32:07
224
+ Epoch 1 | iter 3744 step 117 | loss train: 1.431, val: 1.557 | iter time: 360.89 ms (step) remaining time: 0:31:55
225
+ Epoch 1 | iter 3776 step 118 | loss train: 1.428, val: 1.557 | iter time: 358.67 ms (step) remaining time: 0:31:44
226
+ Epoch 1 | iter 3808 step 119 | loss train: 1.400, val: 1.557 | iter time: 359.95 ms (step) remaining time: 0:31:32
227
+ Epoch 1 | iter 3840 step 120 | loss train: 1.331, val: 1.557 | iter time: 358.79 ms (step) remaining time: 0:31:20
228
+ Epoch 1 | iter 3872 step 121 | loss train: 1.394, val: 1.557 | iter time: 360.93 ms (step) remaining time: 0:31:09
229
+ Epoch 1 | iter 3904 step 122 | loss train: 1.409, val: 1.557 | iter time: 360.29 ms (step) remaining time: 0:30:57
230
+ Epoch 1 | iter 3936 step 123 | loss train: 1.389, val: 1.557 | iter time: 358.96 ms (step) remaining time: 0:30:45
231
+ Epoch 1 | iter 3968 step 124 | loss train: 1.381, val: 1.557 | iter time: 360.04 ms (step) remaining time: 0:30:34
232
+ Epoch 1 | iter 4000 step 125 | loss train: 1.441, val: 1.557 | iter time: 359.37 ms (step) remaining time: 0:30:22
233
+ Epoch 1 | iter 4032 step 126 | loss train: 1.344, val: 1.557 | iter time: 358.96 ms (step) remaining time: 0:30:11
234
+ Epoch 1 | iter 4064 step 127 | loss train: 1.416, val: 1.557 | iter time: 359.99 ms (step) remaining time: 0:29:59
235
+ Epoch 1 | iter 4096 step 128 | loss train: 1.417, val: 1.557 | iter time: 359.98 ms (step) remaining time: 0:29:47
236
+ Epoch 1 | iter 4128 step 129 | loss train: 1.412, val: 1.557 | iter time: 359.21 ms (step) remaining time: 0:29:36
237
+ Epoch 1 | iter 4160 step 130 | loss train: 1.436, val: 1.557 | iter time: 357.40 ms (step) remaining time: 0:29:24
238
+ Epoch 1 | iter 4192 step 131 | loss train: 1.406, val: 1.557 | iter time: 359.45 ms (step) remaining time: 0:29:13
239
+ Epoch 1 | iter 4224 step 132 | loss train: 1.419, val: 1.557 | iter time: 358.23 ms (step) remaining time: 0:29:01
240
+ Epoch 1 | iter 4256 step 133 | loss train: 1.460, val: 1.557 | iter time: 359.71 ms (step) remaining time: 0:28:50
241
+ Epoch 1 | iter 4288 step 134 | loss train: 1.408, val: 1.557 | iter time: 359.28 ms (step) remaining time: 0:28:38
242
+ Epoch 1 | iter 4320 step 135 | loss train: 1.436, val: 1.557 | iter time: 357.52 ms (step) remaining time: 0:28:27
243
+ Epoch 1 | iter 4352 step 136 | loss train: 1.368, val: 1.557 | iter time: 360.51 ms (step) remaining time: 0:28:15
244
+ Epoch 1 | iter 4384 step 137 | loss train: 1.351, val: 1.557 | iter time: 359.69 ms (step) remaining time: 0:28:04
245
+ Epoch 1 | iter 4416 step 138 | loss train: 1.391, val: 1.557 | iter time: 359.32 ms (step) remaining time: 0:27:52
246
+ Epoch 1 | iter 4448 step 139 | loss train: 1.408, val: 1.557 | iter time: 360.94 ms (step) remaining time: 0:27:41
247
+ Epoch 1 | iter 4480 step 140 | loss train: 1.455, val: 1.557 | iter time: 360.43 ms (step) remaining time: 0:27:29
248
+ Epoch 1 | iter 4512 step 141 | loss train: 1.324, val: 1.557 | iter time: 358.18 ms (step) remaining time: 0:27:18
249
+ Epoch 1 | iter 4544 step 142 | loss train: 1.365, val: 1.557 | iter time: 358.46 ms (step) remaining time: 0:27:07
250
+ Epoch 1 | iter 4576 step 143 | loss train: 1.423, val: 1.557 | iter time: 358.26 ms (step) remaining time: 0:26:56
251
+ Epoch 1 | iter 4608 step 144 | loss train: 1.355, val: 1.557 | iter time: 357.67 ms (step) remaining time: 0:26:44
252
+ Epoch 1 | iter 4640 step 145 | loss train: 1.326, val: 1.557 | iter time: 360.79 ms (step) remaining time: 0:26:33
253
+ Epoch 1 | iter 4672 step 146 | loss train: 1.410, val: 1.557 | iter time: 361.28 ms (step) remaining time: 0:26:21
254
+ Epoch 1 | iter 4704 step 147 | loss train: 1.376, val: 1.557 | iter time: 359.61 ms (step) remaining time: 0:26:10
255
+ Epoch 1 | iter 4736 step 148 | loss train: 1.410, val: 1.557 | iter time: 358.55 ms (step) remaining time: 0:25:58
256
+ Epoch 1 | iter 4768 step 149 | loss train: 1.353, val: 1.557 | iter time: 357.80 ms (step) remaining time: 0:25:47
257
+ Epoch 1 | iter 4800 step 150 | loss train: 1.361, val: 1.557 | iter time: 360.19 ms (step) remaining time: 0:25:36
258
+ Validating ...
259
+ iter 4800: val loss 1.5478, val time: 11291.47 ms
260
+ Epoch 1 | iter 4832 step 151 | loss train: 1.345, val: 1.548 | iter time: 359.93 ms (step) remaining time: 0:25:35
261
+ Epoch 1 | iter 4864 step 152 | loss train: 1.329, val: 1.548 | iter time: 361.09 ms (step) remaining time: 0:25:23
262
+ Epoch 1 | iter 4896 step 153 | loss train: 1.380, val: 1.548 | iter time: 359.32 ms (step) remaining time: 0:25:12
263
+ Epoch 1 | iter 4928 step 154 | loss train: 1.323, val: 1.548 | iter time: 360.58 ms (step) remaining time: 0:25:00
264
+ Epoch 1 | iter 4960 step 155 | loss train: 1.432, val: 1.548 | iter time: 360.24 ms (step) remaining time: 0:24:49
265
+ Epoch 1 | iter 4992 step 156 | loss train: 1.333, val: 1.548 | iter time: 358.40 ms (step) remaining time: 0:24:37
266
+ Epoch 1 | iter 5024 step 157 | loss train: 1.420, val: 1.548 | iter time: 359.48 ms (step) remaining time: 0:24:26
267
+ Epoch 1 | iter 5056 step 158 | loss train: 1.411, val: 1.548 | iter time: 359.26 ms (step) remaining time: 0:24:14
268
+ Epoch 1 | iter 5088 step 159 | loss train: 1.357, val: 1.548 | iter time: 356.81 ms (step) remaining time: 0:24:03
269
+ Epoch 1 | iter 5120 step 160 | loss train: 1.417, val: 1.548 | iter time: 360.77 ms (step) remaining time: 0:23:51
270
+ Epoch 1 | iter 5152 step 161 | loss train: 1.360, val: 1.548 | iter time: 358.44 ms (step) remaining time: 0:23:40
271
+ Epoch 1 | iter 5184 step 162 | loss train: 1.403, val: 1.548 | iter time: 359.83 ms (step) remaining time: 0:23:28
272
+ Epoch 1 | iter 5216 step 163 | loss train: 1.412, val: 1.548 | iter time: 359.03 ms (step) remaining time: 0:23:17
273
+ Epoch 1 | iter 5248 step 164 | loss train: 1.378, val: 1.548 | iter time: 361.43 ms (step) remaining time: 0:23:05
274
+ Epoch 1 | iter 5280 step 165 | loss train: 1.328, val: 1.548 | iter time: 358.02 ms (step) remaining time: 0:22:54
275
+ Epoch 1 | iter 5312 step 166 | loss train: 1.311, val: 1.548 | iter time: 359.26 ms (step) remaining time: 0:22:43
276
+ Epoch 1 | iter 5344 step 167 | loss train: 1.428, val: 1.548 | iter time: 359.76 ms (step) remaining time: 0:22:31
277
+ Epoch 1 | iter 5376 step 168 | loss train: 1.457, val: 1.548 | iter time: 361.20 ms (step) remaining time: 0:22:20
278
+ Epoch 1 | iter 5408 step 169 | loss train: 1.413, val: 1.548 | iter time: 358.80 ms (step) remaining time: 0:22:08
279
+ Epoch 1 | iter 5440 step 170 | loss train: 1.322, val: 1.548 | iter time: 359.81 ms (step) remaining time: 0:21:57
280
+ Epoch 1 | iter 5472 step 171 | loss train: 1.370, val: 1.548 | iter time: 358.45 ms (step) remaining time: 0:21:46
281
+ Epoch 1 | iter 5504 step 172 | loss train: 1.416, val: 1.548 | iter time: 360.15 ms (step) remaining time: 0:21:34
282
+ Epoch 1 | iter 5536 step 173 | loss train: 1.394, val: 1.548 | iter time: 360.54 ms (step) remaining time: 0:21:23
283
+ Epoch 1 | iter 5568 step 174 | loss train: 1.362, val: 1.548 | iter time: 358.27 ms (step) remaining time: 0:21:12
284
+ Epoch 1 | iter 5600 step 175 | loss train: 1.391, val: 1.548 | iter time: 360.58 ms (step) remaining time: 0:21:00
285
+ Epoch 1 | iter 5632 step 176 | loss train: 1.385, val: 1.548 | iter time: 358.95 ms (step) remaining time: 0:20:49
286
+ Epoch 1 | iter 5664 step 177 | loss train: 1.388, val: 1.548 | iter time: 359.80 ms (step) remaining time: 0:20:38
287
+ Epoch 1 | iter 5696 step 178 | loss train: 1.350, val: 1.548 | iter time: 359.06 ms (step) remaining time: 0:20:26
288
+ Epoch 1 | iter 5728 step 179 | loss train: 1.408, val: 1.548 | iter time: 357.48 ms (step) remaining time: 0:20:15
289
+ Epoch 1 | iter 5760 step 180 | loss train: 1.395, val: 1.548 | iter time: 359.03 ms (step) remaining time: 0:20:04
290
+ Epoch 1 | iter 5792 step 181 | loss train: 1.325, val: 1.548 | iter time: 359.54 ms (step) remaining time: 0:19:52
291
+ Epoch 1 | iter 5824 step 182 | loss train: 1.400, val: 1.548 | iter time: 359.30 ms (step) remaining time: 0:19:41
292
+ Epoch 1 | iter 5856 step 183 | loss train: 1.367, val: 1.548 | iter time: 360.67 ms (step) remaining time: 0:19:30
293
+ Epoch 1 | iter 5888 step 184 | loss train: 1.324, val: 1.548 | iter time: 360.37 ms (step) remaining time: 0:19:18
294
+ Epoch 1 | iter 5920 step 185 | loss train: 1.377, val: 1.548 | iter time: 361.38 ms (step) remaining time: 0:19:07
295
+ Epoch 1 | iter 5952 step 186 | loss train: 1.345, val: 1.548 | iter time: 360.53 ms (step) remaining time: 0:18:56
296
+ Epoch 1 | iter 5984 step 187 | loss train: 1.424, val: 1.548 | iter time: 357.06 ms (step) remaining time: 0:18:45
297
+ Epoch 1 | iter 6016 step 188 | loss train: 1.313, val: 1.548 | iter time: 361.73 ms (step) remaining time: 0:18:33
298
+ Epoch 1 | iter 6048 step 189 | loss train: 1.412, val: 1.548 | iter time: 359.61 ms (step) remaining time: 0:18:22
299
+ Epoch 1 | iter 6080 step 190 | loss train: 1.355, val: 1.548 | iter time: 360.04 ms (step) remaining time: 0:18:11
300
+ Epoch 1 | iter 6112 step 191 | loss train: 1.322, val: 1.548 | iter time: 359.21 ms (step) remaining time: 0:17:59
301
+ Epoch 1 | iter 6144 step 192 | loss train: 1.365, val: 1.548 | iter time: 359.09 ms (step) remaining time: 0:17:48
302
+ Epoch 1 | iter 6176 step 193 | loss train: 1.329, val: 1.548 | iter time: 357.82 ms (step) remaining time: 0:17:37
303
+ Epoch 1 | iter 6208 step 194 | loss train: 1.402, val: 1.548 | iter time: 360.41 ms (step) remaining time: 0:17:26
304
+ Epoch 1 | iter 6240 step 195 | loss train: 1.307, val: 1.548 | iter time: 358.97 ms (step) remaining time: 0:17:15
305
+ Epoch 1 | iter 6272 step 196 | loss train: 1.265, val: 1.548 | iter time: 358.39 ms (step) remaining time: 0:17:03
306
+ Epoch 1 | iter 6304 step 197 | loss train: 1.385, val: 1.548 | iter time: 358.00 ms (step) remaining time: 0:16:52
307
+ Epoch 1 | iter 6336 step 198 | loss train: 1.375, val: 1.548 | iter time: 358.06 ms (step) remaining time: 0:16:41
308
+ Epoch 1 | iter 6368 step 199 | loss train: 1.381, val: 1.548 | iter time: 360.08 ms (step) remaining time: 0:16:30
309
+ Epoch 1 | iter 6400 step 200 | loss train: 1.368, val: 1.548 | iter time: 361.09 ms (step) remaining time: 0:16:18
310
+ Validating ...
311
+ iter 6400: val loss 1.5167, val time: 11312.35 ms
312
+ Saving checkpoint to 'out/pretrain/2412/step-00000200/lit_model.pth'
313
+ Epoch 1 | iter 6432 step 201 | loss train: 1.359, val: 1.517 | iter time: 357.64 ms (step) remaining time: 0:16:19
314
+ Epoch 1 | iter 6464 step 202 | loss train: 1.352, val: 1.517 | iter time: 357.31 ms (step) remaining time: 0:16:08
315
+ Epoch 1 | iter 6496 step 203 | loss train: 1.301, val: 1.517 | iter time: 357.70 ms (step) remaining time: 0:15:56
316
+ Epoch 1 | iter 6528 step 204 | loss train: 1.306, val: 1.517 | iter time: 357.61 ms (step) remaining time: 0:15:45
317
+ Epoch 1 | iter 6560 step 205 | loss train: 1.393, val: 1.517 | iter time: 367.74 ms (step) remaining time: 0:15:33
318
+ Epoch 1 | iter 6592 step 206 | loss train: 1.352, val: 1.517 | iter time: 358.15 ms (step) remaining time: 0:15:22
319
+ Epoch 1 | iter 6624 step 207 | loss train: 1.411, val: 1.517 | iter time: 358.99 ms (step) remaining time: 0:15:11
320
+ Epoch 1 | iter 6656 step 208 | loss train: 1.366, val: 1.517 | iter time: 358.44 ms (step) remaining time: 0:14:59
321
+ Epoch 1 | iter 6688 step 209 | loss train: 1.313, val: 1.517 | iter time: 358.03 ms (step) remaining time: 0:14:48
322
+ Epoch 1 | iter 6720 step 210 | loss train: 1.350, val: 1.517 | iter time: 359.49 ms (step) remaining time: 0:14:36
323
+ Epoch 1 | iter 6752 step 211 | loss train: 1.347, val: 1.517 | iter time: 359.26 ms (step) remaining time: 0:14:25
324
+ Epoch 1 | iter 6784 step 212 | loss train: 1.426, val: 1.517 | iter time: 359.41 ms (step) remaining time: 0:14:14
325
+ Epoch 1 | iter 6816 step 213 | loss train: 1.344, val: 1.517 | iter time: 361.01 ms (step) remaining time: 0:14:02
326
+ Epoch 1 | iter 6848 step 214 | loss train: 1.337, val: 1.517 | iter time: 357.72 ms (step) remaining time: 0:13:51
327
+ Epoch 1 | iter 6880 step 215 | loss train: 1.410, val: 1.517 | iter time: 358.95 ms (step) remaining time: 0:13:40
328
+ Epoch 1 | iter 6912 step 216 | loss train: 1.297, val: 1.517 | iter time: 359.36 ms (step) remaining time: 0:13:28
329
+ Epoch 1 | iter 6944 step 217 | loss train: 1.370, val: 1.517 | iter time: 359.56 ms (step) remaining time: 0:13:17
330
+ Epoch 1 | iter 6976 step 218 | loss train: 1.322, val: 1.517 | iter time: 361.88 ms (step) remaining time: 0:13:05
331
+ Epoch 1 | iter 7008 step 219 | loss train: 1.346, val: 1.517 | iter time: 359.38 ms (step) remaining time: 0:12:54
332
+ Epoch 1 | iter 7040 step 220 | loss train: 1.328, val: 1.517 | iter time: 361.68 ms (step) remaining time: 0:12:43
333
+ Epoch 1 | iter 7072 step 221 | loss train: 1.312, val: 1.517 | iter time: 359.43 ms (step) remaining time: 0:12:31
334
+ Epoch 1 | iter 7104 step 222 | loss train: 1.336, val: 1.517 | iter time: 358.97 ms (step) remaining time: 0:12:20
335
+ Epoch 1 | iter 7136 step 223 | loss train: 1.351, val: 1.517 | iter time: 358.65 ms (step) remaining time: 0:12:09
336
+ Epoch 1 | iter 7168 step 224 | loss train: 1.311, val: 1.517 | iter time: 359.52 ms (step) remaining time: 0:11:57
337
+ Epoch 1 | iter 7200 step 225 | loss train: 1.326, val: 1.517 | iter time: 359.56 ms (step) remaining time: 0:11:46
338
+ Epoch 1 | iter 7232 step 226 | loss train: 1.391, val: 1.517 | iter time: 357.92 ms (step) remaining time: 0:11:35
339
+ Epoch 1 | iter 7264 step 227 | loss train: 1.327, val: 1.517 | iter time: 361.14 ms (step) remaining time: 0:11:24
340
+ Epoch 1 | iter 7296 step 228 | loss train: 1.391, val: 1.517 | iter time: 360.86 ms (step) remaining time: 0:11:12
341
+ Epoch 1 | iter 7328 step 229 | loss train: 1.355, val: 1.517 | iter time: 360.88 ms (step) remaining time: 0:11:01
342
+ Epoch 1 | iter 7360 step 230 | loss train: 1.387, val: 1.517 | iter time: 360.25 ms (step) remaining time: 0:10:50
343
+ Epoch 1 | iter 7392 step 231 | loss train: 1.347, val: 1.517 | iter time: 360.12 ms (step) remaining time: 0:10:38
344
+ Epoch 1 | iter 7424 step 232 | loss train: 1.324, val: 1.517 | iter time: 359.50 ms (step) remaining time: 0:10:27
345
+ Epoch 1 | iter 7456 step 233 | loss train: 1.345, val: 1.517 | iter time: 361.46 ms (step) remaining time: 0:10:16
346
+ Epoch 1 | iter 7488 step 234 | loss train: 1.301, val: 1.517 | iter time: 360.08 ms (step) remaining time: 0:10:04
347
+ Epoch 1 | iter 7520 step 235 | loss train: 1.383, val: 1.517 | iter time: 360.12 ms (step) remaining time: 0:09:53
348
+ Epoch 1 | iter 7552 step 236 | loss train: 1.362, val: 1.517 | iter time: 359.22 ms (step) remaining time: 0:09:42
349
+ Epoch 1 | iter 7584 step 237 | loss train: 1.446, val: 1.517 | iter time: 358.76 ms (step) remaining time: 0:09:31
350
+ Epoch 1 | iter 7616 step 238 | loss train: 1.408, val: 1.517 | iter time: 358.99 ms (step) remaining time: 0:09:19
351
+ Epoch 1 | iter 7648 step 239 | loss train: 1.319, val: 1.517 | iter time: 357.99 ms (step) remaining time: 0:09:08
352
+ Epoch 1 | iter 7680 step 240 | loss train: 1.305, val: 1.517 | iter time: 359.95 ms (step) remaining time: 0:08:57
353
+ Epoch 1 | iter 7712 step 241 | loss train: 1.373, val: 1.517 | iter time: 360.20 ms (step) remaining time: 0:08:46
354
+ Epoch 1 | iter 7744 step 242 | loss train: 1.302, val: 1.517 | iter time: 358.69 ms (step) remaining time: 0:08:34
355
+ Epoch 1 | iter 7776 step 243 | loss train: 1.306, val: 1.517 | iter time: 360.55 ms (step) remaining time: 0:08:23
356
+ Epoch 1 | iter 7808 step 244 | loss train: 1.351, val: 1.517 | iter time: 361.13 ms (step) remaining time: 0:08:12
357
+ Epoch 1 | iter 7840 step 245 | loss train: 1.318, val: 1.517 | iter time: 358.61 ms (step) remaining time: 0:08:01
358
+ Epoch 1 | iter 7872 step 246 | loss train: 1.363, val: 1.517 | iter time: 359.24 ms (step) remaining time: 0:07:49
359
+ Epoch 1 | iter 7904 step 247 | loss train: 1.371, val: 1.517 | iter time: 361.89 ms (step) remaining time: 0:07:38
360
+ Epoch 1 | iter 7936 step 248 | loss train: 1.304, val: 1.517 | iter time: 361.06 ms (step) remaining time: 0:07:27
361
+ Epoch 1 | iter 7968 step 249 | loss train: 1.301, val: 1.517 | iter time: 359.34 ms (step) remaining time: 0:07:16
362
+ Epoch 1 | iter 8000 step 250 | loss train: 1.362, val: 1.517 | iter time: 360.73 ms (step) remaining time: 0:07:04
363
+ Validating ...
364
+ iter 8000: val loss 1.5129, val time: 11296.47 ms
365
+ Epoch 1 | iter 8032 step 251 | loss train: 1.402, val: 1.513 | iter time: 358.22 ms (step) remaining time: 0:06:55
366
+ Epoch 1 | iter 8064 step 252 | loss train: 1.334, val: 1.513 | iter time: 358.87 ms (step) remaining time: 0:06:44
367
+ Epoch 1 | iter 8096 step 253 | loss train: 1.387, val: 1.513 | iter time: 360.21 ms (step) remaining time: 0:06:32
368
+ Epoch 1 | iter 8128 step 254 | loss train: 1.346, val: 1.513 | iter time: 358.89 ms (step) remaining time: 0:06:21
369
+ Epoch 1 | iter 8160 step 255 | loss train: 1.288, val: 1.513 | iter time: 359.06 ms (step) remaining time: 0:06:10
370
+ Epoch 1 | iter 8192 step 256 | loss train: 1.320, val: 1.513 | iter time: 357.70 ms (step) remaining time: 0:05:59
371
+ Epoch 1 | iter 8224 step 257 | loss train: 1.288, val: 1.513 | iter time: 359.90 ms (step) remaining time: 0:05:47
372
+ Epoch 1 | iter 8256 step 258 | loss train: 1.400, val: 1.513 | iter time: 358.76 ms (step) remaining time: 0:05:36
373
+ Epoch 1 | iter 8288 step 259 | loss train: 1.351, val: 1.513 | iter time: 358.20 ms (step) remaining time: 0:05:25
374
+ Epoch 1 | iter 8320 step 260 | loss train: 1.350, val: 1.513 | iter time: 358.80 ms (step) remaining time: 0:05:14
375
+ Epoch 1 | iter 8352 step 261 | loss train: 1.428, val: 1.513 | iter time: 359.16 ms (step) remaining time: 0:05:02
376
+ Epoch 1 | iter 8384 step 262 | loss train: 1.390, val: 1.513 | iter time: 360.59 ms (step) remaining time: 0:04:51
377
+ Epoch 1 | iter 8416 step 263 | loss train: 1.396, val: 1.513 | iter time: 358.98 ms (step) remaining time: 0:04:40
378
+ Epoch 1 | iter 8448 step 264 | loss train: 1.297, val: 1.513 | iter time: 360.47 ms (step) remaining time: 0:04:29
379
+ Epoch 1 | iter 8480 step 265 | loss train: 1.331, val: 1.513 | iter time: 358.98 ms (step) remaining time: 0:04:17
380
+ Epoch 1 | iter 8512 step 266 | loss train: 1.341, val: 1.513 | iter time: 360.22 ms (step) remaining time: 0:04:06
381
+ Epoch 1 | iter 8544 step 267 | loss train: 1.339, val: 1.513 | iter time: 360.93 ms (step) remaining time: 0:03:55
382
+ Epoch 1 | iter 8576 step 268 | loss train: 1.369, val: 1.513 | iter time: 359.87 ms (step) remaining time: 0:03:44
383
+ Epoch 1 | iter 8608 step 269 | loss train: 1.329, val: 1.513 | iter time: 359.81 ms (step) remaining time: 0:03:32
384
+ Epoch 1 | iter 8640 step 270 | loss train: 1.325, val: 1.513 | iter time: 359.98 ms (step) remaining time: 0:03:21
385
+ Epoch 1 | iter 8672 step 271 | loss train: 1.344, val: 1.513 | iter time: 358.67 ms (step) remaining time: 0:03:10
386
+ Epoch 1 | iter 8704 step 272 | loss train: 1.380, val: 1.513 | iter time: 359.40 ms (step) remaining time: 0:02:59
387
+ Epoch 1 | iter 8736 step 273 | loss train: 1.355, val: 1.513 | iter time: 359.79 ms (step) remaining time: 0:02:47
388
+ Epoch 1 | iter 8768 step 274 | loss train: 1.335, val: 1.513 | iter time: 357.99 ms (step) remaining time: 0:02:36
389
+ Epoch 1 | iter 8800 step 275 | loss train: 1.354, val: 1.513 | iter time: 360.61 ms (step) remaining time: 0:02:25
390
+ Epoch 1 | iter 8832 step 276 | loss train: 1.368, val: 1.513 | iter time: 360.65 ms (step) remaining time: 0:02:14
391
+ Epoch 1 | iter 8864 step 277 | loss train: 1.358, val: 1.513 | iter time: 358.75 ms (step) remaining time: 0:02:03
392
+ Epoch 1 | iter 8896 step 278 | loss train: 1.335, val: 1.513 | iter time: 360.56 ms (step) remaining time: 0:01:51
393
+ Epoch 1 | iter 8928 step 279 | loss train: 1.334, val: 1.513 | iter time: 359.23 ms (step) remaining time: 0:01:40
394
+ Epoch 1 | iter 8960 step 280 | loss train: 1.324, val: 1.513 | iter time: 360.47 ms (step) remaining time: 0:01:29
395
+ Epoch 1 | iter 8992 step 281 | loss train: 1.363, val: 1.513 | iter time: 360.08 ms (step) remaining time: 0:01:18
396
+ Epoch 1 | iter 9024 step 282 | loss train: 1.341, val: 1.513 | iter time: 359.93 ms (step) remaining time: 0:01:07
397
+ Epoch 1 | iter 9056 step 283 | loss train: 1.330, val: 1.513 | iter time: 359.77 ms (step) remaining time: 0:00:55
398
+ Epoch 1 | iter 9088 step 284 | loss train: 1.378, val: 1.513 | iter time: 359.65 ms (step) remaining time: 0:00:44
399
+ Epoch 1 | iter 9120 step 285 | loss train: 1.394, val: 1.513 | iter time: 360.87 ms (step) remaining time: 0:00:33
400
+ Epoch 1 | iter 9152 step 286 | loss train: 1.375, val: 1.513 | iter time: 358.83 ms (step) remaining time: 0:00:22
401
+ Epoch 1 | iter 9184 step 287 | loss train: 1.343, val: 1.513 | iter time: 358.85 ms (step) remaining time: 0:00:11
402
+ Epoch 1 | iter 9216 step 288 | loss train: 1.378, val: 1.513 | iter time: 357.81 ms (step) remaining time: 0:00:00
403
+ Validating ...
404
+ Final evaluation | val loss: 1.499 | val ppl: 4.477
405
+ Saving checkpoint to 'out/pretrain/2412/final/lit_model.pth'
406
+ ----------------------------------------
407
+ | Performance
408
+ | - Total tokens : 301,989,888
409
+ | - Training Time : 3261.91 s
410
+ | - Tok/sec : 269.31 tok/s
411
+ | ----------------------------------------
412
+ | Memory Usage
413
+ | - Memory Used : 26.32 GB
414
+ ----------------------------------------
out/pretrain/tinyllama/teelogs/2412_full.txt ADDED
@@ -0,0 +1,425 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Initializing distributed: GLOBAL_RANK: 0, MEMBER: 1/4
2
+ Initializing distributed: GLOBAL_RANK: 1, MEMBER: 2/4
3
+ Initializing distributed: GLOBAL_RANK: 2, MEMBER: 3/4
4
+ [rank: 2] Seed set to 42
5
+ [rank: 1] Seed set to 42
6
+ Initializing distributed: GLOBAL_RANK: 3, MEMBER: 4/4
7
+ [rank: 3] Seed set to 42
8
+ ----------------------------------------------------------------------------------------------------
9
+ distributed_backend=nccl
10
+ All distributed processes registered. Starting with 4 processes
11
+ ----------------------------------------------------------------------------------------------------
12
+
13
+ [rank: 0] Seed set to 42
14
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
15
+ train_dataloader = data.train_dataloader()
16
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
17
+ train_dataloader = data.train_dataloader()
18
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
19
+ train_dataloader = data.train_dataloader()
20
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
21
+ train_dataloader = data.train_dataloader()
22
+ All GPUs are fully connected via NVLink.
23
+ {'data': {'batch_size': 1,
24
+ 'data_path': PosixPath('litgpt/data/arxiv'),
25
+ 'num_workers': 8,
26
+ 'seed': 42,
27
+ 'seq_length': 2048,
28
+ 'use_starcoder': True},
29
+ 'data_dir': PosixPath('litgpt/data/arxiv/2412'),
30
+ 'devices': 'auto',
31
+ 'eval': {'evaluate_example': 'first',
32
+ 'final_validation': True,
33
+ 'initial_validation': True,
34
+ 'interval': 50,
35
+ 'max_iters': 200,
36
+ 'max_new_tokens': None},
37
+ 'initial_checkpoint_dir': PosixPath('out/pretrain/2411_full/final'),
38
+ 'log': {'group': None, 'project': None, 'run': None},
39
+ 'logger_name': 'tensorboard',
40
+ 'model_config': {'attention_logit_softcapping': None,
41
+ 'attention_scores_scalar': None,
42
+ 'attn_bias': False,
43
+ 'bias': False,
44
+ 'block_size': 2048,
45
+ 'final_logit_softcapping': None,
46
+ 'gelu_approximate': 'none',
47
+ 'head_size': 64,
48
+ 'hf_config': {'name': 'TinyLlama-1.1B-intermediate-step-1431k-3T',
49
+ 'org': 'TinyLlama'},
50
+ 'intermediate_size': 5632,
51
+ 'lm_head_bias': False,
52
+ 'mlp_class_name': 'LLaMAMLP',
53
+ 'moe_intermediate_size': None,
54
+ 'n_embd': 2048,
55
+ 'n_expert': 0,
56
+ 'n_expert_per_token': 0,
57
+ 'n_head': 32,
58
+ 'n_layer': 22,
59
+ 'n_query_groups': 4,
60
+ 'name': 'tiny-llama-1.1b',
61
+ 'norm_1': True,
62
+ 'norm_2': True,
63
+ 'norm_class_name': 'RMSNorm',
64
+ 'norm_eps': 1e-05,
65
+ 'norm_qk': False,
66
+ 'norm_qk_type': 'default',
67
+ 'padded_vocab_size': 32000,
68
+ 'padding_multiple': 64,
69
+ 'parallel_residual': False,
70
+ 'post_attention_norm': False,
71
+ 'post_mlp_norm': False,
72
+ 'rope_adjustments': None,
73
+ 'rope_base': 10000,
74
+ 'rope_condense_ratio': 1,
75
+ 'rope_indices': None,
76
+ 'rope_local_base_freq': None,
77
+ 'rotary_percentage': 1.0,
78
+ 'scale_embeddings': False,
79
+ 'shared_attention_norm': False,
80
+ 'sliding_window_indices': None,
81
+ 'sliding_window_size': None,
82
+ 'vocab_size': 32000},
83
+ 'model_name': 'tiny-llama-1.1b',
84
+ 'num_nodes': 1,
85
+ 'optimizer': "{'class_path': 'torch.optim.AdamW', 'init_args': {'lr': 0.0004, "
86
+ "'weight_decay': 0.1, 'betas': [0.9, 0.95]}}",
87
+ 'out_dir': PosixPath('out/pretrain/2412_full'),
88
+ 'ppl': False,
89
+ 'precision': 'bf16-mixed',
90
+ 'resume': False,
91
+ 'seed': 42,
92
+ 'tokenizer_dir': PosixPath('checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T'),
93
+ 'train': {'epochs': None,
94
+ 'global_batch_size': 512,
95
+ 'log_interval': 1,
96
+ 'lr_warmup_fraction': None,
97
+ 'lr_warmup_steps': 20,
98
+ 'max_norm': 1.0,
99
+ 'max_seq_length': 2048,
100
+ 'max_steps': None,
101
+ 'max_tokens': 304087040,
102
+ 'micro_batch_size': 4,
103
+ 'min_lr': 4e-05,
104
+ 'save_interval': 100,
105
+ 'tie_embeddings': None}}
106
+ Time to instantiate model: 0.04 seconds.
107
+ Total parameters: 1,100,048,384
108
+ [fix] out/pretrain/2411_full/final/lit_model.pth 是整包 state,提取 model 权重并原子覆盖...
109
+ [fix] 已覆盖为纯权重: out/pretrain/2411_full/final/lit_model.pth
110
+ Validating ...
111
+ Measured TFLOPs: 239.66
112
+ Epoch 1 | iter 32 step 1 | loss train: 1.341, val: 1.354 | iter time: 542.36 ms (step) remaining time: 0:57:31
113
+ Epoch 1 | iter 64 step 2 | loss train: 1.361, val: 1.354 | iter time: 355.55 ms (step) remaining time: 0:54:33
114
+ Epoch 1 | iter 96 step 3 | loss train: 1.447, val: 1.354 | iter time: 356.62 ms (step) remaining time: 0:53:26
115
+ Epoch 1 | iter 128 step 4 | loss train: 1.433, val: 1.354 | iter time: 355.90 ms (step) remaining time: 0:52:49
116
+ Epoch 1 | iter 160 step 5 | loss train: 1.324, val: 1.354 | iter time: 357.80 ms (step) remaining time: 0:52:22
117
+ Epoch 1 | iter 192 step 6 | loss train: 1.344, val: 1.354 | iter time: 356.75 ms (step) remaining time: 0:52:01
118
+ Epoch 1 | iter 224 step 7 | loss train: 1.360, val: 1.354 | iter time: 357.42 ms (step) remaining time: 0:51:44
119
+ Epoch 1 | iter 256 step 8 | loss train: 1.361, val: 1.354 | iter time: 361.01 ms (step) remaining time: 0:51:29
120
+ Epoch 1 | iter 288 step 9 | loss train: 1.357, val: 1.354 | iter time: 358.85 ms (step) remaining time: 0:51:15
121
+ Epoch 1 | iter 320 step 10 | loss train: 1.414, val: 1.354 | iter time: 360.47 ms (step) remaining time: 0:51:01
122
+ Epoch 1 | iter 352 step 11 | loss train: 1.382, val: 1.354 | iter time: 358.64 ms (step) remaining time: 0:50:48
123
+ Epoch 1 | iter 384 step 12 | loss train: 1.444, val: 1.354 | iter time: 359.98 ms (step) remaining time: 0:50:36
124
+ Epoch 1 | iter 416 step 13 | loss train: 1.399, val: 1.354 | iter time: 360.41 ms (step) remaining time: 0:50:23
125
+ Epoch 1 | iter 448 step 14 | loss train: 1.365, val: 1.354 | iter time: 361.38 ms (step) remaining time: 0:50:12
126
+ Epoch 1 | iter 480 step 15 | loss train: 1.500, val: 1.354 | iter time: 359.77 ms (step) remaining time: 0:50:00
127
+ Epoch 1 | iter 512 step 16 | loss train: 1.432, val: 1.354 | iter time: 358.93 ms (step) remaining time: 0:49:48
128
+ Epoch 1 | iter 544 step 17 | loss train: 1.429, val: 1.354 | iter time: 359.86 ms (step) remaining time: 0:49:36
129
+ Epoch 1 | iter 576 step 18 | loss train: 1.400, val: 1.354 | iter time: 358.84 ms (step) remaining time: 0:49:25
130
+ Epoch 1 | iter 608 step 19 | loss train: 1.399, val: 1.354 | iter time: 358.75 ms (step) remaining time: 0:49:13
131
+ Epoch 1 | iter 640 step 20 | loss train: 1.456, val: 1.354 | iter time: 360.98 ms (step) remaining time: 0:49:02
132
+ Epoch 1 | iter 672 step 21 | loss train: 1.386, val: 1.354 | iter time: 358.72 ms (step) remaining time: 0:48:51
133
+ Epoch 1 | iter 704 step 22 | loss train: 1.462, val: 1.354 | iter time: 359.69 ms (step) remaining time: 0:48:39
134
+ Epoch 1 | iter 736 step 23 | loss train: 1.410, val: 1.354 | iter time: 360.88 ms (step) remaining time: 0:48:28
135
+ Epoch 1 | iter 768 step 24 | loss train: 1.469, val: 1.354 | iter time: 359.34 ms (step) remaining time: 0:48:17
136
+ Epoch 1 | iter 800 step 25 | loss train: 1.425, val: 1.354 | iter time: 359.01 ms (step) remaining time: 0:48:06
137
+ Epoch 1 | iter 832 step 26 | loss train: 1.383, val: 1.354 | iter time: 359.88 ms (step) remaining time: 0:47:54
138
+ Epoch 1 | iter 864 step 27 | loss train: 1.419, val: 1.354 | iter time: 360.99 ms (step) remaining time: 0:47:43
139
+ Epoch 1 | iter 896 step 28 | loss train: 1.440, val: 1.354 | iter time: 359.21 ms (step) remaining time: 0:47:32
140
+ Epoch 1 | iter 928 step 29 | loss train: 1.427, val: 1.354 | iter time: 359.62 ms (step) remaining time: 0:47:21
141
+ Epoch 1 | iter 960 step 30 | loss train: 1.465, val: 1.354 | iter time: 360.69 ms (step) remaining time: 0:47:10
142
+ Epoch 1 | iter 992 step 31 | loss train: 1.429, val: 1.354 | iter time: 359.04 ms (step) remaining time: 0:46:59
143
+ Epoch 1 | iter 1024 step 32 | loss train: 1.435, val: 1.354 | iter time: 359.84 ms (step) remaining time: 0:46:48
144
+ Epoch 1 | iter 1056 step 33 | loss train: 1.395, val: 1.354 | iter time: 360.27 ms (step) remaining time: 0:46:37
145
+ Epoch 1 | iter 1088 step 34 | loss train: 1.379, val: 1.354 | iter time: 358.37 ms (step) remaining time: 0:46:26
146
+ Epoch 1 | iter 1120 step 35 | loss train: 1.468, val: 1.354 | iter time: 360.41 ms (step) remaining time: 0:46:15
147
+ Epoch 1 | iter 1152 step 36 | loss train: 1.427, val: 1.354 | iter time: 360.51 ms (step) remaining time: 0:46:04
148
+ Epoch 1 | iter 1184 step 37 | loss train: 1.396, val: 1.354 | iter time: 358.07 ms (step) remaining time: 0:45:53
149
+ Epoch 1 | iter 1216 step 38 | loss train: 1.408, val: 1.354 | iter time: 359.20 ms (step) remaining time: 0:45:42
150
+ Epoch 1 | iter 1248 step 39 | loss train: 1.386, val: 1.354 | iter time: 360.69 ms (step) remaining time: 0:45:31
151
+ Epoch 1 | iter 1280 step 40 | loss train: 1.398, val: 1.354 | iter time: 360.91 ms (step) remaining time: 0:45:20
152
+ Epoch 1 | iter 1312 step 41 | loss train: 1.401, val: 1.354 | iter time: 358.68 ms (step) remaining time: 0:45:09
153
+ Epoch 1 | iter 1344 step 42 | loss train: 1.448, val: 1.354 | iter time: 359.88 ms (step) remaining time: 0:45:00
154
+ Epoch 1 | iter 1376 step 43 | loss train: 1.456, val: 1.354 | iter time: 360.21 ms (step) remaining time: 0:44:49
155
+ Epoch 1 | iter 1408 step 44 | loss train: 1.469, val: 1.354 | iter time: 360.14 ms (step) remaining time: 0:44:38
156
+ Epoch 1 | iter 1440 step 45 | loss train: 1.411, val: 1.354 | iter time: 359.36 ms (step) remaining time: 0:44:26
157
+ Epoch 1 | iter 1472 step 46 | loss train: 1.456, val: 1.354 | iter time: 358.13 ms (step) remaining time: 0:44:15
158
+ Epoch 1 | iter 1504 step 47 | loss train: 1.457, val: 1.354 | iter time: 359.10 ms (step) remaining time: 0:44:04
159
+ Epoch 1 | iter 1536 step 48 | loss train: 1.378, val: 1.354 | iter time: 359.98 ms (step) remaining time: 0:43:53
160
+ Epoch 1 | iter 1568 step 49 | loss train: 1.496, val: 1.354 | iter time: 360.10 ms (step) remaining time: 0:43:42
161
+ Epoch 1 | iter 1600 step 50 | loss train: 1.412, val: 1.354 | iter time: 359.07 ms (step) remaining time: 0:43:31
162
+ Validating ...
163
+ iter 1600: val loss 1.3390, val time: 22296.02 ms
164
+ Epoch 1 | iter 1632 step 51 | loss train: 1.401, val: 1.339 | iter time: 360.83 ms (step) remaining time: 0:45:07
165
+ Epoch 1 | iter 1664 step 52 | loss train: 1.461, val: 1.339 | iter time: 359.43 ms (step) remaining time: 0:44:54
166
+ Epoch 1 | iter 1696 step 53 | loss train: 1.402, val: 1.339 | iter time: 360.41 ms (step) remaining time: 0:44:40
167
+ Epoch 1 | iter 1728 step 54 | loss train: 1.380, val: 1.339 | iter time: 360.50 ms (step) remaining time: 0:44:27
168
+ Epoch 1 | iter 1760 step 55 | loss train: 1.410, val: 1.339 | iter time: 359.11 ms (step) remaining time: 0:44:14
169
+ Epoch 1 | iter 1792 step 56 | loss train: 1.419, val: 1.339 | iter time: 361.42 ms (step) remaining time: 0:44:01
170
+ Epoch 1 | iter 1824 step 57 | loss train: 1.434, val: 1.339 | iter time: 360.24 ms (step) remaining time: 0:43:48
171
+ Epoch 1 | iter 1856 step 58 | loss train: 1.456, val: 1.339 | iter time: 359.49 ms (step) remaining time: 0:43:35
172
+ Epoch 1 | iter 1888 step 59 | loss train: 1.432, val: 1.339 | iter time: 359.70 ms (step) remaining time: 0:43:22
173
+ Epoch 1 | iter 1920 step 60 | loss train: 1.395, val: 1.339 | iter time: 359.47 ms (step) remaining time: 0:43:09
174
+ Epoch 1 | iter 1952 step 61 | loss train: 1.513, val: 1.339 | iter time: 361.67 ms (step) remaining time: 0:42:56
175
+ Epoch 1 | iter 1984 step 62 | loss train: 1.384, val: 1.339 | iter time: 357.72 ms (step) remaining time: 0:42:44
176
+ Epoch 1 | iter 2016 step 63 | loss train: 1.498, val: 1.339 | iter time: 360.05 ms (step) remaining time: 0:42:31
177
+ Epoch 1 | iter 2048 step 64 | loss train: 1.348, val: 1.339 | iter time: 360.51 ms (step) remaining time: 0:42:18
178
+ Epoch 1 | iter 2080 step 65 | loss train: 1.450, val: 1.339 | iter time: 360.83 ms (step) remaining time: 0:42:06
179
+ Epoch 1 | iter 2112 step 66 | loss train: 1.393, val: 1.339 | iter time: 360.45 ms (step) remaining time: 0:41:53
180
+ Epoch 1 | iter 2144 step 67 | loss train: 1.395, val: 1.339 | iter time: 358.96 ms (step) remaining time: 0:41:41
181
+ Epoch 1 | iter 2176 step 68 | loss train: 1.405, val: 1.339 | iter time: 359.69 ms (step) remaining time: 0:41:29
182
+ Epoch 1 | iter 2208 step 69 | loss train: 1.447, val: 1.339 | iter time: 358.44 ms (step) remaining time: 0:41:16
183
+ Epoch 1 | iter 2240 step 70 | loss train: 1.493, val: 1.339 | iter time: 360.57 ms (step) remaining time: 0:41:04
184
+ Epoch 1 | iter 2272 step 71 | loss train: 1.385, val: 1.339 | iter time: 359.78 ms (step) remaining time: 0:40:52
185
+ Epoch 1 | iter 2304 step 72 | loss train: 1.380, val: 1.339 | iter time: 358.99 ms (step) remaining time: 0:40:39
186
+ Epoch 1 | iter 2336 step 73 | loss train: 1.397, val: 1.339 | iter time: 359.30 ms (step) remaining time: 0:40:27
187
+ Epoch 1 | iter 2368 step 74 | loss train: 1.458, val: 1.339 | iter time: 360.61 ms (step) remaining time: 0:40:15
188
+ Epoch 1 | iter 2400 step 75 | loss train: 1.435, val: 1.339 | iter time: 359.19 ms (step) remaining time: 0:40:03
189
+ Epoch 1 | iter 2432 step 76 | loss train: 1.380, val: 1.339 | iter time: 361.20 ms (step) remaining time: 0:39:51
190
+ Epoch 1 | iter 2464 step 77 | loss train: 1.413, val: 1.339 | iter time: 360.82 ms (step) remaining time: 0:39:39
191
+ Epoch 1 | iter 2496 step 78 | loss train: 1.354, val: 1.339 | iter time: 359.73 ms (step) remaining time: 0:39:27
192
+ Epoch 1 | iter 2528 step 79 | loss train: 1.406, val: 1.339 | iter time: 359.10 ms (step) remaining time: 0:39:15
193
+ Epoch 1 | iter 2560 step 80 | loss train: 1.461, val: 1.339 | iter time: 357.56 ms (step) remaining time: 0:39:03
194
+ Epoch 1 | iter 2592 step 81 | loss train: 1.402, val: 1.339 | iter time: 359.97 ms (step) remaining time: 0:38:51
195
+ Epoch 1 | iter 2624 step 82 | loss train: 1.394, val: 1.339 | iter time: 357.17 ms (step) remaining time: 0:38:39
196
+ Epoch 1 | iter 2656 step 83 | loss train: 1.353, val: 1.339 | iter time: 359.08 ms (step) remaining time: 0:38:27
197
+ Epoch 1 | iter 2688 step 84 | loss train: 1.380, val: 1.339 | iter time: 358.63 ms (step) remaining time: 0:38:15
198
+ Epoch 1 | iter 2720 step 85 | loss train: 1.481, val: 1.339 | iter time: 358.72 ms (step) remaining time: 0:38:03
199
+ Epoch 1 | iter 2752 step 86 | loss train: 1.417, val: 1.339 | iter time: 359.02 ms (step) remaining time: 0:37:52
200
+ Epoch 1 | iter 2784 step 87 | loss train: 1.368, val: 1.339 | iter time: 359.64 ms (step) remaining time: 0:37:40
201
+ Epoch 1 | iter 2816 step 88 | loss train: 1.408, val: 1.339 | iter time: 360.00 ms (step) remaining time: 0:37:28
202
+ Epoch 1 | iter 2848 step 89 | loss train: 1.381, val: 1.339 | iter time: 358.61 ms (step) remaining time: 0:37:16
203
+ Epoch 1 | iter 2880 step 90 | loss train: 1.395, val: 1.339 | iter time: 358.45 ms (step) remaining time: 0:37:05
204
+ Epoch 1 | iter 2912 step 91 | loss train: 1.469, val: 1.339 | iter time: 359.47 ms (step) remaining time: 0:36:53
205
+ Epoch 1 | iter 2944 step 92 | loss train: 1.402, val: 1.339 | iter time: 359.16 ms (step) remaining time: 0:36:41
206
+ Epoch 1 | iter 2976 step 93 | loss train: 1.373, val: 1.339 | iter time: 359.54 ms (step) remaining time: 0:36:30
207
+ Epoch 1 | iter 3008 step 94 | loss train: 1.373, val: 1.339 | iter time: 359.46 ms (step) remaining time: 0:36:18
208
+ Epoch 1 | iter 3040 step 95 | loss train: 1.403, val: 1.339 | iter time: 361.33 ms (step) remaining time: 0:36:06
209
+ Epoch 1 | iter 3072 step 96 | loss train: 1.419, val: 1.339 | iter time: 358.25 ms (step) remaining time: 0:35:55
210
+ Epoch 1 | iter 3104 step 97 | loss train: 1.343, val: 1.339 | iter time: 358.54 ms (step) remaining time: 0:35:43
211
+ Epoch 1 | iter 3136 step 98 | loss train: 1.441, val: 1.339 | iter time: 359.17 ms (step) remaining time: 0:35:31
212
+ Epoch 1 | iter 3168 step 99 | loss train: 1.431, val: 1.339 | iter time: 358.26 ms (step) remaining time: 0:35:20
213
+ Epoch 1 | iter 3200 step 100 | loss train: 1.475, val: 1.339 | iter time: 360.35 ms (step) remaining time: 0:35:08
214
+ Validating ...
215
+ iter 3200: val loss 1.3051, val time: 22338.71 ms
216
+ Saving checkpoint to 'out/pretrain/2412_full/step-00000100/lit_model.pth'
217
+ Epoch 1 | iter 3232 step 101 | loss train: 1.356, val: 1.305 | iter time: 357.25 ms (step) remaining time: 0:36:10
218
+ Epoch 1 | iter 3264 step 102 | loss train: 1.421, val: 1.305 | iter time: 355.99 ms (step) remaining time: 0:35:57
219
+ Epoch 1 | iter 3296 step 103 | loss train: 1.403, val: 1.305 | iter time: 357.58 ms (step) remaining time: 0:35:45
220
+ Epoch 1 | iter 3328 step 104 | loss train: 1.402, val: 1.305 | iter time: 360.22 ms (step) remaining time: 0:35:32
221
+ Epoch 1 | iter 3360 step 105 | loss train: 1.499, val: 1.305 | iter time: 360.35 ms (step) remaining time: 0:35:20
222
+ Epoch 1 | iter 3392 step 106 | loss train: 1.496, val: 1.305 | iter time: 358.34 ms (step) remaining time: 0:35:07
223
+ Epoch 1 | iter 3424 step 107 | loss train: 1.446, val: 1.305 | iter time: 360.35 ms (step) remaining time: 0:34:55
224
+ Epoch 1 | iter 3456 step 108 | loss train: 1.433, val: 1.305 | iter time: 359.43 ms (step) remaining time: 0:34:42
225
+ Epoch 1 | iter 3488 step 109 | loss train: 1.400, val: 1.305 | iter time: 359.45 ms (step) remaining time: 0:34:30
226
+ Epoch 1 | iter 3520 step 110 | loss train: 1.435, val: 1.305 | iter time: 359.26 ms (step) remaining time: 0:34:18
227
+ Epoch 1 | iter 3552 step 111 | loss train: 1.474, val: 1.305 | iter time: 358.92 ms (step) remaining time: 0:34:05
228
+ Epoch 1 | iter 3584 step 112 | loss train: 1.418, val: 1.305 | iter time: 359.54 ms (step) remaining time: 0:33:53
229
+ Epoch 1 | iter 3616 step 113 | loss train: 1.369, val: 1.305 | iter time: 358.52 ms (step) remaining time: 0:33:41
230
+ Epoch 1 | iter 3648 step 114 | loss train: 1.389, val: 1.305 | iter time: 359.86 ms (step) remaining time: 0:33:28
231
+ Epoch 1 | iter 3680 step 115 | loss train: 1.388, val: 1.305 | iter time: 358.30 ms (step) remaining time: 0:33:16
232
+ Epoch 1 | iter 3712 step 116 | loss train: 1.421, val: 1.305 | iter time: 358.77 ms (step) remaining time: 0:33:04
233
+ Epoch 1 | iter 3744 step 117 | loss train: 1.442, val: 1.305 | iter time: 358.14 ms (step) remaining time: 0:32:52
234
+ Epoch 1 | iter 3776 step 118 | loss train: 1.436, val: 1.305 | iter time: 360.16 ms (step) remaining time: 0:32:40
235
+ Epoch 1 | iter 3808 step 119 | loss train: 1.407, val: 1.305 | iter time: 361.65 ms (step) remaining time: 0:32:28
236
+ Epoch 1 | iter 3840 step 120 | loss train: 1.337, val: 1.305 | iter time: 360.64 ms (step) remaining time: 0:32:16
237
+ Epoch 1 | iter 3872 step 121 | loss train: 1.400, val: 1.305 | iter time: 357.63 ms (step) remaining time: 0:32:04
238
+ Epoch 1 | iter 3904 step 122 | loss train: 1.412, val: 1.305 | iter time: 359.94 ms (step) remaining time: 0:31:51
239
+ Epoch 1 | iter 3936 step 123 | loss train: 1.392, val: 1.305 | iter time: 359.11 ms (step) remaining time: 0:31:39
240
+ Epoch 1 | iter 3968 step 124 | loss train: 1.383, val: 1.305 | iter time: 361.00 ms (step) remaining time: 0:31:27
241
+ Epoch 1 | iter 4000 step 125 | loss train: 1.445, val: 1.305 | iter time: 357.56 ms (step) remaining time: 0:31:15
242
+ Epoch 1 | iter 4032 step 126 | loss train: 1.344, val: 1.305 | iter time: 359.05 ms (step) remaining time: 0:31:03
243
+ Epoch 1 | iter 4064 step 127 | loss train: 1.416, val: 1.305 | iter time: 358.36 ms (step) remaining time: 0:30:51
244
+ Epoch 1 | iter 4096 step 128 | loss train: 1.418, val: 1.305 | iter time: 357.78 ms (step) remaining time: 0:30:39
245
+ Epoch 1 | iter 4128 step 129 | loss train: 1.414, val: 1.305 | iter time: 360.90 ms (step) remaining time: 0:30:27
246
+ Epoch 1 | iter 4160 step 130 | loss train: 1.438, val: 1.305 | iter time: 358.08 ms (step) remaining time: 0:30:15
247
+ Epoch 1 | iter 4192 step 131 | loss train: 1.408, val: 1.305 | iter time: 360.55 ms (step) remaining time: 0:30:03
248
+ Epoch 1 | iter 4224 step 132 | loss train: 1.420, val: 1.305 | iter time: 360.67 ms (step) remaining time: 0:29:51
249
+ Epoch 1 | iter 4256 step 133 | loss train: 1.462, val: 1.305 | iter time: 358.36 ms (step) remaining time: 0:29:39
250
+ Epoch 1 | iter 4288 step 134 | loss train: 1.410, val: 1.305 | iter time: 360.03 ms (step) remaining time: 0:29:28
251
+ Epoch 1 | iter 4320 step 135 | loss train: 1.436, val: 1.305 | iter time: 359.46 ms (step) remaining time: 0:29:16
252
+ Epoch 1 | iter 4352 step 136 | loss train: 1.366, val: 1.305 | iter time: 360.06 ms (step) remaining time: 0:29:04
253
+ Epoch 1 | iter 4384 step 137 | loss train: 1.352, val: 1.305 | iter time: 360.91 ms (step) remaining time: 0:28:52
254
+ Epoch 1 | iter 4416 step 138 | loss train: 1.390, val: 1.305 | iter time: 359.05 ms (step) remaining time: 0:28:40
255
+ Epoch 1 | iter 4448 step 139 | loss train: 1.408, val: 1.305 | iter time: 360.61 ms (step) remaining time: 0:28:28
256
+ Epoch 1 | iter 4480 step 140 | loss train: 1.454, val: 1.305 | iter time: 358.86 ms (step) remaining time: 0:28:16
257
+ Epoch 1 | iter 4512 step 141 | loss train: 1.325, val: 1.305 | iter time: 361.05 ms (step) remaining time: 0:28:05
258
+ Epoch 1 | iter 4544 step 142 | loss train: 1.364, val: 1.305 | iter time: 358.21 ms (step) remaining time: 0:27:53
259
+ Epoch 1 | iter 4576 step 143 | loss train: 1.424, val: 1.305 | iter time: 359.12 ms (step) remaining time: 0:27:41
260
+ Epoch 1 | iter 4608 step 144 | loss train: 1.354, val: 1.305 | iter time: 360.74 ms (step) remaining time: 0:27:29
261
+ Epoch 1 | iter 4640 step 145 | loss train: 1.325, val: 1.305 | iter time: 358.66 ms (step) remaining time: 0:27:18
262
+ Epoch 1 | iter 4672 step 146 | loss train: 1.403, val: 1.305 | iter time: 359.61 ms (step) remaining time: 0:27:06
263
+ Epoch 1 | iter 4704 step 147 | loss train: 1.376, val: 1.305 | iter time: 358.40 ms (step) remaining time: 0:26:54
264
+ Epoch 1 | iter 4736 step 148 | loss train: 1.410, val: 1.305 | iter time: 357.32 ms (step) remaining time: 0:26:42
265
+ Epoch 1 | iter 4768 step 149 | loss train: 1.352, val: 1.305 | iter time: 358.96 ms (step) remaining time: 0:26:31
266
+ Epoch 1 | iter 4800 step 150 | loss train: 1.362, val: 1.305 | iter time: 358.69 ms (step) remaining time: 0:26:19
267
+ Validating ...
268
+ iter 4800: val loss 1.2641, val time: 22318.71 ms
269
+ Epoch 1 | iter 4832 step 151 | loss train: 1.345, val: 1.264 | iter time: 359.29 ms (step) remaining time: 0:26:28
270
+ Epoch 1 | iter 4864 step 152 | loss train: 1.330, val: 1.264 | iter time: 360.07 ms (step) remaining time: 0:26:16
271
+ Epoch 1 | iter 4896 step 153 | loss train: 1.381, val: 1.264 | iter time: 360.40 ms (step) remaining time: 0:26:04
272
+ Epoch 1 | iter 4928 step 154 | loss train: 1.323, val: 1.264 | iter time: 360.91 ms (step) remaining time: 0:25:52
273
+ Epoch 1 | iter 4960 step 155 | loss train: 1.432, val: 1.264 | iter time: 359.53 ms (step) remaining time: 0:25:40
274
+ Epoch 1 | iter 4992 step 156 | loss train: 1.333, val: 1.264 | iter time: 360.37 ms (step) remaining time: 0:25:28
275
+ Epoch 1 | iter 5024 step 157 | loss train: 1.412, val: 1.264 | iter time: 360.80 ms (step) remaining time: 0:25:17
276
+ Epoch 1 | iter 5056 step 158 | loss train: 1.410, val: 1.264 | iter time: 358.73 ms (step) remaining time: 0:25:05
277
+ Epoch 1 | iter 5088 step 159 | loss train: 1.357, val: 1.264 | iter time: 358.80 ms (step) remaining time: 0:24:53
278
+ Epoch 1 | iter 5120 step 160 | loss train: 1.416, val: 1.264 | iter time: 358.54 ms (step) remaining time: 0:24:41
279
+ Epoch 1 | iter 5152 step 161 | loss train: 1.361, val: 1.264 | iter time: 359.90 ms (step) remaining time: 0:24:29
280
+ Epoch 1 | iter 5184 step 162 | loss train: 1.404, val: 1.264 | iter time: 360.05 ms (step) remaining time: 0:24:17
281
+ Epoch 1 | iter 5216 step 163 | loss train: 1.413, val: 1.264 | iter time: 357.68 ms (step) remaining time: 0:24:06
282
+ Epoch 1 | iter 5248 step 164 | loss train: 1.378, val: 1.264 | iter time: 358.85 ms (step) remaining time: 0:23:54
283
+ Epoch 1 | iter 5280 step 165 | loss train: 1.327, val: 1.264 | iter time: 358.98 ms (step) remaining time: 0:23:42
284
+ Epoch 1 | iter 5312 step 166 | loss train: 1.312, val: 1.264 | iter time: 358.73 ms (step) remaining time: 0:23:30
285
+ Epoch 1 | iter 5344 step 167 | loss train: 1.428, val: 1.264 | iter time: 360.91 ms (step) remaining time: 0:23:18
286
+ Epoch 1 | iter 5376 step 168 | loss train: 1.457, val: 1.264 | iter time: 359.35 ms (step) remaining time: 0:23:07
287
+ Epoch 1 | iter 5408 step 169 | loss train: 1.415, val: 1.264 | iter time: 358.68 ms (step) remaining time: 0:22:55
288
+ Epoch 1 | iter 5440 step 170 | loss train: 1.322, val: 1.264 | iter time: 360.00 ms (step) remaining time: 0:22:43
289
+ Epoch 1 | iter 5472 step 171 | loss train: 1.370, val: 1.264 | iter time: 361.17 ms (step) remaining time: 0:22:32
290
+ Epoch 1 | iter 5504 step 172 | loss train: 1.417, val: 1.264 | iter time: 361.49 ms (step) remaining time: 0:22:20
291
+ Epoch 1 | iter 5536 step 173 | loss train: 1.394, val: 1.264 | iter time: 360.35 ms (step) remaining time: 0:22:08
292
+ Epoch 1 | iter 5568 step 174 | loss train: 1.361, val: 1.264 | iter time: 360.22 ms (step) remaining time: 0:21:57
293
+ Epoch 1 | iter 5600 step 175 | loss train: 1.391, val: 1.264 | iter time: 357.35 ms (step) remaining time: 0:21:45
294
+ Epoch 1 | iter 5632 step 176 | loss train: 1.384, val: 1.264 | iter time: 357.40 ms (step) remaining time: 0:21:33
295
+ Epoch 1 | iter 5664 step 177 | loss train: 1.388, val: 1.264 | iter time: 360.11 ms (step) remaining time: 0:21:22
296
+ Epoch 1 | iter 5696 step 178 | loss train: 1.350, val: 1.264 | iter time: 360.21 ms (step) remaining time: 0:21:10
297
+ Epoch 1 | iter 5728 step 179 | loss train: 1.408, val: 1.264 | iter time: 357.78 ms (step) remaining time: 0:20:58
298
+ Epoch 1 | iter 5760 step 180 | loss train: 1.393, val: 1.264 | iter time: 358.86 ms (step) remaining time: 0:20:47
299
+ Epoch 1 | iter 5792 step 181 | loss train: 1.325, val: 1.264 | iter time: 359.74 ms (step) remaining time: 0:20:35
300
+ Epoch 1 | iter 5824 step 182 | loss train: 1.401, val: 1.264 | iter time: 359.26 ms (step) remaining time: 0:20:24
301
+ Epoch 1 | iter 5856 step 183 | loss train: 1.367, val: 1.264 | iter time: 359.59 ms (step) remaining time: 0:20:12
302
+ Epoch 1 | iter 5888 step 184 | loss train: 1.324, val: 1.264 | iter time: 357.58 ms (step) remaining time: 0:20:00
303
+ Epoch 1 | iter 5920 step 185 | loss train: 1.377, val: 1.264 | iter time: 359.21 ms (step) remaining time: 0:19:49
304
+ Epoch 1 | iter 5952 step 186 | loss train: 1.343, val: 1.264 | iter time: 359.98 ms (step) remaining time: 0:19:37
305
+ Epoch 1 | iter 5984 step 187 | loss train: 1.425, val: 1.264 | iter time: 358.59 ms (step) remaining time: 0:19:26
306
+ Epoch 1 | iter 6016 step 188 | loss train: 1.313, val: 1.264 | iter time: 359.82 ms (step) remaining time: 0:19:14
307
+ Epoch 1 | iter 6048 step 189 | loss train: 1.411, val: 1.264 | iter time: 359.79 ms (step) remaining time: 0:19:03
308
+ Epoch 1 | iter 6080 step 190 | loss train: 1.353, val: 1.264 | iter time: 359.23 ms (step) remaining time: 0:18:51
309
+ Epoch 1 | iter 6112 step 191 | loss train: 1.320, val: 1.264 | iter time: 359.70 ms (step) remaining time: 0:18:39
310
+ Epoch 1 | iter 6144 step 192 | loss train: 1.365, val: 1.264 | iter time: 360.71 ms (step) remaining time: 0:18:28
311
+ Epoch 1 | iter 6176 step 193 | loss train: 1.328, val: 1.264 | iter time: 358.91 ms (step) remaining time: 0:18:16
312
+ Epoch 1 | iter 6208 step 194 | loss train: 1.401, val: 1.264 | iter time: 358.67 ms (step) remaining time: 0:18:05
313
+ Epoch 1 | iter 6240 step 195 | loss train: 1.308, val: 1.264 | iter time: 359.17 ms (step) remaining time: 0:17:53
314
+ Epoch 1 | iter 6272 step 196 | loss train: 1.264, val: 1.264 | iter time: 358.97 ms (step) remaining time: 0:17:42
315
+ Epoch 1 | iter 6304 step 197 | loss train: 1.383, val: 1.264 | iter time: 358.90 ms (step) remaining time: 0:17:30
316
+ Epoch 1 | iter 6336 step 198 | loss train: 1.374, val: 1.264 | iter time: 360.07 ms (step) remaining time: 0:17:19
317
+ Epoch 1 | iter 6368 step 199 | loss train: 1.380, val: 1.264 | iter time: 358.77 ms (step) remaining time: 0:17:07
318
+ Epoch 1 | iter 6400 step 200 | loss train: 1.368, val: 1.264 | iter time: 361.04 ms (step) remaining time: 0:16:56
319
+ Validating ...
320
+ iter 6400: val loss 1.2153, val time: 22309.80 ms
321
+ Saving checkpoint to 'out/pretrain/2412_full/step-00000200/lit_model.pth'
322
+ Epoch 1 | iter 6432 step 201 | loss train: 1.359, val: 1.215 | iter time: 355.81 ms (step) remaining time: 0:17:02
323
+ Epoch 1 | iter 6464 step 202 | loss train: 1.352, val: 1.215 | iter time: 359.06 ms (step) remaining time: 0:16:50
324
+ Epoch 1 | iter 6496 step 203 | loss train: 1.301, val: 1.215 | iter time: 359.85 ms (step) remaining time: 0:16:38
325
+ Epoch 1 | iter 6528 step 204 | loss train: 1.308, val: 1.215 | iter time: 358.79 ms (step) remaining time: 0:16:26
326
+ Epoch 1 | iter 6560 step 205 | loss train: 1.392, val: 1.215 | iter time: 359.65 ms (step) remaining time: 0:16:15
327
+ Epoch 1 | iter 6592 step 206 | loss train: 1.352, val: 1.215 | iter time: 359.50 ms (step) remaining time: 0:16:03
328
+ Epoch 1 | iter 6624 step 207 | loss train: 1.410, val: 1.215 | iter time: 357.80 ms (step) remaining time: 0:15:51
329
+ Epoch 1 | iter 6656 step 208 | loss train: 1.366, val: 1.215 | iter time: 358.14 ms (step) remaining time: 0:15:39
330
+ Epoch 1 | iter 6688 step 209 | loss train: 1.313, val: 1.215 | iter time: 358.15 ms (step) remaining time: 0:15:28
331
+ Epoch 1 | iter 6720 step 210 | loss train: 1.350, val: 1.215 | iter time: 359.08 ms (step) remaining time: 0:15:16
332
+ Epoch 1 | iter 6752 step 211 | loss train: 1.346, val: 1.215 | iter time: 360.07 ms (step) remaining time: 0:15:04
333
+ Epoch 1 | iter 6784 step 212 | loss train: 1.426, val: 1.215 | iter time: 360.54 ms (step) remaining time: 0:14:53
334
+ Epoch 1 | iter 6816 step 213 | loss train: 1.358, val: 1.215 | iter time: 360.48 ms (step) remaining time: 0:14:41
335
+ Epoch 1 | iter 6848 step 214 | loss train: 1.337, val: 1.215 | iter time: 360.18 ms (step) remaining time: 0:14:29
336
+ Epoch 1 | iter 6880 step 215 | loss train: 1.411, val: 1.215 | iter time: 359.64 ms (step) remaining time: 0:14:18
337
+ Epoch 1 | iter 6912 step 216 | loss train: 1.298, val: 1.215 | iter time: 358.69 ms (step) remaining time: 0:14:06
338
+ Epoch 1 | iter 6944 step 217 | loss train: 1.372, val: 1.215 | iter time: 360.14 ms (step) remaining time: 0:13:54
339
+ Epoch 1 | iter 6976 step 218 | loss train: 1.322, val: 1.215 | iter time: 361.39 ms (step) remaining time: 0:13:43
340
+ Epoch 1 | iter 7008 step 219 | loss train: 1.347, val: 1.215 | iter time: 359.88 ms (step) remaining time: 0:13:31
341
+ Epoch 1 | iter 7040 step 220 | loss train: 1.327, val: 1.215 | iter time: 359.27 ms (step) remaining time: 0:13:20
342
+ Epoch 1 | iter 7072 step 221 | loss train: 1.312, val: 1.215 | iter time: 359.89 ms (step) remaining time: 0:13:08
343
+ Epoch 1 | iter 7104 step 222 | loss train: 1.337, val: 1.215 | iter time: 359.85 ms (step) remaining time: 0:12:56
344
+ Epoch 1 | iter 7136 step 223 | loss train: 1.350, val: 1.215 | iter time: 358.61 ms (step) remaining time: 0:12:45
345
+ Epoch 1 | iter 7168 step 224 | loss train: 1.312, val: 1.215 | iter time: 358.78 ms (step) remaining time: 0:12:33
346
+ Epoch 1 | iter 7200 step 225 | loss train: 1.328, val: 1.215 | iter time: 360.38 ms (step) remaining time: 0:12:22
347
+ Epoch 1 | iter 7232 step 226 | loss train: 1.392, val: 1.215 | iter time: 359.40 ms (step) remaining time: 0:12:10
348
+ Epoch 1 | iter 7264 step 227 | loss train: 1.327, val: 1.215 | iter time: 359.22 ms (step) remaining time: 0:11:58
349
+ Epoch 1 | iter 7296 step 228 | loss train: 1.390, val: 1.215 | iter time: 360.04 ms (step) remaining time: 0:11:47
350
+ Epoch 1 | iter 7328 step 229 | loss train: 1.355, val: 1.215 | iter time: 359.28 ms (step) remaining time: 0:11:35
351
+ Epoch 1 | iter 7360 step 230 | loss train: 1.386, val: 1.215 | iter time: 360.59 ms (step) remaining time: 0:11:24
352
+ Epoch 1 | iter 7392 step 231 | loss train: 1.347, val: 1.215 | iter time: 358.41 ms (step) remaining time: 0:11:12
353
+ Epoch 1 | iter 7424 step 232 | loss train: 1.324, val: 1.215 | iter time: 358.11 ms (step) remaining time: 0:11:01
354
+ Epoch 1 | iter 7456 step 233 | loss train: 1.345, val: 1.215 | iter time: 359.01 ms (step) remaining time: 0:10:49
355
+ Epoch 1 | iter 7488 step 234 | loss train: 1.301, val: 1.215 | iter time: 357.50 ms (step) remaining time: 0:10:38
356
+ Epoch 1 | iter 7520 step 235 | loss train: 1.382, val: 1.215 | iter time: 360.48 ms (step) remaining time: 0:10:26
357
+ Epoch 1 | iter 7552 step 236 | loss train: 1.362, val: 1.215 | iter time: 359.96 ms (step) remaining time: 0:10:15
358
+ Epoch 1 | iter 7584 step 237 | loss train: 1.447, val: 1.215 | iter time: 359.46 ms (step) remaining time: 0:10:03
359
+ Epoch 1 | iter 7616 step 238 | loss train: 1.407, val: 1.215 | iter time: 359.74 ms (step) remaining time: 0:09:52
360
+ Epoch 1 | iter 7648 step 239 | loss train: 1.319, val: 1.215 | iter time: 360.08 ms (step) remaining time: 0:09:40
361
+ Epoch 1 | iter 7680 step 240 | loss train: 1.305, val: 1.215 | iter time: 358.30 ms (step) remaining time: 0:09:29
362
+ Epoch 1 | iter 7712 step 241 | loss train: 1.373, val: 1.215 | iter time: 358.91 ms (step) remaining time: 0:09:17
363
+ Epoch 1 | iter 7744 step 242 | loss train: 1.302, val: 1.215 | iter time: 359.06 ms (step) remaining time: 0:09:06
364
+ Epoch 1 | iter 7776 step 243 | loss train: 1.306, val: 1.215 | iter time: 359.05 ms (step) remaining time: 0:08:54
365
+ Epoch 1 | iter 7808 step 244 | loss train: 1.350, val: 1.215 | iter time: 360.22 ms (step) remaining time: 0:08:43
366
+ Epoch 1 | iter 7840 step 245 | loss train: 1.405, val: 1.215 | iter time: 360.15 ms (step) remaining time: 0:08:31
367
+ Epoch 1 | iter 7872 step 246 | loss train: 1.363, val: 1.215 | iter time: 359.16 ms (step) remaining time: 0:08:20
368
+ Epoch 1 | iter 7904 step 247 | loss train: 1.371, val: 1.215 | iter time: 360.34 ms (step) remaining time: 0:08:08
369
+ Epoch 1 | iter 7936 step 248 | loss train: 1.305, val: 1.215 | iter time: 360.11 ms (step) remaining time: 0:07:57
370
+ Epoch 1 | iter 7968 step 249 | loss train: 1.302, val: 1.215 | iter time: 358.10 ms (step) remaining time: 0:07:45
371
+ Epoch 1 | iter 8000 step 250 | loss train: 1.362, val: 1.215 | iter time: 361.27 ms (step) remaining time: 0:07:34
372
+ Validating ...
373
+ iter 8000: val loss 1.1898, val time: 22341.58 ms
374
+ Epoch 1 | iter 8032 step 251 | loss train: 1.403, val: 1.190 | iter time: 360.25 ms (step) remaining time: 0:07:26
375
+ Epoch 1 | iter 8064 step 252 | loss train: 1.336, val: 1.190 | iter time: 359.44 ms (step) remaining time: 0:07:15
376
+ Epoch 1 | iter 8096 step 253 | loss train: 1.387, val: 1.190 | iter time: 358.14 ms (step) remaining time: 0:07:03
377
+ Epoch 1 | iter 8128 step 254 | loss train: 1.346, val: 1.190 | iter time: 358.79 ms (step) remaining time: 0:06:51
378
+ Epoch 1 | iter 8160 step 255 | loss train: 1.287, val: 1.190 | iter time: 359.99 ms (step) remaining time: 0:06:40
379
+ Epoch 1 | iter 8192 step 256 | loss train: 1.320, val: 1.190 | iter time: 358.56 ms (step) remaining time: 0:06:28
380
+ Epoch 1 | iter 8224 step 257 | loss train: 1.288, val: 1.190 | iter time: 359.50 ms (step) remaining time: 0:06:17
381
+ Epoch 1 | iter 8256 step 258 | loss train: 1.401, val: 1.190 | iter time: 358.16 ms (step) remaining time: 0:06:05
382
+ Epoch 1 | iter 8288 step 259 | loss train: 1.352, val: 1.190 | iter time: 359.35 ms (step) remaining time: 0:05:54
383
+ Epoch 1 | iter 8320 step 260 | loss train: 1.350, val: 1.190 | iter time: 359.49 ms (step) remaining time: 0:05:42
384
+ Epoch 1 | iter 8352 step 261 | loss train: 1.428, val: 1.190 | iter time: 359.55 ms (step) remaining time: 0:05:31
385
+ Epoch 1 | iter 8384 step 262 | loss train: 1.390, val: 1.190 | iter time: 359.85 ms (step) remaining time: 0:05:19
386
+ Epoch 1 | iter 8416 step 263 | loss train: 1.397, val: 1.190 | iter time: 360.74 ms (step) remaining time: 0:05:08
387
+ Epoch 1 | iter 8448 step 264 | loss train: 1.296, val: 1.190 | iter time: 359.49 ms (step) remaining time: 0:04:56
388
+ Epoch 1 | iter 8480 step 265 | loss train: 1.330, val: 1.190 | iter time: 360.23 ms (step) remaining time: 0:04:45
389
+ Epoch 1 | iter 8512 step 266 | loss train: 1.342, val: 1.190 | iter time: 359.45 ms (step) remaining time: 0:04:34
390
+ Epoch 1 | iter 8544 step 267 | loss train: 1.339, val: 1.190 | iter time: 360.97 ms (step) remaining time: 0:04:22
391
+ Epoch 1 | iter 8576 step 268 | loss train: 1.369, val: 1.190 | iter time: 360.16 ms (step) remaining time: 0:04:11
392
+ Epoch 1 | iter 8608 step 269 | loss train: 1.330, val: 1.190 | iter time: 360.27 ms (step) remaining time: 0:03:59
393
+ Epoch 1 | iter 8640 step 270 | loss train: 1.327, val: 1.190 | iter time: 359.12 ms (step) remaining time: 0:03:48
394
+ Epoch 1 | iter 8672 step 271 | loss train: 1.323, val: 1.190 | iter time: 357.73 ms (step) remaining time: 0:03:36
395
+ Epoch 1 | iter 8704 step 272 | loss train: 1.382, val: 1.190 | iter time: 358.08 ms (step) remaining time: 0:03:25
396
+ Epoch 1 | iter 8736 step 273 | loss train: 1.356, val: 1.190 | iter time: 359.43 ms (step) remaining time: 0:03:13
397
+ Epoch 1 | iter 8768 step 274 | loss train: 1.337, val: 1.190 | iter time: 359.24 ms (step) remaining time: 0:03:02
398
+ Epoch 1 | iter 8800 step 275 | loss train: 1.355, val: 1.190 | iter time: 359.00 ms (step) remaining time: 0:02:50
399
+ Epoch 1 | iter 8832 step 276 | loss train: 1.368, val: 1.190 | iter time: 359.62 ms (step) remaining time: 0:02:39
400
+ Epoch 1 | iter 8864 step 277 | loss train: 1.358, val: 1.190 | iter time: 358.51 ms (step) remaining time: 0:02:28
401
+ Epoch 1 | iter 8896 step 278 | loss train: 1.335, val: 1.190 | iter time: 358.77 ms (step) remaining time: 0:02:16
402
+ Epoch 1 | iter 8928 step 279 | loss train: 1.335, val: 1.190 | iter time: 359.75 ms (step) remaining time: 0:02:05
403
+ Epoch 1 | iter 8960 step 280 | loss train: 1.324, val: 1.190 | iter time: 358.75 ms (step) remaining time: 0:01:53
404
+ Epoch 1 | iter 8992 step 281 | loss train: 1.363, val: 1.190 | iter time: 360.08 ms (step) remaining time: 0:01:42
405
+ Epoch 1 | iter 9024 step 282 | loss train: 1.342, val: 1.190 | iter time: 357.96 ms (step) remaining time: 0:01:31
406
+ Epoch 1 | iter 9056 step 283 | loss train: 1.330, val: 1.190 | iter time: 359.58 ms (step) remaining time: 0:01:19
407
+ Epoch 1 | iter 9088 step 284 | loss train: 1.378, val: 1.190 | iter time: 359.22 ms (step) remaining time: 0:01:08
408
+ Epoch 1 | iter 9120 step 285 | loss train: 1.395, val: 1.190 | iter time: 359.31 ms (step) remaining time: 0:00:56
409
+ Epoch 1 | iter 9152 step 286 | loss train: 1.375, val: 1.190 | iter time: 358.42 ms (step) remaining time: 0:00:45
410
+ Epoch 1 | iter 9184 step 287 | loss train: 1.343, val: 1.190 | iter time: 359.37 ms (step) remaining time: 0:00:34
411
+ Epoch 1 | iter 9216 step 288 | loss train: 1.376, val: 1.190 | iter time: 359.54 ms (step) remaining time: 0:00:22
412
+ Epoch 2 | iter 9248 step 289 | loss train: 1.302, val: 1.190 | iter time: 358.39 ms (step) remaining time: 0:00:11
413
+ Epoch 2 | iter 9280 step 290 | loss train: 1.167, val: 1.190 | iter time: 358.25 ms (step) remaining time: 0:00:00
414
+ Validating ...
415
+ Final evaluation | val loss: 1.173 | val ppl: 3.232
416
+ Saving checkpoint to 'out/pretrain/2412_full/final/lit_model.pth'
417
+ ----------------------------------------
418
+ | Performance
419
+ | - Total tokens : 304,087,040
420
+ | - Training Time : 3361.34 s
421
+ | - Tok/sec : 163.63 tok/s
422
+ | ----------------------------------------
423
+ | Memory Usage
424
+ | - Memory Used : 26.32 GB
425
+ ----------------------------------------
out/pretrain/tinyllama/teelogs/2412_lr4e-5.txt ADDED
@@ -0,0 +1,424 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Initializing distributed: GLOBAL_RANK: 0, MEMBER: 1/4
2
+ Initializing distributed: GLOBAL_RANK: 1, MEMBER: 2/4
3
+ [rank: 1] Seed set to 42
4
+ Initializing distributed: GLOBAL_RANK: 3, MEMBER: 4/4
5
+ [rank: 3] Seed set to 42
6
+ Initializing distributed: GLOBAL_RANK: 2, MEMBER: 3/4
7
+ [rank: 2] Seed set to 42
8
+ ----------------------------------------------------------------------------------------------------
9
+ distributed_backend=nccl
10
+ All distributed processes registered. Starting with 4 processes
11
+ ----------------------------------------------------------------------------------------------------
12
+
13
+ [rank: 0] Seed set to 42
14
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
15
+ train_dataloader = data.train_dataloader()
16
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
17
+ train_dataloader = data.train_dataloader()
18
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
19
+ train_dataloader = data.train_dataloader()
20
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
21
+ train_dataloader = data.train_dataloader()
22
+ All GPUs are fully connected via NVLink.
23
+ {'data': {'batch_size': 1,
24
+ 'data_path': PosixPath('litgpt/data/arxiv'),
25
+ 'num_workers': 0,
26
+ 'seed': 42,
27
+ 'seq_length': 2048,
28
+ 'use_starcoder': True},
29
+ 'data_dir': PosixPath('litgpt/data/arxiv/2412'),
30
+ 'devices': 'auto',
31
+ 'eval': {'evaluate_example': 'first',
32
+ 'final_validation': True,
33
+ 'initial_validation': True,
34
+ 'interval': 50,
35
+ 'max_iters': 200,
36
+ 'max_new_tokens': None},
37
+ 'initial_checkpoint_dir': PosixPath('out/pretrain/tinyllama/2411_full/final'),
38
+ 'log': {'group': None, 'project': None, 'run': None},
39
+ 'logger_name': 'tensorboard',
40
+ 'model_config': {'attention_logit_softcapping': None,
41
+ 'attention_scores_scalar': None,
42
+ 'attn_bias': False,
43
+ 'bias': False,
44
+ 'block_size': 2048,
45
+ 'final_logit_softcapping': None,
46
+ 'gelu_approximate': 'none',
47
+ 'head_size': 64,
48
+ 'hf_config': {'name': 'TinyLlama-1.1B-intermediate-step-1431k-3T',
49
+ 'org': 'TinyLlama'},
50
+ 'intermediate_size': 5632,
51
+ 'lm_head_bias': False,
52
+ 'mlp_class_name': 'LLaMAMLP',
53
+ 'moe_intermediate_size': None,
54
+ 'n_embd': 2048,
55
+ 'n_expert': 0,
56
+ 'n_expert_per_token': 0,
57
+ 'n_head': 32,
58
+ 'n_layer': 22,
59
+ 'n_query_groups': 4,
60
+ 'name': 'tiny-llama-1.1b',
61
+ 'norm_1': True,
62
+ 'norm_2': True,
63
+ 'norm_class_name': 'RMSNorm',
64
+ 'norm_eps': 1e-05,
65
+ 'norm_qk': False,
66
+ 'norm_qk_type': 'default',
67
+ 'padded_vocab_size': 32000,
68
+ 'padding_multiple': 64,
69
+ 'parallel_residual': False,
70
+ 'post_attention_norm': False,
71
+ 'post_mlp_norm': False,
72
+ 'rope_adjustments': None,
73
+ 'rope_base': 10000,
74
+ 'rope_condense_ratio': 1,
75
+ 'rope_indices': None,
76
+ 'rope_local_base_freq': None,
77
+ 'rotary_percentage': 1.0,
78
+ 'scale_embeddings': False,
79
+ 'shared_attention_norm': False,
80
+ 'sliding_window_indices': None,
81
+ 'sliding_window_size': None,
82
+ 'vocab_size': 32000},
83
+ 'model_name': 'tiny-llama-1.1b',
84
+ 'num_nodes': 1,
85
+ 'optimizer': "{'class_path': 'torch.optim.AdamW', 'init_args': {'lr': 4e-05, "
86
+ "'weight_decay': 0.1, 'betas': [0.9, 0.95]}}",
87
+ 'out_dir': PosixPath('out/pretrain/tinyllama/2412_lr4e-5'),
88
+ 'ppl': False,
89
+ 'precision': 'bf16-mixed',
90
+ 'resume': False,
91
+ 'seed': 42,
92
+ 'tokenizer_dir': PosixPath('checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T'),
93
+ 'train': {'epochs': None,
94
+ 'global_batch_size': 512,
95
+ 'log_interval': 1,
96
+ 'lr_warmup_fraction': None,
97
+ 'lr_warmup_steps': 20,
98
+ 'max_norm': 1.0,
99
+ 'max_seq_length': 2048,
100
+ 'max_steps': None,
101
+ 'max_tokens': 304087040,
102
+ 'micro_batch_size': 4,
103
+ 'min_lr': 4e-05,
104
+ 'save_interval': 100,
105
+ 'tie_embeddings': None}}
106
+ Time to instantiate model: 0.02 seconds.
107
+ Total parameters: 1,100,048,384
108
+ [ok] out/pretrain/tinyllama/2411_full/final/lit_model.pth 已是纯权重
109
+ Validating ...
110
+ Measured TFLOPs: 239.66
111
+ Epoch 1 | iter 32 step 1 | loss train: 1.336, val: 1.313 | iter time: 565.83 ms (step) remaining time: 0:56:26
112
+ Epoch 1 | iter 64 step 2 | loss train: 1.317, val: 1.313 | iter time: 356.53 ms (step) remaining time: 0:54:00
113
+ Epoch 1 | iter 96 step 3 | loss train: 1.342, val: 1.313 | iter time: 359.62 ms (step) remaining time: 0:53:05
114
+ Epoch 1 | iter 128 step 4 | loss train: 1.431, val: 1.313 | iter time: 358.74 ms (step) remaining time: 0:52:41
115
+ Epoch 1 | iter 160 step 5 | loss train: 1.400, val: 1.313 | iter time: 359.17 ms (step) remaining time: 0:52:17
116
+ Epoch 1 | iter 192 step 6 | loss train: 1.420, val: 1.313 | iter time: 362.10 ms (step) remaining time: 0:51:59
117
+ Epoch 1 | iter 224 step 7 | loss train: 1.456, val: 1.313 | iter time: 362.36 ms (step) remaining time: 0:51:42
118
+ Epoch 1 | iter 256 step 8 | loss train: 1.355, val: 1.313 | iter time: 359.39 ms (step) remaining time: 0:51:28
119
+ Epoch 1 | iter 288 step 9 | loss train: 1.346, val: 1.313 | iter time: 359.07 ms (step) remaining time: 0:51:14
120
+ Epoch 1 | iter 320 step 10 | loss train: 1.458, val: 1.313 | iter time: 360.41 ms (step) remaining time: 0:51:01
121
+ Epoch 1 | iter 352 step 11 | loss train: 1.388, val: 1.313 | iter time: 359.18 ms (step) remaining time: 0:50:48
122
+ Epoch 1 | iter 384 step 12 | loss train: 1.411, val: 1.313 | iter time: 358.91 ms (step) remaining time: 0:50:35
123
+ Epoch 1 | iter 416 step 13 | loss train: 1.331, val: 1.313 | iter time: 360.22 ms (step) remaining time: 0:50:23
124
+ Epoch 1 | iter 448 step 14 | loss train: 1.439, val: 1.313 | iter time: 359.52 ms (step) remaining time: 0:50:11
125
+ Epoch 1 | iter 480 step 15 | loss train: 1.357, val: 1.313 | iter time: 360.56 ms (step) remaining time: 0:49:59
126
+ Epoch 1 | iter 512 step 16 | loss train: 1.350, val: 1.313 | iter time: 360.92 ms (step) remaining time: 0:49:48
127
+ Epoch 1 | iter 544 step 17 | loss train: 1.353, val: 1.313 | iter time: 358.80 ms (step) remaining time: 0:49:36
128
+ Epoch 1 | iter 576 step 18 | loss train: 1.354, val: 1.313 | iter time: 360.08 ms (step) remaining time: 0:49:25
129
+ Epoch 1 | iter 608 step 19 | loss train: 1.396, val: 1.313 | iter time: 357.80 ms (step) remaining time: 0:49:14
130
+ Epoch 1 | iter 640 step 20 | loss train: 1.383, val: 1.313 | iter time: 358.83 ms (step) remaining time: 0:49:04
131
+ Epoch 1 | iter 672 step 21 | loss train: 1.401, val: 1.313 | iter time: 360.35 ms (step) remaining time: 0:48:52
132
+ Epoch 1 | iter 704 step 22 | loss train: 1.397, val: 1.313 | iter time: 358.78 ms (step) remaining time: 0:48:41
133
+ Epoch 1 | iter 736 step 23 | loss train: 1.378, val: 1.313 | iter time: 358.63 ms (step) remaining time: 0:48:30
134
+ Epoch 1 | iter 768 step 24 | loss train: 1.405, val: 1.313 | iter time: 360.41 ms (step) remaining time: 0:48:19
135
+ Epoch 1 | iter 800 step 25 | loss train: 1.383, val: 1.313 | iter time: 358.98 ms (step) remaining time: 0:48:07
136
+ Epoch 1 | iter 832 step 26 | loss train: 1.374, val: 1.313 | iter time: 360.33 ms (step) remaining time: 0:47:56
137
+ Epoch 1 | iter 864 step 27 | loss train: 1.372, val: 1.313 | iter time: 361.14 ms (step) remaining time: 0:47:45
138
+ Epoch 1 | iter 896 step 28 | loss train: 1.336, val: 1.313 | iter time: 359.37 ms (step) remaining time: 0:47:34
139
+ Epoch 1 | iter 928 step 29 | loss train: 1.453, val: 1.313 | iter time: 358.61 ms (step) remaining time: 0:47:23
140
+ Epoch 1 | iter 960 step 30 | loss train: 1.349, val: 1.313 | iter time: 360.64 ms (step) remaining time: 0:47:12
141
+ Epoch 1 | iter 992 step 31 | loss train: 1.352, val: 1.313 | iter time: 359.15 ms (step) remaining time: 0:47:01
142
+ Epoch 1 | iter 1024 step 32 | loss train: 1.361, val: 1.313 | iter time: 358.99 ms (step) remaining time: 0:46:50
143
+ Epoch 1 | iter 1056 step 33 | loss train: 1.384, val: 1.313 | iter time: 360.68 ms (step) remaining time: 0:46:38
144
+ Epoch 1 | iter 1088 step 34 | loss train: 1.324, val: 1.313 | iter time: 361.32 ms (step) remaining time: 0:46:27
145
+ Epoch 1 | iter 1120 step 35 | loss train: 1.345, val: 1.313 | iter time: 360.46 ms (step) remaining time: 0:46:16
146
+ Epoch 1 | iter 1152 step 36 | loss train: 1.323, val: 1.313 | iter time: 358.31 ms (step) remaining time: 0:46:05
147
+ Epoch 1 | iter 1184 step 37 | loss train: 1.361, val: 1.313 | iter time: 360.53 ms (step) remaining time: 0:45:54
148
+ Epoch 1 | iter 1216 step 38 | loss train: 1.354, val: 1.313 | iter time: 361.09 ms (step) remaining time: 0:45:43
149
+ Epoch 1 | iter 1248 step 39 | loss train: 1.399, val: 1.313 | iter time: 358.49 ms (step) remaining time: 0:45:32
150
+ Epoch 1 | iter 1280 step 40 | loss train: 1.402, val: 1.313 | iter time: 359.77 ms (step) remaining time: 0:45:21
151
+ Epoch 1 | iter 1312 step 41 | loss train: 1.361, val: 1.313 | iter time: 360.00 ms (step) remaining time: 0:45:11
152
+ Epoch 1 | iter 1344 step 42 | loss train: 1.378, val: 1.313 | iter time: 360.63 ms (step) remaining time: 0:44:59
153
+ Epoch 1 | iter 1376 step 43 | loss train: 1.383, val: 1.313 | iter time: 358.34 ms (step) remaining time: 0:44:48
154
+ Epoch 1 | iter 1408 step 44 | loss train: 1.435, val: 1.313 | iter time: 359.95 ms (step) remaining time: 0:44:37
155
+ Epoch 1 | iter 1440 step 45 | loss train: 1.423, val: 1.313 | iter time: 360.90 ms (step) remaining time: 0:44:27
156
+ Epoch 1 | iter 1472 step 46 | loss train: 1.322, val: 1.313 | iter time: 360.88 ms (step) remaining time: 0:44:17
157
+ Epoch 1 | iter 1504 step 47 | loss train: 1.326, val: 1.313 | iter time: 360.33 ms (step) remaining time: 0:44:06
158
+ Epoch 1 | iter 1536 step 48 | loss train: 1.383, val: 1.313 | iter time: 358.00 ms (step) remaining time: 0:43:55
159
+ Epoch 1 | iter 1568 step 49 | loss train: 1.384, val: 1.313 | iter time: 358.84 ms (step) remaining time: 0:43:44
160
+ Epoch 1 | iter 1600 step 50 | loss train: 1.391, val: 1.313 | iter time: 358.96 ms (step) remaining time: 0:43:33
161
+ Validating ...
162
+ iter 1600: val loss 1.3029, val time: 21936.26 ms
163
+ Epoch 1 | iter 1632 step 51 | loss train: 1.355, val: 1.303 | iter time: 359.70 ms (step) remaining time: 0:45:05
164
+ Epoch 1 | iter 1664 step 52 | loss train: 1.364, val: 1.303 | iter time: 357.90 ms (step) remaining time: 0:44:51
165
+ Epoch 1 | iter 1696 step 53 | loss train: 1.319, val: 1.303 | iter time: 360.23 ms (step) remaining time: 0:44:38
166
+ Epoch 1 | iter 1728 step 54 | loss train: 1.359, val: 1.303 | iter time: 360.56 ms (step) remaining time: 0:44:25
167
+ Epoch 1 | iter 1760 step 55 | loss train: 1.322, val: 1.303 | iter time: 809.47 ms (step) remaining time: 0:44:13
168
+ Epoch 1 | iter 1792 step 56 | loss train: 1.336, val: 1.303 | iter time: 358.64 ms (step) remaining time: 0:44:00
169
+ Epoch 1 | iter 1824 step 57 | loss train: 1.309, val: 1.303 | iter time: 360.66 ms (step) remaining time: 0:43:47
170
+ Epoch 1 | iter 1856 step 58 | loss train: 1.324, val: 1.303 | iter time: 360.01 ms (step) remaining time: 0:43:34
171
+ Epoch 1 | iter 1888 step 59 | loss train: 1.394, val: 1.303 | iter time: 360.62 ms (step) remaining time: 0:43:21
172
+ Epoch 1 | iter 1920 step 60 | loss train: 1.447, val: 1.303 | iter time: 361.11 ms (step) remaining time: 0:43:09
173
+ Epoch 1 | iter 1952 step 61 | loss train: 1.324, val: 1.303 | iter time: 360.09 ms (step) remaining time: 0:42:56
174
+ Epoch 1 | iter 1984 step 62 | loss train: 1.373, val: 1.303 | iter time: 359.53 ms (step) remaining time: 0:42:43
175
+ Epoch 1 | iter 2016 step 63 | loss train: 1.415, val: 1.303 | iter time: 359.52 ms (step) remaining time: 0:42:31
176
+ Epoch 1 | iter 2048 step 64 | loss train: 1.373, val: 1.303 | iter time: 358.80 ms (step) remaining time: 0:42:18
177
+ Epoch 1 | iter 2080 step 65 | loss train: 1.323, val: 1.303 | iter time: 360.27 ms (step) remaining time: 0:42:06
178
+ Epoch 1 | iter 2112 step 66 | loss train: 1.322, val: 1.303 | iter time: 358.27 ms (step) remaining time: 0:41:53
179
+ Epoch 1 | iter 2144 step 67 | loss train: 1.304, val: 1.303 | iter time: 359.32 ms (step) remaining time: 0:41:41
180
+ Epoch 1 | iter 2176 step 68 | loss train: 1.318, val: 1.303 | iter time: 360.43 ms (step) remaining time: 0:41:28
181
+ Epoch 1 | iter 2208 step 69 | loss train: 1.369, val: 1.303 | iter time: 359.58 ms (step) remaining time: 0:41:16
182
+ Epoch 1 | iter 2240 step 70 | loss train: 1.404, val: 1.303 | iter time: 359.33 ms (step) remaining time: 0:41:04
183
+ Epoch 1 | iter 2272 step 71 | loss train: 1.351, val: 1.303 | iter time: 358.02 ms (step) remaining time: 0:40:52
184
+ Epoch 1 | iter 2304 step 72 | loss train: 1.377, val: 1.303 | iter time: 359.25 ms (step) remaining time: 0:40:40
185
+ Epoch 1 | iter 2336 step 73 | loss train: 1.363, val: 1.303 | iter time: 359.37 ms (step) remaining time: 0:40:27
186
+ Epoch 1 | iter 2368 step 74 | loss train: 1.281, val: 1.303 | iter time: 359.60 ms (step) remaining time: 0:40:15
187
+ Epoch 1 | iter 2400 step 75 | loss train: 1.375, val: 1.303 | iter time: 361.50 ms (step) remaining time: 0:40:03
188
+ Epoch 1 | iter 2432 step 76 | loss train: 1.347, val: 1.303 | iter time: 359.78 ms (step) remaining time: 0:39:51
189
+ Epoch 1 | iter 2464 step 77 | loss train: 1.368, val: 1.303 | iter time: 359.86 ms (step) remaining time: 0:39:39
190
+ Epoch 1 | iter 2496 step 78 | loss train: 1.347, val: 1.303 | iter time: 359.10 ms (step) remaining time: 0:39:27
191
+ Epoch 1 | iter 2528 step 79 | loss train: 1.366, val: 1.303 | iter time: 359.23 ms (step) remaining time: 0:39:15
192
+ Epoch 1 | iter 2560 step 80 | loss train: 1.331, val: 1.303 | iter time: 358.89 ms (step) remaining time: 0:39:03
193
+ Epoch 1 | iter 2592 step 81 | loss train: 1.390, val: 1.303 | iter time: 361.51 ms (step) remaining time: 0:38:51
194
+ Epoch 1 | iter 2624 step 82 | loss train: 1.341, val: 1.303 | iter time: 360.83 ms (step) remaining time: 0:38:39
195
+ Epoch 1 | iter 2656 step 83 | loss train: 1.359, val: 1.303 | iter time: 360.05 ms (step) remaining time: 0:38:28
196
+ Epoch 1 | iter 2688 step 84 | loss train: 1.384, val: 1.303 | iter time: 359.42 ms (step) remaining time: 0:38:16
197
+ Epoch 1 | iter 2720 step 85 | loss train: 1.318, val: 1.303 | iter time: 358.95 ms (step) remaining time: 0:38:04
198
+ Epoch 1 | iter 2752 step 86 | loss train: 1.457, val: 1.303 | iter time: 359.61 ms (step) remaining time: 0:37:52
199
+ Epoch 1 | iter 2784 step 87 | loss train: 1.348, val: 1.303 | iter time: 358.33 ms (step) remaining time: 0:37:40
200
+ Epoch 1 | iter 2816 step 88 | loss train: 1.373, val: 1.303 | iter time: 359.67 ms (step) remaining time: 0:37:29
201
+ Epoch 1 | iter 2848 step 89 | loss train: 1.376, val: 1.303 | iter time: 359.15 ms (step) remaining time: 0:37:17
202
+ Epoch 1 | iter 2880 step 90 | loss train: 1.308, val: 1.303 | iter time: 360.99 ms (step) remaining time: 0:37:05
203
+ Epoch 1 | iter 2912 step 91 | loss train: 1.350, val: 1.303 | iter time: 359.87 ms (step) remaining time: 0:36:54
204
+ Epoch 1 | iter 2944 step 92 | loss train: 1.389, val: 1.303 | iter time: 360.11 ms (step) remaining time: 0:36:42
205
+ Epoch 1 | iter 2976 step 93 | loss train: 1.359, val: 1.303 | iter time: 360.67 ms (step) remaining time: 0:36:30
206
+ Epoch 1 | iter 3008 step 94 | loss train: 1.400, val: 1.303 | iter time: 359.70 ms (step) remaining time: 0:36:19
207
+ Epoch 1 | iter 3040 step 95 | loss train: 1.350, val: 1.303 | iter time: 360.29 ms (step) remaining time: 0:36:07
208
+ Epoch 1 | iter 3072 step 96 | loss train: 1.348, val: 1.303 | iter time: 361.10 ms (step) remaining time: 0:35:55
209
+ Epoch 1 | iter 3104 step 97 | loss train: 1.299, val: 1.303 | iter time: 360.33 ms (step) remaining time: 0:35:44
210
+ Epoch 1 | iter 3136 step 98 | loss train: 1.291, val: 1.303 | iter time: 360.33 ms (step) remaining time: 0:35:32
211
+ Epoch 1 | iter 3168 step 99 | loss train: 1.389, val: 1.303 | iter time: 359.23 ms (step) remaining time: 0:35:21
212
+ Epoch 1 | iter 3200 step 100 | loss train: 1.363, val: 1.303 | iter time: 359.37 ms (step) remaining time: 0:35:09
213
+ Validating ...
214
+ iter 3200: val loss 1.2997, val time: 21941.50 ms
215
+ Saving checkpoint to 'out/pretrain/tinyllama/2412_lr4e-5/step-00000100/lit_model.pth'
216
+ Epoch 1 | iter 3232 step 101 | loss train: 1.409, val: 1.300 | iter time: 356.05 ms (step) remaining time: 0:36:09
217
+ Epoch 1 | iter 3264 step 102 | loss train: 1.378, val: 1.300 | iter time: 358.00 ms (step) remaining time: 0:35:56
218
+ Epoch 1 | iter 3296 step 103 | loss train: 1.386, val: 1.300 | iter time: 360.82 ms (step) remaining time: 0:35:43
219
+ Epoch 1 | iter 3328 step 104 | loss train: 1.346, val: 1.300 | iter time: 359.41 ms (step) remaining time: 0:35:31
220
+ Epoch 1 | iter 3360 step 105 | loss train: 1.372, val: 1.300 | iter time: 360.93 ms (step) remaining time: 0:35:18
221
+ Epoch 1 | iter 3392 step 106 | loss train: 1.320, val: 1.300 | iter time: 359.57 ms (step) remaining time: 0:35:06
222
+ Epoch 1 | iter 3424 step 107 | loss train: 1.369, val: 1.300 | iter time: 358.71 ms (step) remaining time: 0:34:53
223
+ Epoch 1 | iter 3456 step 108 | loss train: 1.364, val: 1.300 | iter time: 359.31 ms (step) remaining time: 0:34:41
224
+ Epoch 1 | iter 3488 step 109 | loss train: 1.312, val: 1.300 | iter time: 361.07 ms (step) remaining time: 0:34:28
225
+ Epoch 1 | iter 3520 step 110 | loss train: 1.353, val: 1.300 | iter time: 358.59 ms (step) remaining time: 0:34:16
226
+ Epoch 1 | iter 3552 step 111 | loss train: 1.331, val: 1.300 | iter time: 359.16 ms (step) remaining time: 0:34:04
227
+ Epoch 1 | iter 3584 step 112 | loss train: 1.360, val: 1.300 | iter time: 357.47 ms (step) remaining time: 0:33:51
228
+ Epoch 1 | iter 3616 step 113 | loss train: 1.284, val: 1.300 | iter time: 358.67 ms (step) remaining time: 0:33:40
229
+ Epoch 1 | iter 3648 step 114 | loss train: 1.303, val: 1.300 | iter time: 360.30 ms (step) remaining time: 0:33:27
230
+ Epoch 1 | iter 3680 step 115 | loss train: 1.309, val: 1.300 | iter time: 360.21 ms (step) remaining time: 0:33:15
231
+ Epoch 1 | iter 3712 step 116 | loss train: 1.314, val: 1.300 | iter time: 360.69 ms (step) remaining time: 0:33:03
232
+ Epoch 1 | iter 3744 step 117 | loss train: 1.395, val: 1.300 | iter time: 359.26 ms (step) remaining time: 0:32:51
233
+ Epoch 1 | iter 3776 step 118 | loss train: 1.342, val: 1.300 | iter time: 358.67 ms (step) remaining time: 0:32:38
234
+ Epoch 1 | iter 3808 step 119 | loss train: 1.335, val: 1.300 | iter time: 360.89 ms (step) remaining time: 0:32:26
235
+ Epoch 1 | iter 3840 step 120 | loss train: 1.376, val: 1.300 | iter time: 359.61 ms (step) remaining time: 0:32:14
236
+ Epoch 1 | iter 3872 step 121 | loss train: 1.390, val: 1.300 | iter time: 360.92 ms (step) remaining time: 0:32:02
237
+ Epoch 1 | iter 3904 step 122 | loss train: 1.353, val: 1.300 | iter time: 362.29 ms (step) remaining time: 0:31:50
238
+ Epoch 1 | iter 3936 step 123 | loss train: 1.321, val: 1.300 | iter time: 358.52 ms (step) remaining time: 0:31:38
239
+ Epoch 1 | iter 3968 step 124 | loss train: 1.308, val: 1.300 | iter time: 359.64 ms (step) remaining time: 0:31:26
240
+ Epoch 1 | iter 4000 step 125 | loss train: 1.405, val: 1.300 | iter time: 361.00 ms (step) remaining time: 0:31:15
241
+ Epoch 1 | iter 4032 step 126 | loss train: 1.355, val: 1.300 | iter time: 359.20 ms (step) remaining time: 0:31:03
242
+ Epoch 1 | iter 4064 step 127 | loss train: 1.379, val: 1.300 | iter time: 360.45 ms (step) remaining time: 0:30:51
243
+ Epoch 1 | iter 4096 step 128 | loss train: 1.381, val: 1.300 | iter time: 354.76 ms (step) remaining time: 0:30:39
244
+ Epoch 1 | iter 4128 step 129 | loss train: 1.356, val: 1.300 | iter time: 360.52 ms (step) remaining time: 0:30:27
245
+ Epoch 1 | iter 4160 step 130 | loss train: 1.332, val: 1.300 | iter time: 360.45 ms (step) remaining time: 0:30:15
246
+ Epoch 1 | iter 4192 step 131 | loss train: 1.392, val: 1.300 | iter time: 360.49 ms (step) remaining time: 0:30:03
247
+ Epoch 1 | iter 4224 step 132 | loss train: 1.309, val: 1.300 | iter time: 361.96 ms (step) remaining time: 0:29:51
248
+ Epoch 1 | iter 4256 step 133 | loss train: 1.348, val: 1.300 | iter time: 358.25 ms (step) remaining time: 0:29:39
249
+ Epoch 1 | iter 4288 step 134 | loss train: 1.386, val: 1.300 | iter time: 358.28 ms (step) remaining time: 0:29:27
250
+ Epoch 1 | iter 4320 step 135 | loss train: 1.332, val: 1.300 | iter time: 358.33 ms (step) remaining time: 0:29:15
251
+ Epoch 1 | iter 4352 step 136 | loss train: 1.395, val: 1.300 | iter time: 358.01 ms (step) remaining time: 0:29:04
252
+ Epoch 1 | iter 4384 step 137 | loss train: 1.333, val: 1.300 | iter time: 359.89 ms (step) remaining time: 0:28:52
253
+ Epoch 1 | iter 4416 step 138 | loss train: 1.349, val: 1.300 | iter time: 360.63 ms (step) remaining time: 0:28:40
254
+ Epoch 1 | iter 4448 step 139 | loss train: 1.407, val: 1.300 | iter time: 361.40 ms (step) remaining time: 0:28:28
255
+ Epoch 1 | iter 4480 step 140 | loss train: 1.380, val: 1.300 | iter time: 359.52 ms (step) remaining time: 0:28:16
256
+ Epoch 1 | iter 4512 step 141 | loss train: 1.394, val: 1.300 | iter time: 358.69 ms (step) remaining time: 0:28:05
257
+ Epoch 1 | iter 4544 step 142 | loss train: 1.403, val: 1.300 | iter time: 357.73 ms (step) remaining time: 0:27:53
258
+ Epoch 1 | iter 4576 step 143 | loss train: 1.450, val: 1.300 | iter time: 359.22 ms (step) remaining time: 0:27:41
259
+ Epoch 1 | iter 4608 step 144 | loss train: 1.413, val: 1.300 | iter time: 358.98 ms (step) remaining time: 0:27:29
260
+ Epoch 1 | iter 4640 step 145 | loss train: 1.424, val: 1.300 | iter time: 360.01 ms (step) remaining time: 0:27:18
261
+ Epoch 1 | iter 4672 step 146 | loss train: 1.338, val: 1.300 | iter time: 359.50 ms (step) remaining time: 0:27:06
262
+ Epoch 1 | iter 4704 step 147 | loss train: 1.383, val: 1.300 | iter time: 359.39 ms (step) remaining time: 0:26:54
263
+ Epoch 1 | iter 4736 step 148 | loss train: 1.414, val: 1.300 | iter time: 358.73 ms (step) remaining time: 0:26:43
264
+ Epoch 1 | iter 4768 step 149 | loss train: 1.394, val: 1.300 | iter time: 359.10 ms (step) remaining time: 0:26:31
265
+ Epoch 1 | iter 4800 step 150 | loss train: 1.386, val: 1.300 | iter time: 359.23 ms (step) remaining time: 0:26:19
266
+ Validating ...
267
+ iter 4800: val loss 1.2971, val time: 21947.04 ms
268
+ Epoch 1 | iter 4832 step 151 | loss train: 1.409, val: 1.297 | iter time: 359.90 ms (step) remaining time: 0:26:28
269
+ Epoch 1 | iter 4864 step 152 | loss train: 1.348, val: 1.297 | iter time: 359.18 ms (step) remaining time: 0:26:16
270
+ Epoch 1 | iter 4896 step 153 | loss train: 1.340, val: 1.297 | iter time: 358.93 ms (step) remaining time: 0:26:04
271
+ Epoch 1 | iter 4928 step 154 | loss train: 1.445, val: 1.297 | iter time: 360.32 ms (step) remaining time: 0:25:52
272
+ Epoch 1 | iter 4960 step 155 | loss train: 1.437, val: 1.297 | iter time: 359.25 ms (step) remaining time: 0:25:40
273
+ Epoch 1 | iter 4992 step 156 | loss train: 1.339, val: 1.297 | iter time: 359.60 ms (step) remaining time: 0:25:28
274
+ Epoch 1 | iter 5024 step 157 | loss train: 1.345, val: 1.297 | iter time: 360.62 ms (step) remaining time: 0:25:16
275
+ Epoch 1 | iter 5056 step 158 | loss train: 1.361, val: 1.297 | iter time: 358.48 ms (step) remaining time: 0:25:05
276
+ Epoch 1 | iter 5088 step 159 | loss train: 1.368, val: 1.297 | iter time: 360.55 ms (step) remaining time: 0:24:53
277
+ Epoch 1 | iter 5120 step 160 | loss train: 1.329, val: 1.297 | iter time: 360.70 ms (step) remaining time: 0:24:41
278
+ Epoch 1 | iter 5152 step 161 | loss train: 1.374, val: 1.297 | iter time: 357.78 ms (step) remaining time: 0:24:29
279
+ Epoch 1 | iter 5184 step 162 | loss train: 1.464, val: 1.297 | iter time: 360.20 ms (step) remaining time: 0:24:17
280
+ Epoch 1 | iter 5216 step 163 | loss train: 1.368, val: 1.297 | iter time: 360.98 ms (step) remaining time: 0:24:05
281
+ Epoch 1 | iter 5248 step 164 | loss train: 1.447, val: 1.297 | iter time: 360.17 ms (step) remaining time: 0:23:54
282
+ Epoch 1 | iter 5280 step 165 | loss train: 1.434, val: 1.297 | iter time: 358.76 ms (step) remaining time: 0:23:42
283
+ Epoch 1 | iter 5312 step 166 | loss train: 1.419, val: 1.297 | iter time: 359.84 ms (step) remaining time: 0:23:30
284
+ Epoch 1 | iter 5344 step 167 | loss train: 1.411, val: 1.297 | iter time: 359.60 ms (step) remaining time: 0:23:18
285
+ Epoch 1 | iter 5376 step 168 | loss train: 1.357, val: 1.297 | iter time: 359.87 ms (step) remaining time: 0:23:07
286
+ Epoch 1 | iter 5408 step 169 | loss train: 1.365, val: 1.297 | iter time: 359.00 ms (step) remaining time: 0:22:55
287
+ Epoch 1 | iter 5440 step 170 | loss train: 1.300, val: 1.297 | iter time: 360.21 ms (step) remaining time: 0:22:43
288
+ Epoch 1 | iter 5472 step 171 | loss train: 1.318, val: 1.297 | iter time: 361.20 ms (step) remaining time: 0:22:32
289
+ Epoch 1 | iter 5504 step 172 | loss train: 1.367, val: 1.297 | iter time: 360.54 ms (step) remaining time: 0:22:20
290
+ Epoch 1 | iter 5536 step 173 | loss train: 1.385, val: 1.297 | iter time: 358.10 ms (step) remaining time: 0:22:08
291
+ Epoch 1 | iter 5568 step 174 | loss train: 1.393, val: 1.297 | iter time: 359.94 ms (step) remaining time: 0:21:56
292
+ Epoch 1 | iter 5600 step 175 | loss train: 1.387, val: 1.297 | iter time: 358.91 ms (step) remaining time: 0:21:45
293
+ Epoch 1 | iter 5632 step 176 | loss train: 1.371, val: 1.297 | iter time: 358.64 ms (step) remaining time: 0:21:33
294
+ Epoch 1 | iter 5664 step 177 | loss train: 1.347, val: 1.297 | iter time: 361.96 ms (step) remaining time: 0:21:21
295
+ Epoch 1 | iter 5696 step 178 | loss train: 1.445, val: 1.297 | iter time: 360.14 ms (step) remaining time: 0:21:10
296
+ Epoch 1 | iter 5728 step 179 | loss train: 1.352, val: 1.297 | iter time: 359.23 ms (step) remaining time: 0:20:58
297
+ Epoch 1 | iter 5760 step 180 | loss train: 1.355, val: 1.297 | iter time: 359.65 ms (step) remaining time: 0:20:47
298
+ Epoch 1 | iter 5792 step 181 | loss train: 1.381, val: 1.297 | iter time: 359.39 ms (step) remaining time: 0:20:35
299
+ Epoch 1 | iter 5824 step 182 | loss train: 1.354, val: 1.297 | iter time: 358.25 ms (step) remaining time: 0:20:23
300
+ Epoch 1 | iter 5856 step 183 | loss train: 1.379, val: 1.297 | iter time: 359.01 ms (step) remaining time: 0:20:12
301
+ Epoch 1 | iter 5888 step 184 | loss train: 1.398, val: 1.297 | iter time: 357.69 ms (step) remaining time: 0:20:00
302
+ Epoch 1 | iter 5920 step 185 | loss train: 1.347, val: 1.297 | iter time: 361.26 ms (step) remaining time: 0:19:49
303
+ Epoch 1 | iter 5952 step 186 | loss train: 1.393, val: 1.297 | iter time: 358.92 ms (step) remaining time: 0:19:37
304
+ Epoch 1 | iter 5984 step 187 | loss train: 1.402, val: 1.297 | iter time: 359.91 ms (step) remaining time: 0:19:26
305
+ Epoch 1 | iter 6016 step 188 | loss train: 1.469, val: 1.297 | iter time: 359.22 ms (step) remaining time: 0:19:14
306
+ Epoch 1 | iter 6048 step 189 | loss train: 1.360, val: 1.297 | iter time: 360.93 ms (step) remaining time: 0:19:02
307
+ Epoch 1 | iter 6080 step 190 | loss train: 1.397, val: 1.297 | iter time: 361.50 ms (step) remaining time: 0:18:51
308
+ Epoch 1 | iter 6112 step 191 | loss train: 1.342, val: 1.297 | iter time: 358.85 ms (step) remaining time: 0:18:39
309
+ Epoch 1 | iter 6144 step 192 | loss train: 1.362, val: 1.297 | iter time: 358.55 ms (step) remaining time: 0:18:28
310
+ Epoch 1 | iter 6176 step 193 | loss train: 1.396, val: 1.297 | iter time: 360.94 ms (step) remaining time: 0:18:16
311
+ Epoch 1 | iter 6208 step 194 | loss train: 1.442, val: 1.297 | iter time: 607.16 ms (step) remaining time: 0:18:05
312
+ Epoch 1 | iter 6240 step 195 | loss train: 1.380, val: 1.297 | iter time: 358.51 ms (step) remaining time: 0:17:53
313
+ Epoch 1 | iter 6272 step 196 | loss train: 1.402, val: 1.297 | iter time: 358.22 ms (step) remaining time: 0:17:42
314
+ Epoch 1 | iter 6304 step 197 | loss train: 1.315, val: 1.297 | iter time: 359.23 ms (step) remaining time: 0:17:30
315
+ Epoch 1 | iter 6336 step 198 | loss train: 1.343, val: 1.297 | iter time: 358.04 ms (step) remaining time: 0:17:19
316
+ Epoch 1 | iter 6368 step 199 | loss train: 1.394, val: 1.297 | iter time: 359.15 ms (step) remaining time: 0:17:07
317
+ Epoch 1 | iter 6400 step 200 | loss train: 1.381, val: 1.297 | iter time: 360.85 ms (step) remaining time: 0:16:56
318
+ Validating ...
319
+ iter 6400: val loss 1.2924, val time: 21942.30 ms
320
+ Saving checkpoint to 'out/pretrain/tinyllama/2412_lr4e-5/step-00000200/lit_model.pth'
321
+ Epoch 1 | iter 6432 step 201 | loss train: 1.364, val: 1.292 | iter time: 356.68 ms (step) remaining time: 0:17:01
322
+ Epoch 1 | iter 6464 step 202 | loss train: 1.424, val: 1.292 | iter time: 357.73 ms (step) remaining time: 0:16:49
323
+ Epoch 1 | iter 6496 step 203 | loss train: 1.355, val: 1.292 | iter time: 358.41 ms (step) remaining time: 0:16:38
324
+ Epoch 1 | iter 6528 step 204 | loss train: 1.373, val: 1.292 | iter time: 359.90 ms (step) remaining time: 0:16:26
325
+ Epoch 1 | iter 6560 step 205 | loss train: 1.342, val: 1.292 | iter time: 359.21 ms (step) remaining time: 0:16:14
326
+ Epoch 1 | iter 6592 step 206 | loss train: 1.420, val: 1.292 | iter time: 357.64 ms (step) remaining time: 0:16:02
327
+ Epoch 1 | iter 6624 step 207 | loss train: 1.355, val: 1.292 | iter time: 360.14 ms (step) remaining time: 0:15:51
328
+ Epoch 1 | iter 6656 step 208 | loss train: 1.328, val: 1.292 | iter time: 358.94 ms (step) remaining time: 0:15:39
329
+ Epoch 1 | iter 6688 step 209 | loss train: 1.408, val: 1.292 | iter time: 358.67 ms (step) remaining time: 0:15:27
330
+ Epoch 1 | iter 6720 step 210 | loss train: 1.340, val: 1.292 | iter time: 358.83 ms (step) remaining time: 0:15:16
331
+ Epoch 1 | iter 6752 step 211 | loss train: 1.366, val: 1.292 | iter time: 360.99 ms (step) remaining time: 0:15:04
332
+ Epoch 1 | iter 6784 step 212 | loss train: 1.425, val: 1.292 | iter time: 360.72 ms (step) remaining time: 0:14:52
333
+ Epoch 1 | iter 6816 step 213 | loss train: 1.419, val: 1.292 | iter time: 359.38 ms (step) remaining time: 0:14:41
334
+ Epoch 1 | iter 6848 step 214 | loss train: 1.332, val: 1.292 | iter time: 357.71 ms (step) remaining time: 0:14:29
335
+ Epoch 1 | iter 6880 step 215 | loss train: 1.437, val: 1.292 | iter time: 360.60 ms (step) remaining time: 0:14:17
336
+ Epoch 1 | iter 6912 step 216 | loss train: 1.399, val: 1.292 | iter time: 359.07 ms (step) remaining time: 0:14:06
337
+ Epoch 1 | iter 6944 step 217 | loss train: 1.338, val: 1.292 | iter time: 360.92 ms (step) remaining time: 0:13:54
338
+ Epoch 1 | iter 6976 step 218 | loss train: 1.336, val: 1.292 | iter time: 359.81 ms (step) remaining time: 0:13:43
339
+ Epoch 1 | iter 7008 step 219 | loss train: 1.346, val: 1.292 | iter time: 360.16 ms (step) remaining time: 0:13:31
340
+ Epoch 1 | iter 7040 step 220 | loss train: 1.305, val: 1.292 | iter time: 359.56 ms (step) remaining time: 0:13:19
341
+ Epoch 1 | iter 7072 step 221 | loss train: 1.328, val: 1.292 | iter time: 358.76 ms (step) remaining time: 0:13:08
342
+ Epoch 1 | iter 7104 step 222 | loss train: 1.400, val: 1.292 | iter time: 360.18 ms (step) remaining time: 0:12:56
343
+ Epoch 1 | iter 7136 step 223 | loss train: 1.305, val: 1.292 | iter time: 359.27 ms (step) remaining time: 0:12:45
344
+ Epoch 1 | iter 7168 step 224 | loss train: 1.370, val: 1.292 | iter time: 358.90 ms (step) remaining time: 0:12:33
345
+ Epoch 1 | iter 7200 step 225 | loss train: 1.322, val: 1.292 | iter time: 360.03 ms (step) remaining time: 0:12:21
346
+ Epoch 1 | iter 7232 step 226 | loss train: 1.406, val: 1.292 | iter time: 360.34 ms (step) remaining time: 0:12:10
347
+ Epoch 1 | iter 7264 step 227 | loss train: 1.360, val: 1.292 | iter time: 358.39 ms (step) remaining time: 0:11:58
348
+ Epoch 1 | iter 7296 step 228 | loss train: 1.296, val: 1.292 | iter time: 360.99 ms (step) remaining time: 0:11:47
349
+ Epoch 1 | iter 7328 step 229 | loss train: 1.506, val: 1.292 | iter time: 358.73 ms (step) remaining time: 0:11:35
350
+ Epoch 1 | iter 7360 step 230 | loss train: 1.343, val: 1.292 | iter time: 359.70 ms (step) remaining time: 0:11:24
351
+ Epoch 1 | iter 7392 step 231 | loss train: 1.392, val: 1.292 | iter time: 360.83 ms (step) remaining time: 0:11:12
352
+ Epoch 1 | iter 7424 step 232 | loss train: 1.325, val: 1.292 | iter time: 359.22 ms (step) remaining time: 0:11:01
353
+ Epoch 1 | iter 7456 step 233 | loss train: 1.350, val: 1.292 | iter time: 360.45 ms (step) remaining time: 0:10:49
354
+ Epoch 1 | iter 7488 step 234 | loss train: 1.432, val: 1.292 | iter time: 360.91 ms (step) remaining time: 0:10:38
355
+ Epoch 1 | iter 7520 step 235 | loss train: 1.354, val: 1.292 | iter time: 358.74 ms (step) remaining time: 0:10:26
356
+ Epoch 1 | iter 7552 step 236 | loss train: 1.391, val: 1.292 | iter time: 360.07 ms (step) remaining time: 0:10:14
357
+ Epoch 1 | iter 7584 step 237 | loss train: 1.333, val: 1.292 | iter time: 359.95 ms (step) remaining time: 0:10:03
358
+ Epoch 1 | iter 7616 step 238 | loss train: 1.344, val: 1.292 | iter time: 359.28 ms (step) remaining time: 0:09:51
359
+ Epoch 1 | iter 7648 step 239 | loss train: 1.386, val: 1.292 | iter time: 362.00 ms (step) remaining time: 0:09:40
360
+ Epoch 1 | iter 7680 step 240 | loss train: 1.430, val: 1.292 | iter time: 359.55 ms (step) remaining time: 0:09:29
361
+ Epoch 1 | iter 7712 step 241 | loss train: 1.328, val: 1.292 | iter time: 359.32 ms (step) remaining time: 0:09:17
362
+ Epoch 1 | iter 7744 step 242 | loss train: 1.419, val: 1.292 | iter time: 360.56 ms (step) remaining time: 0:09:06
363
+ Epoch 1 | iter 7776 step 243 | loss train: 1.343, val: 1.292 | iter time: 361.35 ms (step) remaining time: 0:08:54
364
+ Epoch 1 | iter 7808 step 244 | loss train: 1.374, val: 1.292 | iter time: 360.30 ms (step) remaining time: 0:08:43
365
+ Epoch 1 | iter 7840 step 245 | loss train: 1.376, val: 1.292 | iter time: 361.11 ms (step) remaining time: 0:08:31
366
+ Epoch 1 | iter 7872 step 246 | loss train: 1.384, val: 1.292 | iter time: 359.56 ms (step) remaining time: 0:08:20
367
+ Epoch 1 | iter 7904 step 247 | loss train: 1.308, val: 1.292 | iter time: 359.60 ms (step) remaining time: 0:08:08
368
+ Epoch 1 | iter 7936 step 248 | loss train: 1.346, val: 1.292 | iter time: 358.91 ms (step) remaining time: 0:07:57
369
+ Epoch 1 | iter 7968 step 249 | loss train: 1.375, val: 1.292 | iter time: 359.99 ms (step) remaining time: 0:07:45
370
+ Epoch 1 | iter 8000 step 250 | loss train: 1.336, val: 1.292 | iter time: 360.09 ms (step) remaining time: 0:07:34
371
+ Validating ...
372
+ iter 8000: val loss 1.2909, val time: 21936.55 ms
373
+ Epoch 1 | iter 8032 step 251 | loss train: 1.326, val: 1.291 | iter time: 360.83 ms (step) remaining time: 0:07:26
374
+ Epoch 1 | iter 8064 step 252 | loss train: 1.370, val: 1.291 | iter time: 359.27 ms (step) remaining time: 0:07:14
375
+ Epoch 1 | iter 8096 step 253 | loss train: 1.377, val: 1.291 | iter time: 358.56 ms (step) remaining time: 0:07:03
376
+ Epoch 1 | iter 8128 step 254 | loss train: 1.378, val: 1.291 | iter time: 358.74 ms (step) remaining time: 0:06:51
377
+ Epoch 1 | iter 8160 step 255 | loss train: 1.377, val: 1.291 | iter time: 358.21 ms (step) remaining time: 0:06:40
378
+ Epoch 1 | iter 8192 step 256 | loss train: 1.409, val: 1.291 | iter time: 358.67 ms (step) remaining time: 0:06:28
379
+ Epoch 1 | iter 8224 step 257 | loss train: 1.385, val: 1.291 | iter time: 358.52 ms (step) remaining time: 0:06:17
380
+ Epoch 1 | iter 8256 step 258 | loss train: 1.357, val: 1.291 | iter time: 359.40 ms (step) remaining time: 0:06:05
381
+ Epoch 1 | iter 8288 step 259 | loss train: 1.464, val: 1.291 | iter time: 360.44 ms (step) remaining time: 0:05:54
382
+ Epoch 1 | iter 8320 step 260 | loss train: 1.405, val: 1.291 | iter time: 360.00 ms (step) remaining time: 0:05:42
383
+ Epoch 1 | iter 8352 step 261 | loss train: 1.397, val: 1.291 | iter time: 358.48 ms (step) remaining time: 0:05:31
384
+ Epoch 1 | iter 8384 step 262 | loss train: 1.367, val: 1.291 | iter time: 358.87 ms (step) remaining time: 0:05:19
385
+ Epoch 1 | iter 8416 step 263 | loss train: 1.448, val: 1.291 | iter time: 356.68 ms (step) remaining time: 0:05:08
386
+ Epoch 1 | iter 8448 step 264 | loss train: 1.280, val: 1.291 | iter time: 360.27 ms (step) remaining time: 0:04:56
387
+ Epoch 1 | iter 8480 step 265 | loss train: 1.335, val: 1.291 | iter time: 360.53 ms (step) remaining time: 0:04:45
388
+ Epoch 1 | iter 8512 step 266 | loss train: 1.367, val: 1.291 | iter time: 357.93 ms (step) remaining time: 0:04:33
389
+ Epoch 1 | iter 8544 step 267 | loss train: 1.408, val: 1.291 | iter time: 360.38 ms (step) remaining time: 0:04:22
390
+ Epoch 1 | iter 8576 step 268 | loss train: 1.362, val: 1.291 | iter time: 359.66 ms (step) remaining time: 0:04:11
391
+ Epoch 1 | iter 8608 step 269 | loss train: 1.410, val: 1.291 | iter time: 360.35 ms (step) remaining time: 0:03:59
392
+ Epoch 1 | iter 8640 step 270 | loss train: 1.342, val: 1.291 | iter time: 359.21 ms (step) remaining time: 0:03:48
393
+ Epoch 1 | iter 8672 step 271 | loss train: 1.380, val: 1.291 | iter time: 360.04 ms (step) remaining time: 0:03:36
394
+ Epoch 1 | iter 8704 step 272 | loss train: 1.380, val: 1.291 | iter time: 358.57 ms (step) remaining time: 0:03:25
395
+ Epoch 1 | iter 8736 step 273 | loss train: 1.305, val: 1.291 | iter time: 358.47 ms (step) remaining time: 0:03:13
396
+ Epoch 1 | iter 8768 step 274 | loss train: 1.349, val: 1.291 | iter time: 359.82 ms (step) remaining time: 0:03:02
397
+ Epoch 1 | iter 8800 step 275 | loss train: 1.379, val: 1.291 | iter time: 359.03 ms (step) remaining time: 0:02:50
398
+ Epoch 1 | iter 8832 step 276 | loss train: 1.371, val: 1.291 | iter time: 360.44 ms (step) remaining time: 0:02:39
399
+ Epoch 1 | iter 8864 step 277 | loss train: 1.441, val: 1.291 | iter time: 360.93 ms (step) remaining time: 0:02:28
400
+ Epoch 1 | iter 8896 step 278 | loss train: 1.361, val: 1.291 | iter time: 359.28 ms (step) remaining time: 0:02:16
401
+ Epoch 1 | iter 8928 step 279 | loss train: 1.410, val: 1.291 | iter time: 359.84 ms (step) remaining time: 0:02:05
402
+ Epoch 1 | iter 8960 step 280 | loss train: 1.382, val: 1.291 | iter time: 359.06 ms (step) remaining time: 0:01:53
403
+ Epoch 1 | iter 8992 step 281 | loss train: 1.391, val: 1.291 | iter time: 358.93 ms (step) remaining time: 0:01:42
404
+ Epoch 1 | iter 9024 step 282 | loss train: 1.405, val: 1.291 | iter time: 361.64 ms (step) remaining time: 0:01:31
405
+ Epoch 1 | iter 9056 step 283 | loss train: 1.283, val: 1.291 | iter time: 360.10 ms (step) remaining time: 0:01:19
406
+ Epoch 1 | iter 9088 step 284 | loss train: 1.435, val: 1.291 | iter time: 359.89 ms (step) remaining time: 0:01:08
407
+ Epoch 1 | iter 9120 step 285 | loss train: 1.389, val: 1.291 | iter time: 359.18 ms (step) remaining time: 0:00:56
408
+ Epoch 1 | iter 9152 step 286 | loss train: 1.360, val: 1.291 | iter time: 359.35 ms (step) remaining time: 0:00:45
409
+ Epoch 1 | iter 9184 step 287 | loss train: 1.383, val: 1.291 | iter time: 359.96 ms (step) remaining time: 0:00:34
410
+ Epoch 1 | iter 9216 step 288 | loss train: 1.421, val: 1.291 | iter time: 360.37 ms (step) remaining time: 0:00:22
411
+ Epoch 2 | iter 9248 step 289 | loss train: 1.394, val: 1.291 | iter time: 357.32 ms (step) remaining time: 0:00:11
412
+ Epoch 2 | iter 9280 step 290 | loss train: 1.326, val: 1.291 | iter time: 361.11 ms (step) remaining time: 0:00:00
413
+ Validating ...
414
+ Final evaluation | val loss: 1.227 | val ppl: 3.412
415
+ Saving checkpoint to 'out/pretrain/tinyllama/2412_lr4e-5/final/lit_model.pth'
416
+ ----------------------------------------
417
+ | Performance
418
+ | - Total tokens : 304,087,040
419
+ | - Training Time : 3359.33 s
420
+ | - Tok/sec : 156.37 tok/s
421
+ | ----------------------------------------
422
+ | Memory Usage
423
+ | - Memory Used : 26.32 GB
424
+ ----------------------------------------
out/pretrain/tinyllama/teelogs/2501.txt ADDED
@@ -0,0 +1,378 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Initializing distributed: GLOBAL_RANK: 0, MEMBER: 1/4
2
+ Initializing distributed: GLOBAL_RANK: 1, MEMBER: 2/4
3
+ [rank: 1] Seed set to 42
4
+ Initializing distributed: GLOBAL_RANK: 3, MEMBER: 4/4
5
+ [rank: 3] Seed set to 42
6
+ Initializing distributed: GLOBAL_RANK: 2, MEMBER: 3/4
7
+ [rank: 2] Seed set to 42
8
+ ----------------------------------------------------------------------------------------------------
9
+ distributed_backend=nccl
10
+ All distributed processes registered. Starting with 4 processes
11
+ ----------------------------------------------------------------------------------------------------
12
+
13
+ [rank: 0] Seed set to 42
14
+ All GPUs are fully connected via NVLink.
15
+ {'data': {'batch_size': 1,
16
+ 'data_path': PosixPath('litgpt/data/arxiv'),
17
+ 'num_workers': 8,
18
+ 'seed': 42,
19
+ 'seq_length': 2048,
20
+ 'use_starcoder': True},
21
+ 'data_dir': PosixPath('litgpt/data/arxiv/2501'),
22
+ 'devices': 'auto',
23
+ 'eval': {'evaluate_example': 'first',
24
+ 'final_validation': True,
25
+ 'initial_validation': True,
26
+ 'interval': 50,
27
+ 'max_iters': 100,
28
+ 'max_new_tokens': None},
29
+ 'initial_checkpoint_dir': PosixPath('out/pretrain/2412/final'),
30
+ 'log': {'group': None, 'project': None, 'run': None},
31
+ 'logger_name': 'tensorboard',
32
+ 'model_config': {'attention_logit_softcapping': None,
33
+ 'attention_scores_scalar': None,
34
+ 'attn_bias': False,
35
+ 'bias': False,
36
+ 'block_size': 2048,
37
+ 'final_logit_softcapping': None,
38
+ 'gelu_approximate': 'none',
39
+ 'head_size': 64,
40
+ 'hf_config': {'name': 'TinyLlama-1.1B-intermediate-step-1431k-3T',
41
+ 'org': 'TinyLlama'},
42
+ 'intermediate_size': 5632,
43
+ 'lm_head_bias': False,
44
+ 'mlp_class_name': 'LLaMAMLP',
45
+ 'moe_intermediate_size': None,
46
+ 'n_embd': 2048,
47
+ 'n_expert': 0,
48
+ 'n_expert_per_token': 0,
49
+ 'n_head': 32,
50
+ 'n_layer': 22,
51
+ 'n_query_groups': 4,
52
+ 'name': 'tiny-llama-1.1b',
53
+ 'norm_1': True,
54
+ 'norm_2': True,
55
+ 'norm_class_name': 'RMSNorm',
56
+ 'norm_eps': 1e-05,
57
+ 'norm_qk': False,
58
+ 'norm_qk_type': 'default',
59
+ 'padded_vocab_size': 32000,
60
+ 'padding_multiple': 64,
61
+ 'parallel_residual': False,
62
+ 'post_attention_norm': False,
63
+ 'post_mlp_norm': False,
64
+ 'rope_adjustments': None,
65
+ 'rope_base': 10000,
66
+ 'rope_condense_ratio': 1,
67
+ 'rope_indices': None,
68
+ 'rope_local_base_freq': None,
69
+ 'rotary_percentage': 1.0,
70
+ 'scale_embeddings': False,
71
+ 'shared_attention_norm': False,
72
+ 'sliding_window_indices': None,
73
+ 'sliding_window_size': None,
74
+ 'vocab_size': 32000},
75
+ 'model_name': 'tiny-llama-1.1b',
76
+ 'num_nodes': 1,
77
+ 'optimizer': "{'class_path': 'torch.optim.AdamW', 'init_args': {'lr': 0.0004, "
78
+ "'weight_decay': 0.1, 'betas': [0.9, 0.95]}}",
79
+ 'out_dir': PosixPath('out/pretrain/2501'),
80
+ 'precision': 'bf16-mixed',
81
+ 'resume': False,
82
+ 'seed': 42,
83
+ 'tokenizer_dir': PosixPath('checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T'),
84
+ 'train': {'epochs': None,
85
+ 'global_batch_size': 512,
86
+ 'log_interval': 1,
87
+ 'lr_warmup_fraction': None,
88
+ 'lr_warmup_steps': 20,
89
+ 'max_norm': 1.0,
90
+ 'max_seq_length': 2048,
91
+ 'max_steps': None,
92
+ 'max_tokens': 264241152,
93
+ 'micro_batch_size': 4,
94
+ 'min_lr': 4e-05,
95
+ 'save_interval': 100,
96
+ 'tie_embeddings': None}}
97
+ Time to instantiate model: 0.02 seconds.
98
+ Total parameters: 1,100,048,384
99
+ [fix] out/pretrain/2412/final/lit_model.pth 是整包 state,提取 model 权重并原子覆盖...
100
+ [fix] 已覆盖为纯权重: out/pretrain/2412/final/lit_model.pth
101
+ Validating ...
102
+ Measured TFLOPs: 239.66
103
+ Epoch 1 | iter 32 step 1 | loss train: 1.400, val: 1.332 | iter time: 542.67 ms (step) remaining time: 0:49:50
104
+ Epoch 1 | iter 64 step 2 | loss train: 1.346, val: 1.332 | iter time: 354.95 ms (step) remaining time: 0:47:12
105
+ Epoch 1 | iter 96 step 3 | loss train: 1.349, val: 1.332 | iter time: 356.42 ms (step) remaining time: 0:46:14
106
+ Epoch 1 | iter 128 step 4 | loss train: 1.348, val: 1.332 | iter time: 359.29 ms (step) remaining time: 0:45:41
107
+ Epoch 1 | iter 160 step 5 | loss train: 1.309, val: 1.332 | iter time: 357.98 ms (step) remaining time: 0:45:17
108
+ Epoch 1 | iter 192 step 6 | loss train: 1.392, val: 1.332 | iter time: 357.45 ms (step) remaining time: 0:44:59
109
+ Epoch 1 | iter 224 step 7 | loss train: 1.365, val: 1.332 | iter time: 359.96 ms (step) remaining time: 0:44:43
110
+ Epoch 1 | iter 256 step 8 | loss train: 1.357, val: 1.332 | iter time: 360.01 ms (step) remaining time: 0:44:29
111
+ Epoch 1 | iter 288 step 9 | loss train: 1.361, val: 1.332 | iter time: 359.18 ms (step) remaining time: 0:44:15
112
+ Epoch 1 | iter 320 step 10 | loss train: 1.356, val: 1.332 | iter time: 358.22 ms (step) remaining time: 0:44:02
113
+ Epoch 1 | iter 352 step 11 | loss train: 1.460, val: 1.332 | iter time: 357.89 ms (step) remaining time: 0:43:50
114
+ Epoch 1 | iter 384 step 12 | loss train: 1.357, val: 1.332 | iter time: 358.98 ms (step) remaining time: 0:43:38
115
+ Epoch 1 | iter 416 step 13 | loss train: 1.317, val: 1.332 | iter time: 359.74 ms (step) remaining time: 0:43:26
116
+ Epoch 1 | iter 448 step 14 | loss train: 1.365, val: 1.332 | iter time: 359.22 ms (step) remaining time: 0:43:14
117
+ Epoch 1 | iter 480 step 15 | loss train: 1.356, val: 1.332 | iter time: 360.00 ms (step) remaining time: 0:43:02
118
+ Epoch 1 | iter 512 step 16 | loss train: 1.438, val: 1.332 | iter time: 358.80 ms (step) remaining time: 0:42:51
119
+ Epoch 1 | iter 544 step 17 | loss train: 1.357, val: 1.332 | iter time: 360.71 ms (step) remaining time: 0:42:39
120
+ Epoch 1 | iter 576 step 18 | loss train: 1.392, val: 1.332 | iter time: 359.65 ms (step) remaining time: 0:42:28
121
+ Epoch 1 | iter 608 step 19 | loss train: 1.368, val: 1.332 | iter time: 357.81 ms (step) remaining time: 0:42:17
122
+ Epoch 1 | iter 640 step 20 | loss train: 1.386, val: 1.332 | iter time: 358.55 ms (step) remaining time: 0:42:06
123
+ Epoch 1 | iter 672 step 21 | loss train: 1.378, val: 1.332 | iter time: 359.05 ms (step) remaining time: 0:41:54
124
+ Epoch 1 | iter 704 step 22 | loss train: 1.421, val: 1.332 | iter time: 358.99 ms (step) remaining time: 0:41:43
125
+ Epoch 1 | iter 736 step 23 | loss train: 1.430, val: 1.332 | iter time: 358.35 ms (step) remaining time: 0:41:32
126
+ Epoch 1 | iter 768 step 24 | loss train: 1.417, val: 1.332 | iter time: 360.41 ms (step) remaining time: 0:41:21
127
+ Epoch 1 | iter 800 step 25 | loss train: 1.410, val: 1.332 | iter time: 360.08 ms (step) remaining time: 0:41:10
128
+ Epoch 1 | iter 832 step 26 | loss train: 1.370, val: 1.332 | iter time: 359.93 ms (step) remaining time: 0:40:59
129
+ Epoch 1 | iter 864 step 27 | loss train: 1.462, val: 1.332 | iter time: 358.41 ms (step) remaining time: 0:40:48
130
+ Epoch 1 | iter 896 step 28 | loss train: 1.384, val: 1.332 | iter time: 360.39 ms (step) remaining time: 0:40:37
131
+ Epoch 1 | iter 928 step 29 | loss train: 1.419, val: 1.332 | iter time: 359.62 ms (step) remaining time: 0:40:26
132
+ Epoch 1 | iter 960 step 30 | loss train: 1.397, val: 1.332 | iter time: 360.96 ms (step) remaining time: 0:40:15
133
+ Epoch 1 | iter 992 step 31 | loss train: 1.438, val: 1.332 | iter time: 359.77 ms (step) remaining time: 0:40:04
134
+ Epoch 1 | iter 1024 step 32 | loss train: 1.413, val: 1.332 | iter time: 358.74 ms (step) remaining time: 0:39:54
135
+ Epoch 1 | iter 1056 step 33 | loss train: 1.313, val: 1.332 | iter time: 358.83 ms (step) remaining time: 0:39:43
136
+ Epoch 1 | iter 1088 step 34 | loss train: 1.467, val: 1.332 | iter time: 359.99 ms (step) remaining time: 0:39:32
137
+ Epoch 1 | iter 1120 step 35 | loss train: 1.367, val: 1.332 | iter time: 360.26 ms (step) remaining time: 0:39:21
138
+ Epoch 1 | iter 1152 step 36 | loss train: 1.371, val: 1.332 | iter time: 358.26 ms (step) remaining time: 0:39:10
139
+ Epoch 1 | iter 1184 step 37 | loss train: 1.384, val: 1.332 | iter time: 360.89 ms (step) remaining time: 0:38:59
140
+ Epoch 1 | iter 1216 step 38 | loss train: 1.473, val: 1.332 | iter time: 360.28 ms (step) remaining time: 0:38:48
141
+ Epoch 1 | iter 1248 step 39 | loss train: 1.339, val: 1.332 | iter time: 567.37 ms (step) remaining time: 0:38:38
142
+ Epoch 1 | iter 1280 step 40 | loss train: 1.378, val: 1.332 | iter time: 359.83 ms (step) remaining time: 0:38:27
143
+ Epoch 1 | iter 1312 step 41 | loss train: 1.402, val: 1.332 | iter time: 359.55 ms (step) remaining time: 0:38:16
144
+ Epoch 1 | iter 1344 step 42 | loss train: 1.378, val: 1.332 | iter time: 359.78 ms (step) remaining time: 0:38:05
145
+ Epoch 1 | iter 1376 step 43 | loss train: 1.412, val: 1.332 | iter time: 358.24 ms (step) remaining time: 0:37:54
146
+ Epoch 1 | iter 1408 step 44 | loss train: 1.351, val: 1.332 | iter time: 359.95 ms (step) remaining time: 0:37:43
147
+ Epoch 1 | iter 1440 step 45 | loss train: 1.361, val: 1.332 | iter time: 359.66 ms (step) remaining time: 0:37:32
148
+ Epoch 1 | iter 1472 step 46 | loss train: 1.381, val: 1.332 | iter time: 360.26 ms (step) remaining time: 0:37:21
149
+ Epoch 1 | iter 1504 step 47 | loss train: 1.390, val: 1.332 | iter time: 359.66 ms (step) remaining time: 0:37:10
150
+ Epoch 1 | iter 1536 step 48 | loss train: 1.385, val: 1.332 | iter time: 359.21 ms (step) remaining time: 0:36:59
151
+ Epoch 1 | iter 1568 step 49 | loss train: 1.336, val: 1.332 | iter time: 359.84 ms (step) remaining time: 0:36:48
152
+ Epoch 1 | iter 1600 step 50 | loss train: 1.439, val: 1.332 | iter time: 360.70 ms (step) remaining time: 0:36:37
153
+ Validating ...
154
+ iter 1600: val loss 1.3841, val time: 9333.33 ms
155
+ Epoch 1 | iter 1632 step 51 | loss train: 1.399, val: 1.384 | iter time: 359.46 ms (step) remaining time: 0:37:03
156
+ Epoch 1 | iter 1664 step 52 | loss train: 1.382, val: 1.384 | iter time: 358.16 ms (step) remaining time: 0:36:51
157
+ Epoch 1 | iter 1696 step 53 | loss train: 1.403, val: 1.384 | iter time: 358.36 ms (step) remaining time: 0:36:39
158
+ Epoch 1 | iter 1728 step 54 | loss train: 1.356, val: 1.384 | iter time: 361.38 ms (step) remaining time: 0:36:27
159
+ Epoch 1 | iter 1760 step 55 | loss train: 1.366, val: 1.384 | iter time: 359.32 ms (step) remaining time: 0:36:15
160
+ Epoch 1 | iter 1792 step 56 | loss train: 1.372, val: 1.384 | iter time: 359.30 ms (step) remaining time: 0:36:04
161
+ Epoch 1 | iter 1824 step 57 | loss train: 1.346, val: 1.384 | iter time: 359.75 ms (step) remaining time: 0:35:52
162
+ Epoch 1 | iter 1856 step 58 | loss train: 1.389, val: 1.384 | iter time: 358.76 ms (step) remaining time: 0:35:40
163
+ Epoch 1 | iter 1888 step 59 | loss train: 1.383, val: 1.384 | iter time: 358.26 ms (step) remaining time: 0:35:29
164
+ Epoch 1 | iter 1920 step 60 | loss train: 1.376, val: 1.384 | iter time: 359.70 ms (step) remaining time: 0:35:17
165
+ Epoch 1 | iter 1952 step 61 | loss train: 1.416, val: 1.384 | iter time: 360.33 ms (step) remaining time: 0:35:06
166
+ Epoch 1 | iter 1984 step 62 | loss train: 1.413, val: 1.384 | iter time: 359.84 ms (step) remaining time: 0:34:54
167
+ Epoch 1 | iter 2016 step 63 | loss train: 1.403, val: 1.384 | iter time: 359.84 ms (step) remaining time: 0:34:43
168
+ Epoch 1 | iter 2048 step 64 | loss train: 1.422, val: 1.384 | iter time: 359.36 ms (step) remaining time: 0:34:31
169
+ Epoch 1 | iter 2080 step 65 | loss train: 1.403, val: 1.384 | iter time: 359.96 ms (step) remaining time: 0:34:20
170
+ Epoch 1 | iter 2112 step 66 | loss train: 1.423, val: 1.384 | iter time: 359.89 ms (step) remaining time: 0:34:08
171
+ Epoch 1 | iter 2144 step 67 | loss train: 1.388, val: 1.384 | iter time: 357.93 ms (step) remaining time: 0:33:57
172
+ Epoch 1 | iter 2176 step 68 | loss train: 1.387, val: 1.384 | iter time: 360.34 ms (step) remaining time: 0:33:45
173
+ Epoch 1 | iter 2208 step 69 | loss train: 1.378, val: 1.384 | iter time: 360.62 ms (step) remaining time: 0:33:34
174
+ Epoch 1 | iter 2240 step 70 | loss train: 1.456, val: 1.384 | iter time: 360.33 ms (step) remaining time: 0:33:22
175
+ Epoch 1 | iter 2272 step 71 | loss train: 1.331, val: 1.384 | iter time: 358.28 ms (step) remaining time: 0:33:11
176
+ Epoch 1 | iter 2304 step 72 | loss train: 1.450, val: 1.384 | iter time: 361.11 ms (step) remaining time: 0:33:00
177
+ Epoch 1 | iter 2336 step 73 | loss train: 1.370, val: 1.384 | iter time: 358.94 ms (step) remaining time: 0:32:48
178
+ Epoch 1 | iter 2368 step 74 | loss train: 1.393, val: 1.384 | iter time: 360.33 ms (step) remaining time: 0:32:37
179
+ Epoch 1 | iter 2400 step 75 | loss train: 1.346, val: 1.384 | iter time: 360.23 ms (step) remaining time: 0:32:26
180
+ Epoch 1 | iter 2432 step 76 | loss train: 1.390, val: 1.384 | iter time: 359.55 ms (step) remaining time: 0:32:14
181
+ Epoch 1 | iter 2464 step 77 | loss train: 1.430, val: 1.384 | iter time: 360.50 ms (step) remaining time: 0:32:03
182
+ Epoch 1 | iter 2496 step 78 | loss train: 1.400, val: 1.384 | iter time: 359.88 ms (step) remaining time: 0:31:52
183
+ Epoch 1 | iter 2528 step 79 | loss train: 1.414, val: 1.384 | iter time: 358.00 ms (step) remaining time: 0:31:40
184
+ Epoch 1 | iter 2560 step 80 | loss train: 1.360, val: 1.384 | iter time: 359.93 ms (step) remaining time: 0:31:29
185
+ Epoch 1 | iter 2592 step 81 | loss train: 1.370, val: 1.384 | iter time: 360.49 ms (step) remaining time: 0:31:18
186
+ Epoch 1 | iter 2624 step 82 | loss train: 1.357, val: 1.384 | iter time: 359.63 ms (step) remaining time: 0:31:07
187
+ Epoch 1 | iter 2656 step 83 | loss train: 1.475, val: 1.384 | iter time: 358.50 ms (step) remaining time: 0:30:56
188
+ Epoch 1 | iter 2688 step 84 | loss train: 1.418, val: 1.384 | iter time: 359.98 ms (step) remaining time: 0:30:44
189
+ Epoch 1 | iter 2720 step 85 | loss train: 1.399, val: 1.384 | iter time: 358.52 ms (step) remaining time: 0:30:33
190
+ Epoch 1 | iter 2752 step 86 | loss train: 1.314, val: 1.384 | iter time: 359.80 ms (step) remaining time: 0:30:22
191
+ Epoch 1 | iter 2784 step 87 | loss train: 1.379, val: 1.384 | iter time: 359.09 ms (step) remaining time: 0:30:11
192
+ Epoch 1 | iter 2816 step 88 | loss train: 1.389, val: 1.384 | iter time: 360.58 ms (step) remaining time: 0:30:00
193
+ Epoch 1 | iter 2848 step 89 | loss train: 1.462, val: 1.384 | iter time: 359.08 ms (step) remaining time: 0:29:48
194
+ Epoch 1 | iter 2880 step 90 | loss train: 1.416, val: 1.384 | iter time: 358.58 ms (step) remaining time: 0:29:37
195
+ Epoch 1 | iter 2912 step 91 | loss train: 1.358, val: 1.384 | iter time: 360.25 ms (step) remaining time: 0:29:27
196
+ Epoch 1 | iter 2944 step 92 | loss train: 1.353, val: 1.384 | iter time: 359.92 ms (step) remaining time: 0:29:15
197
+ Epoch 1 | iter 2976 step 93 | loss train: 1.371, val: 1.384 | iter time: 358.83 ms (step) remaining time: 0:29:04
198
+ Epoch 1 | iter 3008 step 94 | loss train: 1.398, val: 1.384 | iter time: 358.68 ms (step) remaining time: 0:28:53
199
+ Epoch 1 | iter 3040 step 95 | loss train: 1.385, val: 1.384 | iter time: 358.23 ms (step) remaining time: 0:28:42
200
+ Epoch 1 | iter 3072 step 96 | loss train: 1.368, val: 1.384 | iter time: 359.37 ms (step) remaining time: 0:28:31
201
+ Epoch 1 | iter 3104 step 97 | loss train: 1.405, val: 1.384 | iter time: 362.03 ms (step) remaining time: 0:28:20
202
+ Epoch 1 | iter 3136 step 98 | loss train: 1.418, val: 1.384 | iter time: 358.84 ms (step) remaining time: 0:28:08
203
+ Epoch 1 | iter 3168 step 99 | loss train: 1.383, val: 1.384 | iter time: 361.50 ms (step) remaining time: 0:27:57
204
+ Epoch 1 | iter 3200 step 100 | loss train: 1.381, val: 1.384 | iter time: 357.64 ms (step) remaining time: 0:27:46
205
+ Validating ...
206
+ iter 3200: val loss 1.3758, val time: 9316.28 ms
207
+ Saving checkpoint to 'out/pretrain/2501/step-00000100/lit_model.pth'
208
+ Epoch 1 | iter 3232 step 101 | loss train: 1.369, val: 1.376 | iter time: 356.94 ms (step) remaining time: 0:28:14
209
+ Epoch 1 | iter 3264 step 102 | loss train: 1.419, val: 1.376 | iter time: 359.22 ms (step) remaining time: 0:28:03
210
+ Epoch 1 | iter 3296 step 103 | loss train: 1.440, val: 1.376 | iter time: 360.22 ms (step) remaining time: 0:27:51
211
+ Epoch 1 | iter 3328 step 104 | loss train: 1.407, val: 1.376 | iter time: 358.35 ms (step) remaining time: 0:27:39
212
+ Epoch 1 | iter 3360 step 105 | loss train: 1.383, val: 1.376 | iter time: 359.82 ms (step) remaining time: 0:27:27
213
+ Epoch 1 | iter 3392 step 106 | loss train: 1.396, val: 1.376 | iter time: 359.48 ms (step) remaining time: 0:27:16
214
+ Epoch 1 | iter 3424 step 107 | loss train: 1.320, val: 1.376 | iter time: 358.96 ms (step) remaining time: 0:27:04
215
+ Epoch 1 | iter 3456 step 108 | loss train: 1.330, val: 1.376 | iter time: 360.58 ms (step) remaining time: 0:26:52
216
+ Epoch 1 | iter 3488 step 109 | loss train: 1.373, val: 1.376 | iter time: 357.68 ms (step) remaining time: 0:26:41
217
+ Epoch 1 | iter 3520 step 110 | loss train: 1.408, val: 1.376 | iter time: 360.18 ms (step) remaining time: 0:26:29
218
+ Epoch 1 | iter 3552 step 111 | loss train: 1.320, val: 1.376 | iter time: 360.34 ms (step) remaining time: 0:26:17
219
+ Epoch 1 | iter 3584 step 112 | loss train: 1.342, val: 1.376 | iter time: 360.17 ms (step) remaining time: 0:26:06
220
+ Epoch 1 | iter 3616 step 113 | loss train: 1.356, val: 1.376 | iter time: 360.13 ms (step) remaining time: 0:25:54
221
+ Epoch 1 | iter 3648 step 114 | loss train: 1.406, val: 1.376 | iter time: 358.94 ms (step) remaining time: 0:25:42
222
+ Epoch 1 | iter 3680 step 115 | loss train: 1.394, val: 1.376 | iter time: 356.49 ms (step) remaining time: 0:25:31
223
+ Epoch 1 | iter 3712 step 116 | loss train: 1.352, val: 1.376 | iter time: 357.91 ms (step) remaining time: 0:25:19
224
+ Epoch 1 | iter 3744 step 117 | loss train: 1.402, val: 1.376 | iter time: 358.79 ms (step) remaining time: 0:25:08
225
+ Epoch 1 | iter 3776 step 118 | loss train: 1.402, val: 1.376 | iter time: 359.61 ms (step) remaining time: 0:24:56
226
+ Epoch 1 | iter 3808 step 119 | loss train: 1.425, val: 1.376 | iter time: 357.89 ms (step) remaining time: 0:24:45
227
+ Epoch 1 | iter 3840 step 120 | loss train: 1.401, val: 1.376 | iter time: 359.65 ms (step) remaining time: 0:24:33
228
+ Epoch 1 | iter 3872 step 121 | loss train: 1.384, val: 1.376 | iter time: 358.79 ms (step) remaining time: 0:24:22
229
+ Epoch 1 | iter 3904 step 122 | loss train: 1.360, val: 1.376 | iter time: 360.40 ms (step) remaining time: 0:24:10
230
+ Epoch 1 | iter 3936 step 123 | loss train: 1.456, val: 1.376 | iter time: 360.81 ms (step) remaining time: 0:23:59
231
+ Epoch 1 | iter 3968 step 124 | loss train: 1.363, val: 1.376 | iter time: 357.82 ms (step) remaining time: 0:23:47
232
+ Epoch 1 | iter 4000 step 125 | loss train: 1.366, val: 1.376 | iter time: 359.29 ms (step) remaining time: 0:23:36
233
+ Epoch 1 | iter 4032 step 126 | loss train: 1.341, val: 1.376 | iter time: 359.66 ms (step) remaining time: 0:23:24
234
+ Epoch 1 | iter 4064 step 127 | loss train: 1.351, val: 1.376 | iter time: 357.20 ms (step) remaining time: 0:23:13
235
+ Epoch 1 | iter 4096 step 128 | loss train: 1.309, val: 1.376 | iter time: 359.26 ms (step) remaining time: 0:23:02
236
+ Epoch 1 | iter 4128 step 129 | loss train: 1.369, val: 1.376 | iter time: 359.71 ms (step) remaining time: 0:22:50
237
+ Epoch 1 | iter 4160 step 130 | loss train: 1.372, val: 1.376 | iter time: 359.24 ms (step) remaining time: 0:22:39
238
+ Epoch 1 | iter 4192 step 131 | loss train: 1.379, val: 1.376 | iter time: 361.33 ms (step) remaining time: 0:22:27
239
+ Epoch 1 | iter 4224 step 132 | loss train: 1.371, val: 1.376 | iter time: 358.67 ms (step) remaining time: 0:22:16
240
+ Epoch 1 | iter 4256 step 133 | loss train: 1.293, val: 1.376 | iter time: 358.61 ms (step) remaining time: 0:22:05
241
+ Epoch 1 | iter 4288 step 134 | loss train: 1.356, val: 1.376 | iter time: 359.37 ms (step) remaining time: 0:21:53
242
+ Epoch 1 | iter 4320 step 135 | loss train: 1.374, val: 1.376 | iter time: 360.14 ms (step) remaining time: 0:21:42
243
+ Epoch 1 | iter 4352 step 136 | loss train: 1.379, val: 1.376 | iter time: 358.83 ms (step) remaining time: 0:21:31
244
+ Epoch 1 | iter 4384 step 137 | loss train: 1.325, val: 1.376 | iter time: 359.97 ms (step) remaining time: 0:21:19
245
+ Epoch 1 | iter 4416 step 138 | loss train: 1.310, val: 1.376 | iter time: 360.51 ms (step) remaining time: 0:21:08
246
+ Epoch 1 | iter 4448 step 139 | loss train: 1.344, val: 1.376 | iter time: 357.83 ms (step) remaining time: 0:20:56
247
+ Epoch 1 | iter 4480 step 140 | loss train: 1.404, val: 1.376 | iter time: 358.43 ms (step) remaining time: 0:20:45
248
+ Epoch 1 | iter 4512 step 141 | loss train: 1.398, val: 1.376 | iter time: 359.62 ms (step) remaining time: 0:20:34
249
+ Epoch 1 | iter 4544 step 142 | loss train: 1.305, val: 1.376 | iter time: 360.75 ms (step) remaining time: 0:20:23
250
+ Epoch 1 | iter 4576 step 143 | loss train: 1.334, val: 1.376 | iter time: 358.80 ms (step) remaining time: 0:20:12
251
+ Epoch 1 | iter 4608 step 144 | loss train: 1.267, val: 1.376 | iter time: 359.82 ms (step) remaining time: 0:20:00
252
+ Epoch 1 | iter 4640 step 145 | loss train: 1.310, val: 1.376 | iter time: 359.23 ms (step) remaining time: 0:19:49
253
+ Epoch 1 | iter 4672 step 146 | loss train: 1.347, val: 1.376 | iter time: 360.55 ms (step) remaining time: 0:19:38
254
+ Epoch 1 | iter 4704 step 147 | loss train: 1.339, val: 1.376 | iter time: 360.26 ms (step) remaining time: 0:19:26
255
+ Epoch 1 | iter 4736 step 148 | loss train: 1.320, val: 1.376 | iter time: 359.16 ms (step) remaining time: 0:19:15
256
+ Epoch 1 | iter 4768 step 149 | loss train: 1.307, val: 1.376 | iter time: 358.87 ms (step) remaining time: 0:19:04
257
+ Epoch 1 | iter 4800 step 150 | loss train: 1.353, val: 1.376 | iter time: 359.04 ms (step) remaining time: 0:18:52
258
+ Validating ...
259
+ iter 4800: val loss 1.3584, val time: 9331.75 ms
260
+ Epoch 1 | iter 4832 step 151 | loss train: 1.368, val: 1.358 | iter time: 360.64 ms (step) remaining time: 0:18:47
261
+ Epoch 1 | iter 4864 step 152 | loss train: 1.279, val: 1.358 | iter time: 360.04 ms (step) remaining time: 0:18:36
262
+ Epoch 1 | iter 4896 step 153 | loss train: 1.366, val: 1.358 | iter time: 360.63 ms (step) remaining time: 0:18:25
263
+ Epoch 1 | iter 4928 step 154 | loss train: 1.266, val: 1.358 | iter time: 360.45 ms (step) remaining time: 0:18:13
264
+ Epoch 1 | iter 4960 step 155 | loss train: 1.348, val: 1.358 | iter time: 359.06 ms (step) remaining time: 0:18:02
265
+ Epoch 1 | iter 4992 step 156 | loss train: 1.389, val: 1.358 | iter time: 358.50 ms (step) remaining time: 0:17:51
266
+ Epoch 1 | iter 5024 step 157 | loss train: 1.293, val: 1.358 | iter time: 358.17 ms (step) remaining time: 0:17:39
267
+ Epoch 1 | iter 5056 step 158 | loss train: 1.321, val: 1.358 | iter time: 358.70 ms (step) remaining time: 0:17:28
268
+ Epoch 1 | iter 5088 step 159 | loss train: 1.378, val: 1.358 | iter time: 357.95 ms (step) remaining time: 0:17:17
269
+ Epoch 1 | iter 5120 step 160 | loss train: 1.392, val: 1.358 | iter time: 360.34 ms (step) remaining time: 0:17:05
270
+ Epoch 1 | iter 5152 step 161 | loss train: 1.413, val: 1.358 | iter time: 359.66 ms (step) remaining time: 0:16:54
271
+ Epoch 1 | iter 5184 step 162 | loss train: 1.470, val: 1.358 | iter time: 359.69 ms (step) remaining time: 0:16:43
272
+ Epoch 1 | iter 5216 step 163 | loss train: 1.372, val: 1.358 | iter time: 358.44 ms (step) remaining time: 0:16:31
273
+ Epoch 1 | iter 5248 step 164 | loss train: 1.303, val: 1.358 | iter time: 361.69 ms (step) remaining time: 0:16:20
274
+ Epoch 1 | iter 5280 step 165 | loss train: 1.376, val: 1.358 | iter time: 359.07 ms (step) remaining time: 0:16:09
275
+ Epoch 1 | iter 5312 step 166 | loss train: 1.343, val: 1.358 | iter time: 361.26 ms (step) remaining time: 0:15:57
276
+ Epoch 1 | iter 5344 step 167 | loss train: 1.392, val: 1.358 | iter time: 359.30 ms (step) remaining time: 0:15:46
277
+ Epoch 1 | iter 5376 step 168 | loss train: 1.375, val: 1.358 | iter time: 359.14 ms (step) remaining time: 0:15:35
278
+ Epoch 1 | iter 5408 step 169 | loss train: 1.358, val: 1.358 | iter time: 359.95 ms (step) remaining time: 0:15:24
279
+ Epoch 1 | iter 5440 step 170 | loss train: 1.387, val: 1.358 | iter time: 360.69 ms (step) remaining time: 0:15:12
280
+ Epoch 1 | iter 5472 step 171 | loss train: 1.347, val: 1.358 | iter time: 357.40 ms (step) remaining time: 0:15:01
281
+ Epoch 1 | iter 5504 step 172 | loss train: 1.354, val: 1.358 | iter time: 360.32 ms (step) remaining time: 0:14:50
282
+ Epoch 1 | iter 5536 step 173 | loss train: 1.366, val: 1.358 | iter time: 358.60 ms (step) remaining time: 0:14:39
283
+ Epoch 1 | iter 5568 step 174 | loss train: 1.337, val: 1.358 | iter time: 359.23 ms (step) remaining time: 0:14:27
284
+ Epoch 1 | iter 5600 step 175 | loss train: 1.337, val: 1.358 | iter time: 359.73 ms (step) remaining time: 0:14:16
285
+ Epoch 1 | iter 5632 step 176 | loss train: 1.367, val: 1.358 | iter time: 358.54 ms (step) remaining time: 0:14:05
286
+ Epoch 1 | iter 5664 step 177 | loss train: 1.427, val: 1.358 | iter time: 359.61 ms (step) remaining time: 0:13:54
287
+ Epoch 1 | iter 5696 step 178 | loss train: 1.292, val: 1.358 | iter time: 357.83 ms (step) remaining time: 0:13:42
288
+ Epoch 1 | iter 5728 step 179 | loss train: 1.315, val: 1.358 | iter time: 358.14 ms (step) remaining time: 0:13:31
289
+ Epoch 1 | iter 5760 step 180 | loss train: 1.341, val: 1.358 | iter time: 362.36 ms (step) remaining time: 0:13:20
290
+ Epoch 1 | iter 5792 step 181 | loss train: 1.344, val: 1.358 | iter time: 359.84 ms (step) remaining time: 0:13:09
291
+ Epoch 1 | iter 5824 step 182 | loss train: 1.347, val: 1.358 | iter time: 359.39 ms (step) remaining time: 0:12:58
292
+ Epoch 1 | iter 5856 step 183 | loss train: 1.342, val: 1.358 | iter time: 358.73 ms (step) remaining time: 0:12:46
293
+ Epoch 1 | iter 5888 step 184 | loss train: 1.327, val: 1.358 | iter time: 359.10 ms (step) remaining time: 0:12:35
294
+ Epoch 1 | iter 5920 step 185 | loss train: 1.291, val: 1.358 | iter time: 360.39 ms (step) remaining time: 0:12:24
295
+ Epoch 1 | iter 5952 step 186 | loss train: 1.345, val: 1.358 | iter time: 359.84 ms (step) remaining time: 0:12:13
296
+ Epoch 1 | iter 5984 step 187 | loss train: 1.355, val: 1.358 | iter time: 359.66 ms (step) remaining time: 0:12:02
297
+ Epoch 1 | iter 6016 step 188 | loss train: 1.359, val: 1.358 | iter time: 358.92 ms (step) remaining time: 0:11:50
298
+ Epoch 1 | iter 6048 step 189 | loss train: 1.333, val: 1.358 | iter time: 360.08 ms (step) remaining time: 0:11:39
299
+ Epoch 1 | iter 6080 step 190 | loss train: 1.322, val: 1.358 | iter time: 358.55 ms (step) remaining time: 0:11:28
300
+ Epoch 1 | iter 6112 step 191 | loss train: 1.350, val: 1.358 | iter time: 359.75 ms (step) remaining time: 0:11:17
301
+ Epoch 1 | iter 6144 step 192 | loss train: 1.317, val: 1.358 | iter time: 358.34 ms (step) remaining time: 0:11:06
302
+ Epoch 1 | iter 6176 step 193 | loss train: 1.433, val: 1.358 | iter time: 359.86 ms (step) remaining time: 0:10:54
303
+ Epoch 1 | iter 6208 step 194 | loss train: 1.398, val: 1.358 | iter time: 359.30 ms (step) remaining time: 0:10:43
304
+ Epoch 1 | iter 6240 step 195 | loss train: 1.394, val: 1.358 | iter time: 360.39 ms (step) remaining time: 0:10:32
305
+ Epoch 1 | iter 6272 step 196 | loss train: 1.342, val: 1.358 | iter time: 357.85 ms (step) remaining time: 0:10:21
306
+ Epoch 1 | iter 6304 step 197 | loss train: 1.339, val: 1.358 | iter time: 358.97 ms (step) remaining time: 0:10:10
307
+ Epoch 1 | iter 6336 step 198 | loss train: 1.325, val: 1.358 | iter time: 357.72 ms (step) remaining time: 0:09:59
308
+ Epoch 1 | iter 6368 step 199 | loss train: 1.259, val: 1.358 | iter time: 360.42 ms (step) remaining time: 0:09:48
309
+ Epoch 1 | iter 6400 step 200 | loss train: 1.374, val: 1.358 | iter time: 359.52 ms (step) remaining time: 0:09:36
310
+ Validating ...
311
+ iter 6400: val loss 1.3422, val time: 9310.92 ms
312
+ Saving checkpoint to 'out/pretrain/2501/step-00000200/lit_model.pth'
313
+ Epoch 1 | iter 6432 step 201 | loss train: 1.287, val: 1.342 | iter time: 357.16 ms (step) remaining time: 0:09:32
314
+ Epoch 1 | iter 6464 step 202 | loss train: 1.426, val: 1.342 | iter time: 357.81 ms (step) remaining time: 0:09:21
315
+ Epoch 1 | iter 6496 step 203 | loss train: 1.360, val: 1.342 | iter time: 356.41 ms (step) remaining time: 0:09:09
316
+ Epoch 1 | iter 6528 step 204 | loss train: 1.257, val: 1.342 | iter time: 360.00 ms (step) remaining time: 0:08:58
317
+ Epoch 1 | iter 6560 step 205 | loss train: 1.297, val: 1.342 | iter time: 358.62 ms (step) remaining time: 0:08:47
318
+ Epoch 1 | iter 6592 step 206 | loss train: 1.310, val: 1.342 | iter time: 359.17 ms (step) remaining time: 0:08:35
319
+ Epoch 1 | iter 6624 step 207 | loss train: 1.291, val: 1.342 | iter time: 359.83 ms (step) remaining time: 0:08:24
320
+ Epoch 1 | iter 6656 step 208 | loss train: 1.322, val: 1.342 | iter time: 359.76 ms (step) remaining time: 0:08:13
321
+ Epoch 1 | iter 6688 step 209 | loss train: 1.394, val: 1.342 | iter time: 360.63 ms (step) remaining time: 0:08:01
322
+ Epoch 1 | iter 6720 step 210 | loss train: 1.287, val: 1.342 | iter time: 359.49 ms (step) remaining time: 0:07:50
323
+ Epoch 1 | iter 6752 step 211 | loss train: 1.331, val: 1.342 | iter time: 358.42 ms (step) remaining time: 0:07:39
324
+ Epoch 1 | iter 6784 step 212 | loss train: 1.279, val: 1.342 | iter time: 358.61 ms (step) remaining time: 0:07:28
325
+ Epoch 1 | iter 6816 step 213 | loss train: 1.420, val: 1.342 | iter time: 360.61 ms (step) remaining time: 0:07:16
326
+ Epoch 1 | iter 6848 step 214 | loss train: 1.332, val: 1.342 | iter time: 359.73 ms (step) remaining time: 0:07:05
327
+ Epoch 1 | iter 6880 step 215 | loss train: 1.279, val: 1.342 | iter time: 359.59 ms (step) remaining time: 0:06:54
328
+ Epoch 1 | iter 6912 step 216 | loss train: 1.283, val: 1.342 | iter time: 360.42 ms (step) remaining time: 0:06:43
329
+ Epoch 1 | iter 6944 step 217 | loss train: 1.322, val: 1.342 | iter time: 361.37 ms (step) remaining time: 0:06:31
330
+ Epoch 1 | iter 6976 step 218 | loss train: 1.278, val: 1.342 | iter time: 359.24 ms (step) remaining time: 0:06:20
331
+ Epoch 1 | iter 7008 step 219 | loss train: 1.306, val: 1.342 | iter time: 359.08 ms (step) remaining time: 0:06:09
332
+ Epoch 1 | iter 7040 step 220 | loss train: 1.367, val: 1.342 | iter time: 359.51 ms (step) remaining time: 0:05:58
333
+ Epoch 1 | iter 7072 step 221 | loss train: 1.401, val: 1.342 | iter time: 358.77 ms (step) remaining time: 0:05:46
334
+ Epoch 1 | iter 7104 step 222 | loss train: 1.307, val: 1.342 | iter time: 359.80 ms (step) remaining time: 0:05:35
335
+ Epoch 1 | iter 7136 step 223 | loss train: 1.292, val: 1.342 | iter time: 357.55 ms (step) remaining time: 0:05:24
336
+ Epoch 1 | iter 7168 step 224 | loss train: 1.316, val: 1.342 | iter time: 357.36 ms (step) remaining time: 0:05:13
337
+ Epoch 1 | iter 7200 step 225 | loss train: 1.304, val: 1.342 | iter time: 360.23 ms (step) remaining time: 0:05:01
338
+ Epoch 1 | iter 7232 step 226 | loss train: 1.264, val: 1.342 | iter time: 359.23 ms (step) remaining time: 0:04:50
339
+ Epoch 1 | iter 7264 step 227 | loss train: 1.297, val: 1.342 | iter time: 359.82 ms (step) remaining time: 0:04:39
340
+ Epoch 1 | iter 7296 step 228 | loss train: 1.325, val: 1.342 | iter time: 360.34 ms (step) remaining time: 0:04:28
341
+ Epoch 1 | iter 7328 step 229 | loss train: 1.234, val: 1.342 | iter time: 360.32 ms (step) remaining time: 0:04:17
342
+ Epoch 1 | iter 7360 step 230 | loss train: 1.287, val: 1.342 | iter time: 361.14 ms (step) remaining time: 0:04:05
343
+ Epoch 1 | iter 7392 step 231 | loss train: 1.249, val: 1.342 | iter time: 361.26 ms (step) remaining time: 0:03:54
344
+ Epoch 1 | iter 7424 step 232 | loss train: 1.299, val: 1.342 | iter time: 359.71 ms (step) remaining time: 0:03:43
345
+ Epoch 1 | iter 7456 step 233 | loss train: 1.321, val: 1.342 | iter time: 359.22 ms (step) remaining time: 0:03:32
346
+ Epoch 1 | iter 7488 step 234 | loss train: 1.303, val: 1.342 | iter time: 359.32 ms (step) remaining time: 0:03:21
347
+ Epoch 1 | iter 7520 step 235 | loss train: 1.363, val: 1.342 | iter time: 360.85 ms (step) remaining time: 0:03:09
348
+ Epoch 1 | iter 7552 step 236 | loss train: 1.248, val: 1.342 | iter time: 356.77 ms (step) remaining time: 0:02:58
349
+ Epoch 1 | iter 7584 step 237 | loss train: 1.220, val: 1.342 | iter time: 360.02 ms (step) remaining time: 0:02:47
350
+ Epoch 1 | iter 7616 step 238 | loss train: 1.325, val: 1.342 | iter time: 357.48 ms (step) remaining time: 0:02:36
351
+ Epoch 1 | iter 7648 step 239 | loss train: 1.352, val: 1.342 | iter time: 358.89 ms (step) remaining time: 0:02:25
352
+ Epoch 1 | iter 7680 step 240 | loss train: 1.250, val: 1.342 | iter time: 360.98 ms (step) remaining time: 0:02:13
353
+ Epoch 1 | iter 7712 step 241 | loss train: 1.265, val: 1.342 | iter time: 358.69 ms (step) remaining time: 0:02:02
354
+ Epoch 1 | iter 7744 step 242 | loss train: 1.269, val: 1.342 | iter time: 359.83 ms (step) remaining time: 0:01:51
355
+ Epoch 1 | iter 7776 step 243 | loss train: 1.286, val: 1.342 | iter time: 360.23 ms (step) remaining time: 0:01:40
356
+ Epoch 1 | iter 7808 step 244 | loss train: 1.292, val: 1.342 | iter time: 359.48 ms (step) remaining time: 0:01:29
357
+ Epoch 1 | iter 7840 step 245 | loss train: 1.219, val: 1.342 | iter time: 359.16 ms (step) remaining time: 0:01:18
358
+ Epoch 1 | iter 7872 step 246 | loss train: 1.377, val: 1.342 | iter time: 359.59 ms (step) remaining time: 0:01:06
359
+ Epoch 1 | iter 7904 step 247 | loss train: 1.245, val: 1.342 | iter time: 360.63 ms (step) remaining time: 0:00:55
360
+ Epoch 1 | iter 7936 step 248 | loss train: 1.228, val: 1.342 | iter time: 360.64 ms (step) remaining time: 0:00:44
361
+ Epoch 1 | iter 7968 step 249 | loss train: 1.286, val: 1.342 | iter time: 357.79 ms (step) remaining time: 0:00:33
362
+ Epoch 1 | iter 8000 step 250 | loss train: 1.271, val: 1.342 | iter time: 359.67 ms (step) remaining time: 0:00:22
363
+ Validating ...
364
+ iter 8000: val loss 1.3335, val time: 9326.54 ms
365
+ Epoch 1 | iter 8032 step 251 | loss train: 1.335, val: 1.333 | iter time: 361.00 ms (step) remaining time: 0:00:11
366
+ Epoch 1 | iter 8064 step 252 | loss train: 1.351, val: 1.333 | iter time: 358.17 ms (step) remaining time: 0:00:00
367
+ Validating ...
368
+ Final evaluation | val loss: 1.333 | val ppl: 3.792
369
+ Saving checkpoint to 'out/pretrain/2501/final/lit_model.pth'
370
+ ----------------------------------------
371
+ | Performance
372
+ | - Total tokens : 264,241,152
373
+ | - Training Time : 2857.04 s
374
+ | - Tok/sec : 234.93 tok/s
375
+ | ----------------------------------------
376
+ | Memory Usage
377
+ | - Memory Used : 26.32 GB
378
+ ----------------------------------------
out/pretrain/tinyllama/teelogs/2501_full.txt ADDED
@@ -0,0 +1,389 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Initializing distributed: GLOBAL_RANK: 0, MEMBER: 1/4
2
+ Initializing distributed: GLOBAL_RANK: 1, MEMBER: 2/4
3
+ [rank: 1] Seed set to 42
4
+ Initializing distributed: GLOBAL_RANK: 3, MEMBER: 4/4
5
+ Initializing distributed: GLOBAL_RANK: 2, MEMBER: 3/4
6
+ [rank: 3] Seed set to 42
7
+ [rank: 2] Seed set to 42
8
+ ----------------------------------------------------------------------------------------------------
9
+ distributed_backend=nccl
10
+ All distributed processes registered. Starting with 4 processes
11
+ ----------------------------------------------------------------------------------------------------
12
+
13
+ [rank: 0] Seed set to 42
14
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
15
+ train_dataloader = data.train_dataloader()
16
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
17
+ train_dataloader = data.train_dataloader()
18
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
19
+ train_dataloader = data.train_dataloader()
20
+ /mnt/data/litgpt/litgpt/pretrain.py:515: UserWarning: A newer version of litdata is available (0.2.52). Please consider upgrading with `pip install -U litdata`. Not all functionalities of the platform can be guaranteed to work with the current version.
21
+ train_dataloader = data.train_dataloader()
22
+ All GPUs are fully connected via NVLink.
23
+ {'data': {'batch_size': 1,
24
+ 'data_path': PosixPath('litgpt/data/arxiv'),
25
+ 'num_workers': 8,
26
+ 'seed': 42,
27
+ 'seq_length': 2048,
28
+ 'use_starcoder': True},
29
+ 'data_dir': PosixPath('litgpt/data/arxiv/2501'),
30
+ 'devices': 'auto',
31
+ 'eval': {'evaluate_example': 'first',
32
+ 'final_validation': True,
33
+ 'initial_validation': True,
34
+ 'interval': 50,
35
+ 'max_iters': 200,
36
+ 'max_new_tokens': None},
37
+ 'initial_checkpoint_dir': PosixPath('out/pretrain/2412_full/final'),
38
+ 'log': {'group': None, 'project': None, 'run': None},
39
+ 'logger_name': 'tensorboard',
40
+ 'model_config': {'attention_logit_softcapping': None,
41
+ 'attention_scores_scalar': None,
42
+ 'attn_bias': False,
43
+ 'bias': False,
44
+ 'block_size': 2048,
45
+ 'final_logit_softcapping': None,
46
+ 'gelu_approximate': 'none',
47
+ 'head_size': 64,
48
+ 'hf_config': {'name': 'TinyLlama-1.1B-intermediate-step-1431k-3T',
49
+ 'org': 'TinyLlama'},
50
+ 'intermediate_size': 5632,
51
+ 'lm_head_bias': False,
52
+ 'mlp_class_name': 'LLaMAMLP',
53
+ 'moe_intermediate_size': None,
54
+ 'n_embd': 2048,
55
+ 'n_expert': 0,
56
+ 'n_expert_per_token': 0,
57
+ 'n_head': 32,
58
+ 'n_layer': 22,
59
+ 'n_query_groups': 4,
60
+ 'name': 'tiny-llama-1.1b',
61
+ 'norm_1': True,
62
+ 'norm_2': True,
63
+ 'norm_class_name': 'RMSNorm',
64
+ 'norm_eps': 1e-05,
65
+ 'norm_qk': False,
66
+ 'norm_qk_type': 'default',
67
+ 'padded_vocab_size': 32000,
68
+ 'padding_multiple': 64,
69
+ 'parallel_residual': False,
70
+ 'post_attention_norm': False,
71
+ 'post_mlp_norm': False,
72
+ 'rope_adjustments': None,
73
+ 'rope_base': 10000,
74
+ 'rope_condense_ratio': 1,
75
+ 'rope_indices': None,
76
+ 'rope_local_base_freq': None,
77
+ 'rotary_percentage': 1.0,
78
+ 'scale_embeddings': False,
79
+ 'shared_attention_norm': False,
80
+ 'sliding_window_indices': None,
81
+ 'sliding_window_size': None,
82
+ 'vocab_size': 32000},
83
+ 'model_name': 'tiny-llama-1.1b',
84
+ 'num_nodes': 1,
85
+ 'optimizer': "{'class_path': 'torch.optim.AdamW', 'init_args': {'lr': 0.0004, "
86
+ "'weight_decay': 0.1, 'betas': [0.9, 0.95]}}",
87
+ 'out_dir': PosixPath('out/pretrain/2501_full'),
88
+ 'ppl': False,
89
+ 'precision': 'bf16-mixed',
90
+ 'resume': False,
91
+ 'seed': 42,
92
+ 'tokenizer_dir': PosixPath('checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T'),
93
+ 'train': {'epochs': None,
94
+ 'global_batch_size': 512,
95
+ 'log_interval': 1,
96
+ 'lr_warmup_fraction': None,
97
+ 'lr_warmup_steps': 20,
98
+ 'max_norm': 1.0,
99
+ 'max_seq_length': 2048,
100
+ 'max_steps': None,
101
+ 'max_tokens': 266338304,
102
+ 'micro_batch_size': 4,
103
+ 'min_lr': 4e-05,
104
+ 'save_interval': 100,
105
+ 'tie_embeddings': None}}
106
+ Time to instantiate model: 0.04 seconds.
107
+ Total parameters: 1,100,048,384
108
+ [fix] out/pretrain/2412_full/final/lit_model.pth 是整包 state,提取 model 权重并原子覆盖...
109
+ [fix] 已覆盖为纯权重: out/pretrain/2412_full/final/lit_model.pth
110
+ Validating ...
111
+ Measured TFLOPs: 239.66
112
+ Epoch 1 | iter 32 step 1 | loss train: 1.403, val: 1.397 | iter time: 538.33 ms (step) remaining time: 0:50:36
113
+ Epoch 1 | iter 64 step 2 | loss train: 1.346, val: 1.397 | iter time: 356.55 ms (step) remaining time: 0:47:48
114
+ Epoch 1 | iter 96 step 3 | loss train: 1.353, val: 1.397 | iter time: 358.24 ms (step) remaining time: 0:46:47
115
+ Epoch 1 | iter 128 step 4 | loss train: 1.348, val: 1.397 | iter time: 357.23 ms (step) remaining time: 0:46:12
116
+ Epoch 1 | iter 160 step 5 | loss train: 1.310, val: 1.397 | iter time: 358.87 ms (step) remaining time: 0:45:47
117
+ Epoch 1 | iter 192 step 6 | loss train: 1.394, val: 1.397 | iter time: 358.63 ms (step) remaining time: 0:45:27
118
+ Epoch 1 | iter 224 step 7 | loss train: 1.366, val: 1.397 | iter time: 359.22 ms (step) remaining time: 0:45:11
119
+ Epoch 1 | iter 256 step 8 | loss train: 1.357, val: 1.397 | iter time: 360.37 ms (step) remaining time: 0:44:55
120
+ Epoch 1 | iter 288 step 9 | loss train: 1.362, val: 1.397 | iter time: 358.10 ms (step) remaining time: 0:44:41
121
+ Epoch 1 | iter 320 step 10 | loss train: 1.356, val: 1.397 | iter time: 360.02 ms (step) remaining time: 0:44:28
122
+ Epoch 1 | iter 352 step 11 | loss train: 1.460, val: 1.397 | iter time: 358.79 ms (step) remaining time: 0:44:15
123
+ Epoch 1 | iter 384 step 12 | loss train: 1.355, val: 1.397 | iter time: 360.48 ms (step) remaining time: 0:44:02
124
+ Epoch 1 | iter 416 step 13 | loss train: 1.316, val: 1.397 | iter time: 358.54 ms (step) remaining time: 0:43:50
125
+ Epoch 1 | iter 448 step 14 | loss train: 1.365, val: 1.397 | iter time: 359.56 ms (step) remaining time: 0:43:38
126
+ Epoch 1 | iter 480 step 15 | loss train: 1.356, val: 1.397 | iter time: 358.74 ms (step) remaining time: 0:43:26
127
+ Epoch 1 | iter 512 step 16 | loss train: 1.437, val: 1.397 | iter time: 358.07 ms (step) remaining time: 0:43:14
128
+ Epoch 1 | iter 544 step 17 | loss train: 1.358, val: 1.397 | iter time: 359.89 ms (step) remaining time: 0:43:02
129
+ Epoch 1 | iter 576 step 18 | loss train: 1.393, val: 1.397 | iter time: 359.03 ms (step) remaining time: 0:42:51
130
+ Epoch 1 | iter 608 step 19 | loss train: 1.371, val: 1.397 | iter time: 360.24 ms (step) remaining time: 0:42:39
131
+ Epoch 1 | iter 640 step 20 | loss train: 1.386, val: 1.397 | iter time: 360.41 ms (step) remaining time: 0:42:28
132
+ Epoch 1 | iter 672 step 21 | loss train: 1.377, val: 1.397 | iter time: 359.95 ms (step) remaining time: 0:42:17
133
+ Epoch 1 | iter 704 step 22 | loss train: 1.421, val: 1.397 | iter time: 359.80 ms (step) remaining time: 0:42:05
134
+ Epoch 1 | iter 736 step 23 | loss train: 1.428, val: 1.397 | iter time: 359.09 ms (step) remaining time: 0:41:54
135
+ Epoch 1 | iter 768 step 24 | loss train: 1.415, val: 1.397 | iter time: 359.53 ms (step) remaining time: 0:41:43
136
+ Epoch 1 | iter 800 step 25 | loss train: 1.410, val: 1.397 | iter time: 358.80 ms (step) remaining time: 0:41:32
137
+ Epoch 1 | iter 832 step 26 | loss train: 1.366, val: 1.397 | iter time: 359.72 ms (step) remaining time: 0:41:21
138
+ Epoch 1 | iter 864 step 27 | loss train: 1.458, val: 1.397 | iter time: 359.36 ms (step) remaining time: 0:41:10
139
+ Epoch 1 | iter 896 step 28 | loss train: 1.384, val: 1.397 | iter time: 357.36 ms (step) remaining time: 0:40:59
140
+ Epoch 1 | iter 928 step 29 | loss train: 1.417, val: 1.397 | iter time: 360.41 ms (step) remaining time: 0:40:47
141
+ Epoch 1 | iter 960 step 30 | loss train: 1.395, val: 1.397 | iter time: 358.60 ms (step) remaining time: 0:40:36
142
+ Epoch 1 | iter 992 step 31 | loss train: 1.435, val: 1.397 | iter time: 358.80 ms (step) remaining time: 0:40:25
143
+ Epoch 1 | iter 1024 step 32 | loss train: 1.411, val: 1.397 | iter time: 359.41 ms (step) remaining time: 0:40:14
144
+ Epoch 1 | iter 1056 step 33 | loss train: 1.312, val: 1.397 | iter time: 361.37 ms (step) remaining time: 0:40:03
145
+ Epoch 1 | iter 1088 step 34 | loss train: 1.466, val: 1.397 | iter time: 358.91 ms (step) remaining time: 0:39:52
146
+ Epoch 1 | iter 1120 step 35 | loss train: 1.367, val: 1.397 | iter time: 358.94 ms (step) remaining time: 0:39:41
147
+ Epoch 1 | iter 1152 step 36 | loss train: 1.369, val: 1.397 | iter time: 359.73 ms (step) remaining time: 0:39:30
148
+ Epoch 1 | iter 1184 step 37 | loss train: 1.385, val: 1.397 | iter time: 359.10 ms (step) remaining time: 0:39:19
149
+ Epoch 1 | iter 1216 step 38 | loss train: 1.473, val: 1.397 | iter time: 359.89 ms (step) remaining time: 0:39:08
150
+ Epoch 1 | iter 1248 step 39 | loss train: 1.340, val: 1.397 | iter time: 358.85 ms (step) remaining time: 0:38:57
151
+ Epoch 1 | iter 1280 step 40 | loss train: 1.378, val: 1.397 | iter time: 358.18 ms (step) remaining time: 0:38:46
152
+ Epoch 1 | iter 1312 step 41 | loss train: 1.403, val: 1.397 | iter time: 359.14 ms (step) remaining time: 0:38:36
153
+ Epoch 1 | iter 1344 step 42 | loss train: 1.380, val: 1.397 | iter time: 358.61 ms (step) remaining time: 0:38:26
154
+ Epoch 1 | iter 1376 step 43 | loss train: 1.413, val: 1.397 | iter time: 360.20 ms (step) remaining time: 0:38:15
155
+ Epoch 1 | iter 1408 step 44 | loss train: 1.352, val: 1.397 | iter time: 359.32 ms (step) remaining time: 0:38:04
156
+ Epoch 1 | iter 1440 step 45 | loss train: 1.362, val: 1.397 | iter time: 358.92 ms (step) remaining time: 0:37:53
157
+ Epoch 1 | iter 1472 step 46 | loss train: 1.383, val: 1.397 | iter time: 359.46 ms (step) remaining time: 0:37:42
158
+ Epoch 1 | iter 1504 step 47 | loss train: 1.391, val: 1.397 | iter time: 359.28 ms (step) remaining time: 0:37:31
159
+ Epoch 1 | iter 1536 step 48 | loss train: 1.394, val: 1.397 | iter time: 360.60 ms (step) remaining time: 0:37:20
160
+ Epoch 1 | iter 1568 step 49 | loss train: 1.350, val: 1.397 | iter time: 359.75 ms (step) remaining time: 0:37:09
161
+ Epoch 1 | iter 1600 step 50 | loss train: 1.448, val: 1.397 | iter time: 358.81 ms (step) remaining time: 0:36:58
162
+ Validating ...
163
+ iter 1600: val loss 1.3935, val time: 22346.96 ms
164
+ Epoch 1 | iter 1632 step 51 | loss train: 1.403, val: 1.393 | iter time: 359.56 ms (step) remaining time: 0:38:18
165
+ Epoch 1 | iter 1664 step 52 | loss train: 1.387, val: 1.393 | iter time: 361.43 ms (step) remaining time: 0:38:05
166
+ Epoch 1 | iter 1696 step 53 | loss train: 1.407, val: 1.393 | iter time: 361.57 ms (step) remaining time: 0:37:52
167
+ Epoch 1 | iter 1728 step 54 | loss train: 1.370, val: 1.393 | iter time: 360.65 ms (step) remaining time: 0:37:39
168
+ Epoch 1 | iter 1760 step 55 | loss train: 1.373, val: 1.393 | iter time: 360.84 ms (step) remaining time: 0:37:27
169
+ Epoch 1 | iter 1792 step 56 | loss train: 1.378, val: 1.393 | iter time: 358.78 ms (step) remaining time: 0:37:14
170
+ Epoch 1 | iter 1824 step 57 | loss train: 1.349, val: 1.393 | iter time: 358.85 ms (step) remaining time: 0:37:01
171
+ Epoch 1 | iter 1856 step 58 | loss train: 1.390, val: 1.393 | iter time: 359.22 ms (step) remaining time: 0:36:48
172
+ Epoch 1 | iter 1888 step 59 | loss train: 1.387, val: 1.393 | iter time: 360.96 ms (step) remaining time: 0:36:36
173
+ Epoch 1 | iter 1920 step 60 | loss train: 1.375, val: 1.393 | iter time: 359.90 ms (step) remaining time: 0:36:23
174
+ Epoch 1 | iter 1952 step 61 | loss train: 1.416, val: 1.393 | iter time: 358.98 ms (step) remaining time: 0:36:11
175
+ Epoch 1 | iter 1984 step 62 | loss train: 1.415, val: 1.393 | iter time: 360.22 ms (step) remaining time: 0:35:58
176
+ Epoch 1 | iter 2016 step 63 | loss train: 1.403, val: 1.393 | iter time: 360.85 ms (step) remaining time: 0:35:46
177
+ Epoch 1 | iter 2048 step 64 | loss train: 1.423, val: 1.393 | iter time: 361.80 ms (step) remaining time: 0:35:34
178
+ Epoch 1 | iter 2080 step 65 | loss train: 1.405, val: 1.393 | iter time: 360.37 ms (step) remaining time: 0:35:21
179
+ Epoch 1 | iter 2112 step 66 | loss train: 1.425, val: 1.393 | iter time: 361.07 ms (step) remaining time: 0:35:09
180
+ Epoch 1 | iter 2144 step 67 | loss train: 1.389, val: 1.393 | iter time: 358.03 ms (step) remaining time: 0:34:57
181
+ Epoch 1 | iter 2176 step 68 | loss train: 1.388, val: 1.393 | iter time: 359.64 ms (step) remaining time: 0:34:45
182
+ Epoch 1 | iter 2208 step 69 | loss train: 1.378, val: 1.393 | iter time: 360.29 ms (step) remaining time: 0:34:33
183
+ Epoch 1 | iter 2240 step 70 | loss train: 1.458, val: 1.393 | iter time: 360.03 ms (step) remaining time: 0:34:21
184
+ Epoch 1 | iter 2272 step 71 | loss train: 1.333, val: 1.393 | iter time: 359.79 ms (step) remaining time: 0:34:09
185
+ Epoch 1 | iter 2304 step 72 | loss train: 1.449, val: 1.393 | iter time: 359.22 ms (step) remaining time: 0:33:57
186
+ Epoch 1 | iter 2336 step 73 | loss train: 1.368, val: 1.393 | iter time: 361.36 ms (step) remaining time: 0:33:45
187
+ Epoch 1 | iter 2368 step 74 | loss train: 1.391, val: 1.393 | iter time: 360.59 ms (step) remaining time: 0:33:33
188
+ Epoch 1 | iter 2400 step 75 | loss train: 1.346, val: 1.393 | iter time: 360.80 ms (step) remaining time: 0:33:21
189
+ Epoch 1 | iter 2432 step 76 | loss train: 1.389, val: 1.393 | iter time: 358.67 ms (step) remaining time: 0:33:09
190
+ Epoch 1 | iter 2464 step 77 | loss train: 1.429, val: 1.393 | iter time: 362.44 ms (step) remaining time: 0:32:57
191
+ Epoch 1 | iter 2496 step 78 | loss train: 1.402, val: 1.393 | iter time: 359.60 ms (step) remaining time: 0:32:45
192
+ Epoch 1 | iter 2528 step 79 | loss train: 1.414, val: 1.393 | iter time: 360.31 ms (step) remaining time: 0:32:33
193
+ Epoch 1 | iter 2560 step 80 | loss train: 1.360, val: 1.393 | iter time: 360.49 ms (step) remaining time: 0:32:21
194
+ Epoch 1 | iter 2592 step 81 | loss train: 1.373, val: 1.393 | iter time: 361.48 ms (step) remaining time: 0:32:10
195
+ Epoch 1 | iter 2624 step 82 | loss train: 1.359, val: 1.393 | iter time: 358.61 ms (step) remaining time: 0:31:58
196
+ Epoch 1 | iter 2656 step 83 | loss train: 1.476, val: 1.393 | iter time: 358.76 ms (step) remaining time: 0:31:46
197
+ Epoch 1 | iter 2688 step 84 | loss train: 1.417, val: 1.393 | iter time: 360.95 ms (step) remaining time: 0:31:34
198
+ Epoch 1 | iter 2720 step 85 | loss train: 1.400, val: 1.393 | iter time: 359.84 ms (step) remaining time: 0:31:23
199
+ Epoch 1 | iter 2752 step 86 | loss train: 1.314, val: 1.393 | iter time: 360.06 ms (step) remaining time: 0:31:11
200
+ Epoch 1 | iter 2784 step 87 | loss train: 1.378, val: 1.393 | iter time: 359.69 ms (step) remaining time: 0:30:59
201
+ Epoch 1 | iter 2816 step 88 | loss train: 1.388, val: 1.393 | iter time: 358.09 ms (step) remaining time: 0:30:48
202
+ Epoch 1 | iter 2848 step 89 | loss train: 1.460, val: 1.393 | iter time: 358.62 ms (step) remaining time: 0:30:36
203
+ Epoch 1 | iter 2880 step 90 | loss train: 1.413, val: 1.393 | iter time: 358.29 ms (step) remaining time: 0:30:25
204
+ Epoch 1 | iter 2912 step 91 | loss train: 1.357, val: 1.393 | iter time: 359.89 ms (step) remaining time: 0:30:13
205
+ Epoch 1 | iter 2944 step 92 | loss train: 1.353, val: 1.393 | iter time: 360.57 ms (step) remaining time: 0:30:01
206
+ Epoch 1 | iter 2976 step 93 | loss train: 1.370, val: 1.393 | iter time: 359.00 ms (step) remaining time: 0:29:50
207
+ Epoch 1 | iter 3008 step 94 | loss train: 1.397, val: 1.393 | iter time: 358.75 ms (step) remaining time: 0:29:38
208
+ Epoch 1 | iter 3040 step 95 | loss train: 1.385, val: 1.393 | iter time: 358.80 ms (step) remaining time: 0:29:27
209
+ Epoch 1 | iter 3072 step 96 | loss train: 1.370, val: 1.393 | iter time: 361.56 ms (step) remaining time: 0:29:15
210
+ Epoch 1 | iter 3104 step 97 | loss train: 1.405, val: 1.393 | iter time: 360.19 ms (step) remaining time: 0:29:04
211
+ Epoch 1 | iter 3136 step 98 | loss train: 1.418, val: 1.393 | iter time: 359.04 ms (step) remaining time: 0:28:52
212
+ Epoch 1 | iter 3168 step 99 | loss train: 1.384, val: 1.393 | iter time: 358.87 ms (step) remaining time: 0:28:41
213
+ Epoch 1 | iter 3200 step 100 | loss train: 1.380, val: 1.393 | iter time: 360.67 ms (step) remaining time: 0:28:29
214
+ Validating ...
215
+ iter 3200: val loss 1.3468, val time: 22312.98 ms
216
+ Saving checkpoint to 'out/pretrain/2501_full/step-00000100/lit_model.pth'
217
+ Epoch 1 | iter 3232 step 101 | loss train: 1.368, val: 1.347 | iter time: 358.34 ms (step) remaining time: 0:29:17
218
+ Epoch 1 | iter 3264 step 102 | loss train: 1.419, val: 1.347 | iter time: 359.27 ms (step) remaining time: 0:29:05
219
+ Epoch 1 | iter 3296 step 103 | loss train: 1.440, val: 1.347 | iter time: 359.18 ms (step) remaining time: 0:28:52
220
+ Epoch 1 | iter 3328 step 104 | loss train: 1.408, val: 1.347 | iter time: 361.67 ms (step) remaining time: 0:28:40
221
+ Epoch 1 | iter 3360 step 105 | loss train: 1.382, val: 1.347 | iter time: 359.36 ms (step) remaining time: 0:28:28
222
+ Epoch 1 | iter 3392 step 106 | loss train: 1.395, val: 1.347 | iter time: 358.25 ms (step) remaining time: 0:28:16
223
+ Epoch 1 | iter 3424 step 107 | loss train: 1.321, val: 1.347 | iter time: 359.61 ms (step) remaining time: 0:28:03
224
+ Epoch 1 | iter 3456 step 108 | loss train: 1.329, val: 1.347 | iter time: 359.73 ms (step) remaining time: 0:27:51
225
+ Epoch 1 | iter 3488 step 109 | loss train: 1.373, val: 1.347 | iter time: 361.43 ms (step) remaining time: 0:27:39
226
+ Epoch 1 | iter 3520 step 110 | loss train: 1.408, val: 1.347 | iter time: 359.68 ms (step) remaining time: 0:27:27
227
+ Epoch 1 | iter 3552 step 111 | loss train: 1.320, val: 1.347 | iter time: 359.40 ms (step) remaining time: 0:27:14
228
+ Epoch 1 | iter 3584 step 112 | loss train: 1.341, val: 1.347 | iter time: 358.33 ms (step) remaining time: 0:27:02
229
+ Epoch 1 | iter 3616 step 113 | loss train: 1.358, val: 1.347 | iter time: 359.00 ms (step) remaining time: 0:26:50
230
+ Epoch 1 | iter 3648 step 114 | loss train: 1.407, val: 1.347 | iter time: 359.78 ms (step) remaining time: 0:26:38
231
+ Epoch 1 | iter 3680 step 115 | loss train: 1.394, val: 1.347 | iter time: 357.77 ms (step) remaining time: 0:26:26
232
+ Epoch 1 | iter 3712 step 116 | loss train: 1.353, val: 1.347 | iter time: 361.80 ms (step) remaining time: 0:26:14
233
+ Epoch 1 | iter 3744 step 117 | loss train: 1.401, val: 1.347 | iter time: 359.22 ms (step) remaining time: 0:26:02
234
+ Epoch 1 | iter 3776 step 118 | loss train: 1.401, val: 1.347 | iter time: 359.99 ms (step) remaining time: 0:25:50
235
+ Epoch 1 | iter 3808 step 119 | loss train: 1.425, val: 1.347 | iter time: 359.97 ms (step) remaining time: 0:25:38
236
+ Epoch 1 | iter 3840 step 120 | loss train: 1.401, val: 1.347 | iter time: 360.56 ms (step) remaining time: 0:25:26
237
+ Epoch 1 | iter 3872 step 121 | loss train: 1.383, val: 1.347 | iter time: 358.77 ms (step) remaining time: 0:25:14
238
+ Epoch 1 | iter 3904 step 122 | loss train: 1.359, val: 1.347 | iter time: 361.14 ms (step) remaining time: 0:25:02
239
+ Epoch 1 | iter 3936 step 123 | loss train: 1.454, val: 1.347 | iter time: 359.28 ms (step) remaining time: 0:24:51
240
+ Epoch 1 | iter 3968 step 124 | loss train: 1.363, val: 1.347 | iter time: 360.13 ms (step) remaining time: 0:24:39
241
+ Epoch 1 | iter 4000 step 125 | loss train: 1.365, val: 1.347 | iter time: 360.35 ms (step) remaining time: 0:24:27
242
+ Epoch 1 | iter 4032 step 126 | loss train: 1.340, val: 1.347 | iter time: 358.83 ms (step) remaining time: 0:24:15
243
+ Epoch 1 | iter 4064 step 127 | loss train: 1.352, val: 1.347 | iter time: 362.50 ms (step) remaining time: 0:24:03
244
+ Epoch 1 | iter 4096 step 128 | loss train: 1.310, val: 1.347 | iter time: 359.94 ms (step) remaining time: 0:23:51
245
+ Epoch 1 | iter 4128 step 129 | loss train: 1.367, val: 1.347 | iter time: 359.28 ms (step) remaining time: 0:23:39
246
+ Epoch 1 | iter 4160 step 130 | loss train: 1.372, val: 1.347 | iter time: 359.84 ms (step) remaining time: 0:23:27
247
+ Epoch 1 | iter 4192 step 131 | loss train: 1.379, val: 1.347 | iter time: 359.89 ms (step) remaining time: 0:23:16
248
+ Epoch 1 | iter 4224 step 132 | loss train: 1.371, val: 1.347 | iter time: 359.55 ms (step) remaining time: 0:23:04
249
+ Epoch 1 | iter 4256 step 133 | loss train: 1.293, val: 1.347 | iter time: 359.16 ms (step) remaining time: 0:22:52
250
+ Epoch 1 | iter 4288 step 134 | loss train: 1.357, val: 1.347 | iter time: 360.69 ms (step) remaining time: 0:22:40
251
+ Epoch 1 | iter 4320 step 135 | loss train: 1.372, val: 1.347 | iter time: 358.83 ms (step) remaining time: 0:22:28
252
+ Epoch 1 | iter 4352 step 136 | loss train: 1.380, val: 1.347 | iter time: 359.60 ms (step) remaining time: 0:22:17
253
+ Epoch 1 | iter 4384 step 137 | loss train: 1.325, val: 1.347 | iter time: 359.80 ms (step) remaining time: 0:22:05
254
+ Epoch 1 | iter 4416 step 138 | loss train: 1.310, val: 1.347 | iter time: 360.75 ms (step) remaining time: 0:21:53
255
+ Epoch 1 | iter 4448 step 139 | loss train: 1.343, val: 1.347 | iter time: 360.34 ms (step) remaining time: 0:21:42
256
+ Epoch 1 | iter 4480 step 140 | loss train: 1.405, val: 1.347 | iter time: 358.82 ms (step) remaining time: 0:21:30
257
+ Epoch 1 | iter 4512 step 141 | loss train: 1.398, val: 1.347 | iter time: 359.01 ms (step) remaining time: 0:21:18
258
+ Epoch 1 | iter 4544 step 142 | loss train: 1.304, val: 1.347 | iter time: 361.06 ms (step) remaining time: 0:21:07
259
+ Epoch 1 | iter 4576 step 143 | loss train: 1.335, val: 1.347 | iter time: 361.51 ms (step) remaining time: 0:20:55
260
+ Epoch 1 | iter 4608 step 144 | loss train: 1.267, val: 1.347 | iter time: 360.34 ms (step) remaining time: 0:20:43
261
+ Epoch 1 | iter 4640 step 145 | loss train: 1.310, val: 1.347 | iter time: 359.48 ms (step) remaining time: 0:20:32
262
+ Epoch 1 | iter 4672 step 146 | loss train: 1.347, val: 1.347 | iter time: 360.42 ms (step) remaining time: 0:20:20
263
+ Epoch 1 | iter 4704 step 147 | loss train: 1.340, val: 1.347 | iter time: 360.08 ms (step) remaining time: 0:20:08
264
+ Epoch 1 | iter 4736 step 148 | loss train: 1.320, val: 1.347 | iter time: 360.98 ms (step) remaining time: 0:19:57
265
+ Epoch 1 | iter 4768 step 149 | loss train: 1.306, val: 1.347 | iter time: 359.88 ms (step) remaining time: 0:19:45
266
+ Epoch 1 | iter 4800 step 150 | loss train: 1.354, val: 1.347 | iter time: 358.99 ms (step) remaining time: 0:19:33
267
+ Validating ...
268
+ iter 4800: val loss 1.2983, val time: 22355.73 ms
269
+ Epoch 1 | iter 4832 step 151 | loss train: 1.368, val: 1.298 | iter time: 359.06 ms (step) remaining time: 0:19:37
270
+ Epoch 1 | iter 4864 step 152 | loss train: 1.281, val: 1.298 | iter time: 358.58 ms (step) remaining time: 0:19:25
271
+ Epoch 1 | iter 4896 step 153 | loss train: 1.368, val: 1.298 | iter time: 360.00 ms (step) remaining time: 0:19:14
272
+ Epoch 1 | iter 4928 step 154 | loss train: 1.267, val: 1.298 | iter time: 358.61 ms (step) remaining time: 0:19:02
273
+ Epoch 1 | iter 4960 step 155 | loss train: 1.349, val: 1.298 | iter time: 359.30 ms (step) remaining time: 0:18:50
274
+ Epoch 1 | iter 4992 step 156 | loss train: 1.388, val: 1.298 | iter time: 359.72 ms (step) remaining time: 0:18:38
275
+ Epoch 1 | iter 5024 step 157 | loss train: 1.293, val: 1.298 | iter time: 360.07 ms (step) remaining time: 0:18:26
276
+ Epoch 1 | iter 5056 step 158 | loss train: 1.322, val: 1.298 | iter time: 360.49 ms (step) remaining time: 0:18:15
277
+ Epoch 1 | iter 5088 step 159 | loss train: 1.377, val: 1.298 | iter time: 359.55 ms (step) remaining time: 0:18:03
278
+ Epoch 1 | iter 5120 step 160 | loss train: 1.392, val: 1.298 | iter time: 360.04 ms (step) remaining time: 0:17:51
279
+ Epoch 1 | iter 5152 step 161 | loss train: 1.413, val: 1.298 | iter time: 360.48 ms (step) remaining time: 0:17:39
280
+ Epoch 1 | iter 5184 step 162 | loss train: 1.471, val: 1.298 | iter time: 358.96 ms (step) remaining time: 0:17:28
281
+ Epoch 1 | iter 5216 step 163 | loss train: 1.373, val: 1.298 | iter time: 359.37 ms (step) remaining time: 0:17:16
282
+ Epoch 1 | iter 5248 step 164 | loss train: 1.304, val: 1.298 | iter time: 358.78 ms (step) remaining time: 0:17:04
283
+ Epoch 1 | iter 5280 step 165 | loss train: 1.378, val: 1.298 | iter time: 360.73 ms (step) remaining time: 0:16:53
284
+ Epoch 1 | iter 5312 step 166 | loss train: 1.344, val: 1.298 | iter time: 358.77 ms (step) remaining time: 0:16:41
285
+ Epoch 1 | iter 5344 step 167 | loss train: 1.392, val: 1.298 | iter time: 361.08 ms (step) remaining time: 0:16:29
286
+ Epoch 1 | iter 5376 step 168 | loss train: 1.375, val: 1.298 | iter time: 359.46 ms (step) remaining time: 0:16:18
287
+ Epoch 1 | iter 5408 step 169 | loss train: 1.359, val: 1.298 | iter time: 359.83 ms (step) remaining time: 0:16:06
288
+ Epoch 1 | iter 5440 step 170 | loss train: 1.386, val: 1.298 | iter time: 358.48 ms (step) remaining time: 0:15:55
289
+ Epoch 1 | iter 5472 step 171 | loss train: 1.346, val: 1.298 | iter time: 359.73 ms (step) remaining time: 0:15:43
290
+ Epoch 1 | iter 5504 step 172 | loss train: 1.356, val: 1.298 | iter time: 359.56 ms (step) remaining time: 0:15:31
291
+ Epoch 1 | iter 5536 step 173 | loss train: 1.367, val: 1.298 | iter time: 359.31 ms (step) remaining time: 0:15:20
292
+ Epoch 1 | iter 5568 step 174 | loss train: 1.336, val: 1.298 | iter time: 357.40 ms (step) remaining time: 0:15:08
293
+ Epoch 1 | iter 5600 step 175 | loss train: 1.339, val: 1.298 | iter time: 358.61 ms (step) remaining time: 0:14:57
294
+ Epoch 1 | iter 5632 step 176 | loss train: 1.367, val: 1.298 | iter time: 358.90 ms (step) remaining time: 0:14:45
295
+ Epoch 1 | iter 5664 step 177 | loss train: 1.427, val: 1.298 | iter time: 358.81 ms (step) remaining time: 0:14:33
296
+ Epoch 1 | iter 5696 step 178 | loss train: 1.292, val: 1.298 | iter time: 358.37 ms (step) remaining time: 0:14:22
297
+ Epoch 1 | iter 5728 step 179 | loss train: 1.315, val: 1.298 | iter time: 360.13 ms (step) remaining time: 0:14:10
298
+ Epoch 1 | iter 5760 step 180 | loss train: 1.341, val: 1.298 | iter time: 360.82 ms (step) remaining time: 0:13:59
299
+ Epoch 1 | iter 5792 step 181 | loss train: 1.345, val: 1.298 | iter time: 359.64 ms (step) remaining time: 0:13:47
300
+ Epoch 1 | iter 5824 step 182 | loss train: 1.349, val: 1.298 | iter time: 359.87 ms (step) remaining time: 0:13:36
301
+ Epoch 1 | iter 5856 step 183 | loss train: 1.343, val: 1.298 | iter time: 360.25 ms (step) remaining time: 0:13:24
302
+ Epoch 1 | iter 5888 step 184 | loss train: 1.327, val: 1.298 | iter time: 359.87 ms (step) remaining time: 0:13:13
303
+ Epoch 1 | iter 5920 step 185 | loss train: 1.292, val: 1.298 | iter time: 360.61 ms (step) remaining time: 0:13:01
304
+ Epoch 1 | iter 5952 step 186 | loss train: 1.345, val: 1.298 | iter time: 358.24 ms (step) remaining time: 0:12:50
305
+ Epoch 1 | iter 5984 step 187 | loss train: 1.355, val: 1.298 | iter time: 360.62 ms (step) remaining time: 0:12:38
306
+ Epoch 1 | iter 6016 step 188 | loss train: 1.360, val: 1.298 | iter time: 360.38 ms (step) remaining time: 0:12:27
307
+ Epoch 1 | iter 6048 step 189 | loss train: 1.332, val: 1.298 | iter time: 360.11 ms (step) remaining time: 0:12:15
308
+ Epoch 1 | iter 6080 step 190 | loss train: 1.324, val: 1.298 | iter time: 360.02 ms (step) remaining time: 0:12:04
309
+ Epoch 1 | iter 6112 step 191 | loss train: 1.351, val: 1.298 | iter time: 360.18 ms (step) remaining time: 0:11:52
310
+ Epoch 1 | iter 6144 step 192 | loss train: 1.316, val: 1.298 | iter time: 358.14 ms (step) remaining time: 0:11:41
311
+ Epoch 1 | iter 6176 step 193 | loss train: 1.433, val: 1.298 | iter time: 359.33 ms (step) remaining time: 0:11:30
312
+ Epoch 1 | iter 6208 step 194 | loss train: 1.400, val: 1.298 | iter time: 361.19 ms (step) remaining time: 0:11:18
313
+ Epoch 1 | iter 6240 step 195 | loss train: 1.394, val: 1.298 | iter time: 359.00 ms (step) remaining time: 0:11:07
314
+ Epoch 1 | iter 6272 step 196 | loss train: 1.341, val: 1.298 | iter time: 358.43 ms (step) remaining time: 0:10:55
315
+ Epoch 1 | iter 6304 step 197 | loss train: 1.340, val: 1.298 | iter time: 358.66 ms (step) remaining time: 0:10:44
316
+ Epoch 1 | iter 6336 step 198 | loss train: 1.325, val: 1.298 | iter time: 360.47 ms (step) remaining time: 0:10:32
317
+ Epoch 1 | iter 6368 step 199 | loss train: 1.261, val: 1.298 | iter time: 360.92 ms (step) remaining time: 0:10:21
318
+ Epoch 1 | iter 6400 step 200 | loss train: 1.374, val: 1.298 | iter time: 361.31 ms (step) remaining time: 0:10:09
319
+ Validating ...
320
+ iter 6400: val loss 1.2520, val time: 22361.19 ms
321
+ Saving checkpoint to 'out/pretrain/2501_full/step-00000200/lit_model.pth'
322
+ Epoch 1 | iter 6432 step 201 | loss train: 1.287, val: 1.252 | iter time: 356.95 ms (step) remaining time: 0:10:08
323
+ Epoch 1 | iter 6464 step 202 | loss train: 1.426, val: 1.252 | iter time: 357.79 ms (step) remaining time: 0:09:57
324
+ Epoch 1 | iter 6496 step 203 | loss train: 1.360, val: 1.252 | iter time: 356.87 ms (step) remaining time: 0:09:45
325
+ Epoch 1 | iter 6528 step 204 | loss train: 1.257, val: 1.252 | iter time: 360.23 ms (step) remaining time: 0:09:33
326
+ Epoch 1 | iter 6560 step 205 | loss train: 1.297, val: 1.252 | iter time: 360.28 ms (step) remaining time: 0:09:22
327
+ Epoch 1 | iter 6592 step 206 | loss train: 1.312, val: 1.252 | iter time: 360.69 ms (step) remaining time: 0:09:10
328
+ Epoch 1 | iter 6624 step 207 | loss train: 1.291, val: 1.252 | iter time: 358.29 ms (step) remaining time: 0:08:59
329
+ Epoch 1 | iter 6656 step 208 | loss train: 1.322, val: 1.252 | iter time: 358.72 ms (step) remaining time: 0:08:47
330
+ Epoch 1 | iter 6688 step 209 | loss train: 1.394, val: 1.252 | iter time: 359.99 ms (step) remaining time: 0:08:35
331
+ Epoch 1 | iter 6720 step 210 | loss train: 1.287, val: 1.252 | iter time: 357.99 ms (step) remaining time: 0:08:24
332
+ Epoch 1 | iter 6752 step 211 | loss train: 1.333, val: 1.252 | iter time: 359.27 ms (step) remaining time: 0:08:12
333
+ Epoch 1 | iter 6784 step 212 | loss train: 1.279, val: 1.252 | iter time: 360.06 ms (step) remaining time: 0:08:01
334
+ Epoch 1 | iter 6816 step 213 | loss train: 1.419, val: 1.252 | iter time: 358.95 ms (step) remaining time: 0:07:49
335
+ Epoch 1 | iter 6848 step 214 | loss train: 1.332, val: 1.252 | iter time: 359.75 ms (step) remaining time: 0:07:37
336
+ Epoch 1 | iter 6880 step 215 | loss train: 1.278, val: 1.252 | iter time: 359.60 ms (step) remaining time: 0:07:26
337
+ Epoch 1 | iter 6912 step 216 | loss train: 1.283, val: 1.252 | iter time: 360.78 ms (step) remaining time: 0:07:14
338
+ Epoch 1 | iter 6944 step 217 | loss train: 1.324, val: 1.252 | iter time: 360.59 ms (step) remaining time: 0:07:03
339
+ Epoch 1 | iter 6976 step 218 | loss train: 1.277, val: 1.252 | iter time: 361.70 ms (step) remaining time: 0:06:51
340
+ Epoch 1 | iter 7008 step 219 | loss train: 1.306, val: 1.252 | iter time: 357.83 ms (step) remaining time: 0:06:40
341
+ Epoch 1 | iter 7040 step 220 | loss train: 1.367, val: 1.252 | iter time: 358.98 ms (step) remaining time: 0:06:28
342
+ Epoch 1 | iter 7072 step 221 | loss train: 1.402, val: 1.252 | iter time: 360.28 ms (step) remaining time: 0:06:17
343
+ Epoch 1 | iter 7104 step 222 | loss train: 1.307, val: 1.252 | iter time: 360.30 ms (step) remaining time: 0:06:05
344
+ Epoch 1 | iter 7136 step 223 | loss train: 1.291, val: 1.252 | iter time: 359.98 ms (step) remaining time: 0:05:54
345
+ Epoch 1 | iter 7168 step 224 | loss train: 1.316, val: 1.252 | iter time: 361.93 ms (step) remaining time: 0:05:42
346
+ Epoch 1 | iter 7200 step 225 | loss train: 1.304, val: 1.252 | iter time: 359.45 ms (step) remaining time: 0:05:31
347
+ Epoch 1 | iter 7232 step 226 | loss train: 1.264, val: 1.252 | iter time: 359.71 ms (step) remaining time: 0:05:19
348
+ Epoch 1 | iter 7264 step 227 | loss train: 1.299, val: 1.252 | iter time: 360.36 ms (step) remaining time: 0:05:08
349
+ Epoch 1 | iter 7296 step 228 | loss train: 1.326, val: 1.252 | iter time: 358.74 ms (step) remaining time: 0:04:56
350
+ Epoch 1 | iter 7328 step 229 | loss train: 1.233, val: 1.252 | iter time: 359.42 ms (step) remaining time: 0:04:45
351
+ Epoch 1 | iter 7360 step 230 | loss train: 1.288, val: 1.252 | iter time: 360.07 ms (step) remaining time: 0:04:33
352
+ Epoch 1 | iter 7392 step 231 | loss train: 1.249, val: 1.252 | iter time: 360.96 ms (step) remaining time: 0:04:22
353
+ Epoch 1 | iter 7424 step 232 | loss train: 1.300, val: 1.252 | iter time: 361.62 ms (step) remaining time: 0:04:10
354
+ Epoch 1 | iter 7456 step 233 | loss train: 1.320, val: 1.252 | iter time: 359.72 ms (step) remaining time: 0:03:59
355
+ Epoch 1 | iter 7488 step 234 | loss train: 1.304, val: 1.252 | iter time: 358.41 ms (step) remaining time: 0:03:48
356
+ Epoch 1 | iter 7520 step 235 | loss train: 1.363, val: 1.252 | iter time: 360.15 ms (step) remaining time: 0:03:36
357
+ Epoch 1 | iter 7552 step 236 | loss train: 1.247, val: 1.252 | iter time: 359.70 ms (step) remaining time: 0:03:25
358
+ Epoch 1 | iter 7584 step 237 | loss train: 1.220, val: 1.252 | iter time: 358.00 ms (step) remaining time: 0:03:13
359
+ Epoch 1 | iter 7616 step 238 | loss train: 1.325, val: 1.252 | iter time: 359.85 ms (step) remaining time: 0:03:02
360
+ Epoch 1 | iter 7648 step 239 | loss train: 1.351, val: 1.252 | iter time: 357.77 ms (step) remaining time: 0:02:50
361
+ Epoch 1 | iter 7680 step 240 | loss train: 1.251, val: 1.252 | iter time: 360.29 ms (step) remaining time: 0:02:39
362
+ Epoch 1 | iter 7712 step 241 | loss train: 1.265, val: 1.252 | iter time: 360.02 ms (step) remaining time: 0:02:27
363
+ Epoch 1 | iter 7744 step 242 | loss train: 1.270, val: 1.252 | iter time: 361.47 ms (step) remaining time: 0:02:16
364
+ Epoch 1 | iter 7776 step 243 | loss train: 1.287, val: 1.252 | iter time: 359.72 ms (step) remaining time: 0:02:05
365
+ Epoch 1 | iter 7808 step 244 | loss train: 1.293, val: 1.252 | iter time: 359.49 ms (step) remaining time: 0:01:53
366
+ Epoch 1 | iter 7840 step 245 | loss train: 1.219, val: 1.252 | iter time: 358.47 ms (step) remaining time: 0:01:42
367
+ Epoch 1 | iter 7872 step 246 | loss train: 1.378, val: 1.252 | iter time: 360.27 ms (step) remaining time: 0:01:31
368
+ Epoch 1 | iter 7904 step 247 | loss train: 1.245, val: 1.252 | iter time: 357.73 ms (step) remaining time: 0:01:19
369
+ Epoch 1 | iter 7936 step 248 | loss train: 1.227, val: 1.252 | iter time: 358.04 ms (step) remaining time: 0:01:08
370
+ Epoch 1 | iter 7968 step 249 | loss train: 1.286, val: 1.252 | iter time: 359.20 ms (step) remaining time: 0:00:56
371
+ Epoch 1 | iter 8000 step 250 | loss train: 1.272, val: 1.252 | iter time: 359.58 ms (step) remaining time: 0:00:45
372
+ Validating ...
373
+ iter 8000: val loss 1.2134, val time: 22383.99 ms
374
+ Epoch 1 | iter 8032 step 251 | loss train: 1.333, val: 1.213 | iter time: 360.62 ms (step) remaining time: 0:00:34
375
+ Epoch 1 | iter 8064 step 252 | loss train: 1.351, val: 1.213 | iter time: 359.46 ms (step) remaining time: 0:00:22
376
+ Epoch 1 | iter 8096 step 253 | loss train: 1.283, val: 1.213 | iter time: 361.84 ms (step) remaining time: 0:00:11
377
+ Epoch 2 | iter 8128 step 254 | loss train: 1.321, val: 1.213 | iter time: 360.06 ms (step) remaining time: 0:00:00
378
+ Validating ...
379
+ Final evaluation | val loss: 1.210 | val ppl: 3.354
380
+ Saving checkpoint to 'out/pretrain/2501_full/final/lit_model.pth'
381
+ ----------------------------------------
382
+ | Performance
383
+ | - Total tokens : 266,338,304
384
+ | - Training Time : 2971.64 s
385
+ | - Tok/sec : 143.04 tok/s
386
+ | ----------------------------------------
387
+ | Memory Usage
388
+ | - Memory Used : 26.32 GB
389
+ ----------------------------------------
out/pretrain/tinyllama_3_epoch/2407/final/config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LlamaForCausalLM"
4
+ ],
5
+ "bos_token_id": 1,
6
+ "eos_token_id": 2,
7
+ "hidden_act": "silu",
8
+ "hidden_size": 2048,
9
+ "initializer_range": 0.02,
10
+ "intermediate_size": 5632,
11
+ "max_position_embeddings": 2048,
12
+ "model_type": "llama",
13
+ "num_attention_heads": 32,
14
+ "num_hidden_layers": 22,
15
+ "num_key_value_heads": 4,
16
+ "pretraining_tp": 1,
17
+ "rms_norm_eps": 1e-05,
18
+ "rope_scaling": null,
19
+ "tie_word_embeddings": false,
20
+ "torch_dtype": "float32",
21
+ "transformers_version": "4.31.0.dev0",
22
+ "use_cache": true,
23
+ "vocab_size": 32000
24
+ }
out/pretrain/tinyllama_3_epoch/2407/final/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 1,
3
+ "eos_token_id": 2,
4
+ "pad_token_id": 0,
5
+ "max_length": 2048,
6
+ "transformers_version": "4.31.0.dev0"
7
+ }
out/pretrain/tinyllama_3_epoch/2407/final/hyperparameters.yaml ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model_name: tiny-llama-1.1b
2
+ out_dir: out/pretrain/tinyllama_3_epoch/2407
3
+ precision: bf16-mixed
4
+ initial_checkpoint_dir: checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
5
+ resume: false
6
+ data:
7
+ class_path: litgpt.data.Arxiv
8
+ init_args:
9
+ data_path: litgpt/data/arxiv
10
+ seed: 42
11
+ num_workers: 0
12
+ use_starcoder: true
13
+ ppl: false
14
+ data_dir: litgpt/data/arxiv_tinyllama_tokenized/2407
15
+ train:
16
+ save_interval: 100
17
+ log_interval: 1
18
+ global_batch_size: 512
19
+ micro_batch_size: 4
20
+ lr_warmup_steps: 20
21
+ max_tokens: 635437056
22
+ max_seq_length: 2048
23
+ max_norm: 1.0
24
+ min_lr: 4.0e-05
25
+ eval:
26
+ interval: 50
27
+ max_iters: 200
28
+ initial_validation: true
29
+ final_validation: true
30
+ evaluate_example: first
31
+ log: {}
32
+ optimizer:
33
+ class_path: torch.optim.AdamW
34
+ init_args:
35
+ lr: 4.0e-05
36
+ weight_decay: 0.1
37
+ betas:
38
+ - 0.9
39
+ - 0.95
40
+ devices: auto
41
+ num_nodes: 1
42
+ tokenizer_dir: checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
43
+ logger_name: tensorboard
44
+ seed: 42
out/pretrain/tinyllama_3_epoch/2407/final/model_config.yaml ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ attention_logit_softcapping: null
2
+ attention_scores_scalar: null
3
+ attn_bias: false
4
+ bias: false
5
+ block_size: 2048
6
+ final_logit_softcapping: null
7
+ gelu_approximate: none
8
+ head_size: 64
9
+ hf_config:
10
+ name: TinyLlama-1.1B-intermediate-step-1431k-3T
11
+ org: TinyLlama
12
+ intermediate_size: 5632
13
+ lm_head_bias: false
14
+ mlp_class_name: LLaMAMLP
15
+ moe_intermediate_size: null
16
+ n_embd: 2048
17
+ n_expert: 0
18
+ n_expert_per_token: 0
19
+ n_head: 32
20
+ n_layer: 22
21
+ n_query_groups: 4
22
+ name: tiny-llama-1.1b
23
+ norm_1: true
24
+ norm_2: true
25
+ norm_class_name: RMSNorm
26
+ norm_eps: 1.0e-05
27
+ norm_qk: false
28
+ norm_qk_type: default
29
+ padded_vocab_size: 32000
30
+ padding_multiple: 64
31
+ parallel_residual: false
32
+ post_attention_norm: false
33
+ post_mlp_norm: false
34
+ rope_adjustments: null
35
+ rope_base: 10000
36
+ rope_condense_ratio: 1
37
+ rope_indices: null
38
+ rope_local_base_freq: null
39
+ rotary_percentage: 1.0
40
+ scale_embeddings: false
41
+ shared_attention_norm: false
42
+ sliding_window_indices: null
43
+ sliding_window_size: null
44
+ vocab_size: 32000
out/pretrain/tinyllama_3_epoch/2407/final/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
out/pretrain/tinyllama_3_epoch/2407/final/tokenizer_config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "bos_token": {
5
+ "__type": "AddedToken",
6
+ "content": "<s>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "clean_up_tokenization_spaces": false,
13
+ "eos_token": {
14
+ "__type": "AddedToken",
15
+ "content": "</s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "legacy": false,
22
+ "model_max_length": 1000000000000000019884624838656,
23
+ "pad_token": null,
24
+ "padding_side": "right",
25
+ "sp_model_kwargs": {},
26
+ "tokenizer_class": "LlamaTokenizer",
27
+ "unk_token": {
28
+ "__type": "AddedToken",
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false
34
+ }
35
+ }
out/pretrain/tinyllama_3_epoch/2408/final/config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LlamaForCausalLM"
4
+ ],
5
+ "bos_token_id": 1,
6
+ "eos_token_id": 2,
7
+ "hidden_act": "silu",
8
+ "hidden_size": 2048,
9
+ "initializer_range": 0.02,
10
+ "intermediate_size": 5632,
11
+ "max_position_embeddings": 2048,
12
+ "model_type": "llama",
13
+ "num_attention_heads": 32,
14
+ "num_hidden_layers": 22,
15
+ "num_key_value_heads": 4,
16
+ "pretraining_tp": 1,
17
+ "rms_norm_eps": 1e-05,
18
+ "rope_scaling": null,
19
+ "tie_word_embeddings": false,
20
+ "torch_dtype": "float32",
21
+ "transformers_version": "4.31.0.dev0",
22
+ "use_cache": true,
23
+ "vocab_size": 32000
24
+ }
out/pretrain/tinyllama_3_epoch/2408/final/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 1,
3
+ "eos_token_id": 2,
4
+ "pad_token_id": 0,
5
+ "max_length": 2048,
6
+ "transformers_version": "4.31.0.dev0"
7
+ }
out/pretrain/tinyllama_3_epoch/2408/final/hyperparameters.yaml ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model_name: tiny-llama-1.1b
2
+ out_dir: out/pretrain/tinyllama_3_epoch/2408
3
+ precision: bf16-mixed
4
+ initial_checkpoint_dir: out/pretrain/tinyllama_3_epoch/2407/final
5
+ resume: false
6
+ data:
7
+ class_path: litgpt.data.Arxiv
8
+ init_args:
9
+ data_path: litgpt/data/arxiv
10
+ seed: 42
11
+ num_workers: 0
12
+ use_starcoder: true
13
+ ppl: false
14
+ data_dir: litgpt/data/arxiv_tinyllama_tokenized/2408
15
+ train:
16
+ save_interval: 100
17
+ log_interval: 1
18
+ global_batch_size: 512
19
+ micro_batch_size: 4
20
+ lr_warmup_steps: 20
21
+ max_tokens: 531628032
22
+ max_seq_length: 2048
23
+ max_norm: 1.0
24
+ min_lr: 4.0e-05
25
+ eval:
26
+ interval: 50
27
+ max_iters: 200
28
+ initial_validation: true
29
+ final_validation: true
30
+ evaluate_example: first
31
+ log: {}
32
+ optimizer:
33
+ class_path: torch.optim.AdamW
34
+ init_args:
35
+ lr: 4.0e-05
36
+ weight_decay: 0.1
37
+ betas:
38
+ - 0.9
39
+ - 0.95
40
+ devices: auto
41
+ num_nodes: 1
42
+ tokenizer_dir: checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
43
+ logger_name: tensorboard
44
+ seed: 42
out/pretrain/tinyllama_3_epoch/2408/final/model_config.yaml ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ attention_logit_softcapping: null
2
+ attention_scores_scalar: null
3
+ attn_bias: false
4
+ bias: false
5
+ block_size: 2048
6
+ final_logit_softcapping: null
7
+ gelu_approximate: none
8
+ head_size: 64
9
+ hf_config:
10
+ name: TinyLlama-1.1B-intermediate-step-1431k-3T
11
+ org: TinyLlama
12
+ intermediate_size: 5632
13
+ lm_head_bias: false
14
+ mlp_class_name: LLaMAMLP
15
+ moe_intermediate_size: null
16
+ n_embd: 2048
17
+ n_expert: 0
18
+ n_expert_per_token: 0
19
+ n_head: 32
20
+ n_layer: 22
21
+ n_query_groups: 4
22
+ name: tiny-llama-1.1b
23
+ norm_1: true
24
+ norm_2: true
25
+ norm_class_name: RMSNorm
26
+ norm_eps: 1.0e-05
27
+ norm_qk: false
28
+ norm_qk_type: default
29
+ padded_vocab_size: 32000
30
+ padding_multiple: 64
31
+ parallel_residual: false
32
+ post_attention_norm: false
33
+ post_mlp_norm: false
34
+ rope_adjustments: null
35
+ rope_base: 10000
36
+ rope_condense_ratio: 1
37
+ rope_indices: null
38
+ rope_local_base_freq: null
39
+ rotary_percentage: 1.0
40
+ scale_embeddings: false
41
+ shared_attention_norm: false
42
+ sliding_window_indices: null
43
+ sliding_window_size: null
44
+ vocab_size: 32000
out/pretrain/tinyllama_3_epoch/2408/final/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
out/pretrain/tinyllama_3_epoch/2408/final/tokenizer_config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "bos_token": {
5
+ "__type": "AddedToken",
6
+ "content": "<s>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "clean_up_tokenization_spaces": false,
13
+ "eos_token": {
14
+ "__type": "AddedToken",
15
+ "content": "</s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "legacy": false,
22
+ "model_max_length": 1000000000000000019884624838656,
23
+ "pad_token": null,
24
+ "padding_side": "right",
25
+ "sp_model_kwargs": {},
26
+ "tokenizer_class": "LlamaTokenizer",
27
+ "unk_token": {
28
+ "__type": "AddedToken",
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false
34
+ }
35
+ }
out/pretrain/tinyllama_3_epoch/2409/final/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
out/pretrain/tinyllama_3_epoch/2409/final/tokenizer_config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "bos_token": {
5
+ "__type": "AddedToken",
6
+ "content": "<s>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "clean_up_tokenization_spaces": false,
13
+ "eos_token": {
14
+ "__type": "AddedToken",
15
+ "content": "</s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "legacy": false,
22
+ "model_max_length": 1000000000000000019884624838656,
23
+ "pad_token": null,
24
+ "padding_side": "right",
25
+ "sp_model_kwargs": {},
26
+ "tokenizer_class": "LlamaTokenizer",
27
+ "unk_token": {
28
+ "__type": "AddedToken",
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false
34
+ }
35
+ }
out/pretrain/tinyllama_lr_plus/2501/final/config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LlamaForCausalLM"
4
+ ],
5
+ "bos_token_id": 1,
6
+ "eos_token_id": 2,
7
+ "hidden_act": "silu",
8
+ "hidden_size": 2048,
9
+ "initializer_range": 0.02,
10
+ "intermediate_size": 5632,
11
+ "max_position_embeddings": 2048,
12
+ "model_type": "llama",
13
+ "num_attention_heads": 32,
14
+ "num_hidden_layers": 22,
15
+ "num_key_value_heads": 4,
16
+ "pretraining_tp": 1,
17
+ "rms_norm_eps": 1e-05,
18
+ "rope_scaling": null,
19
+ "tie_word_embeddings": false,
20
+ "torch_dtype": "float32",
21
+ "transformers_version": "4.31.0.dev0",
22
+ "use_cache": true,
23
+ "vocab_size": 32000
24
+ }
out/pretrain/tinyllama_lr_plus/2502/final/config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LlamaForCausalLM"
4
+ ],
5
+ "bos_token_id": 1,
6
+ "eos_token_id": 2,
7
+ "hidden_act": "silu",
8
+ "hidden_size": 2048,
9
+ "initializer_range": 0.02,
10
+ "intermediate_size": 5632,
11
+ "max_position_embeddings": 2048,
12
+ "model_type": "llama",
13
+ "num_attention_heads": 32,
14
+ "num_hidden_layers": 22,
15
+ "num_key_value_heads": 4,
16
+ "pretraining_tp": 1,
17
+ "rms_norm_eps": 1e-05,
18
+ "rope_scaling": null,
19
+ "tie_word_embeddings": false,
20
+ "torch_dtype": "float32",
21
+ "transformers_version": "4.31.0.dev0",
22
+ "use_cache": true,
23
+ "vocab_size": 32000
24
+ }
out/pretrain/tinyllama_lr_plus/2502/final/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 1,
3
+ "eos_token_id": 2,
4
+ "pad_token_id": 0,
5
+ "max_length": 2048,
6
+ "transformers_version": "4.31.0.dev0"
7
+ }
out/pretrain/tinyllama_lr_plus/2502/final/hyperparameters.yaml ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model_name: tiny-llama-1.1b
2
+ out_dir: out/pretrain/tinyllama_lr_plus/2502
3
+ precision: bf16-mixed
4
+ initial_checkpoint_dir: out/pretrain/tinyllama_lr_plus/2501/final
5
+ resume: false
6
+ data:
7
+ class_path: litgpt.data.Arxiv
8
+ init_args:
9
+ data_path: litgpt/data/arxiv
10
+ seed: 42
11
+ num_workers: 0
12
+ use_starcoder: true
13
+ ppl: false
14
+ data_dir: litgpt/data/arxiv_tinyllama_tokenized/2502
15
+ train:
16
+ save_interval: 9999
17
+ log_interval: 1
18
+ global_batch_size: 512
19
+ micro_batch_size: 4
20
+ lr_warmup_steps: 20
21
+ max_tokens: 330301440
22
+ max_seq_length: 2048
23
+ max_norm: 1.0
24
+ min_lr: 0.0008
25
+ eval:
26
+ interval: 50
27
+ max_iters: 200
28
+ initial_validation: true
29
+ final_validation: true
30
+ evaluate_example: first
31
+ log: {}
32
+ optimizer:
33
+ class_path: torch.optim.AdamW
34
+ init_args:
35
+ lr: 0.0008
36
+ weight_decay: 0.1
37
+ betas:
38
+ - 0.9
39
+ - 0.95
40
+ devices: auto
41
+ num_nodes: 1
42
+ tokenizer_dir: checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
43
+ logger_name: tensorboard
44
+ seed: 42
out/pretrain/tinyllama_lr_plus/2502/final/model_config.yaml ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ attention_logit_softcapping: null
2
+ attention_scores_scalar: null
3
+ attn_bias: false
4
+ bias: false
5
+ block_size: 2048
6
+ final_logit_softcapping: null
7
+ gelu_approximate: none
8
+ head_size: 64
9
+ hf_config:
10
+ name: TinyLlama-1.1B-intermediate-step-1431k-3T
11
+ org: TinyLlama
12
+ intermediate_size: 5632
13
+ lm_head_bias: false
14
+ mlp_class_name: LLaMAMLP
15
+ moe_intermediate_size: null
16
+ n_embd: 2048
17
+ n_expert: 0
18
+ n_expert_per_token: 0
19
+ n_head: 32
20
+ n_layer: 22
21
+ n_query_groups: 4
22
+ name: tiny-llama-1.1b
23
+ norm_1: true
24
+ norm_2: true
25
+ norm_class_name: RMSNorm
26
+ norm_eps: 1.0e-05
27
+ norm_qk: false
28
+ norm_qk_type: default
29
+ padded_vocab_size: 32000
30
+ padding_multiple: 64
31
+ parallel_residual: false
32
+ post_attention_norm: false
33
+ post_mlp_norm: false
34
+ rope_adjustments: null
35
+ rope_base: 10000
36
+ rope_condense_ratio: 1
37
+ rope_indices: null
38
+ rope_local_base_freq: null
39
+ rotary_percentage: 1.0
40
+ scale_embeddings: false
41
+ shared_attention_norm: false
42
+ sliding_window_indices: null
43
+ sliding_window_size: null
44
+ vocab_size: 32000
out/pretrain/tinyllama_lr_plus/2502/final/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
out/pretrain/tinyllama_lr_plus/2502/final/tokenizer_config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "bos_token": {
5
+ "__type": "AddedToken",
6
+ "content": "<s>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "clean_up_tokenization_spaces": false,
13
+ "eos_token": {
14
+ "__type": "AddedToken",
15
+ "content": "</s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "legacy": false,
22
+ "model_max_length": 1000000000000000019884624838656,
23
+ "pad_token": null,
24
+ "padding_side": "right",
25
+ "sp_model_kwargs": {},
26
+ "tokenizer_class": "LlamaTokenizer",
27
+ "unk_token": {
28
+ "__type": "AddedToken",
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false
34
+ }
35
+ }
out/pretrain/tinyllama_lr_plus/2503/final/config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LlamaForCausalLM"
4
+ ],
5
+ "bos_token_id": 1,
6
+ "eos_token_id": 2,
7
+ "hidden_act": "silu",
8
+ "hidden_size": 2048,
9
+ "initializer_range": 0.02,
10
+ "intermediate_size": 5632,
11
+ "max_position_embeddings": 2048,
12
+ "model_type": "llama",
13
+ "num_attention_heads": 32,
14
+ "num_hidden_layers": 22,
15
+ "num_key_value_heads": 4,
16
+ "pretraining_tp": 1,
17
+ "rms_norm_eps": 1e-05,
18
+ "rope_scaling": null,
19
+ "tie_word_embeddings": false,
20
+ "torch_dtype": "float32",
21
+ "transformers_version": "4.31.0.dev0",
22
+ "use_cache": true,
23
+ "vocab_size": 32000
24
+ }
out/pretrain/tinyllama_lr_plus/2503/final/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 1,
3
+ "eos_token_id": 2,
4
+ "pad_token_id": 0,
5
+ "max_length": 2048,
6
+ "transformers_version": "4.31.0.dev0"
7
+ }
out/pretrain/tinyllama_lr_plus/2503/final/hyperparameters.yaml ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model_name: tiny-llama-1.1b
2
+ out_dir: out/pretrain/tinyllama_lr_plus/2503
3
+ precision: bf16-mixed
4
+ initial_checkpoint_dir: out/pretrain/tinyllama_lr_plus/2502/final
5
+ resume: false
6
+ data:
7
+ class_path: litgpt.data.Arxiv
8
+ init_args:
9
+ data_path: litgpt/data/arxiv
10
+ seed: 42
11
+ num_workers: 0
12
+ use_starcoder: true
13
+ ppl: false
14
+ data_dir: litgpt/data/arxiv_tinyllama_tokenized/2503
15
+ train:
16
+ save_interval: 9999
17
+ log_interval: 1
18
+ global_batch_size: 512
19
+ micro_batch_size: 4
20
+ lr_warmup_steps: 20
21
+ max_tokens: 394264576
22
+ max_seq_length: 2048
23
+ max_norm: 1.0
24
+ min_lr: 0.0008
25
+ eval:
26
+ interval: 50
27
+ max_iters: 200
28
+ initial_validation: true
29
+ final_validation: true
30
+ evaluate_example: first
31
+ log: {}
32
+ optimizer:
33
+ class_path: torch.optim.AdamW
34
+ init_args:
35
+ lr: 0.0008
36
+ weight_decay: 0.1
37
+ betas:
38
+ - 0.9
39
+ - 0.95
40
+ devices: auto
41
+ num_nodes: 1
42
+ tokenizer_dir: checkpoints/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
43
+ logger_name: tensorboard
44
+ seed: 42
out/pretrain/tinyllama_lr_plus/2503/final/model_config.yaml ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ attention_logit_softcapping: null
2
+ attention_scores_scalar: null
3
+ attn_bias: false
4
+ bias: false
5
+ block_size: 2048
6
+ final_logit_softcapping: null
7
+ gelu_approximate: none
8
+ head_size: 64
9
+ hf_config:
10
+ name: TinyLlama-1.1B-intermediate-step-1431k-3T
11
+ org: TinyLlama
12
+ intermediate_size: 5632
13
+ lm_head_bias: false
14
+ mlp_class_name: LLaMAMLP
15
+ moe_intermediate_size: null
16
+ n_embd: 2048
17
+ n_expert: 0
18
+ n_expert_per_token: 0
19
+ n_head: 32
20
+ n_layer: 22
21
+ n_query_groups: 4
22
+ name: tiny-llama-1.1b
23
+ norm_1: true
24
+ norm_2: true
25
+ norm_class_name: RMSNorm
26
+ norm_eps: 1.0e-05
27
+ norm_qk: false
28
+ norm_qk_type: default
29
+ padded_vocab_size: 32000
30
+ padding_multiple: 64
31
+ parallel_residual: false
32
+ post_attention_norm: false
33
+ post_mlp_norm: false
34
+ rope_adjustments: null
35
+ rope_base: 10000
36
+ rope_condense_ratio: 1
37
+ rope_indices: null
38
+ rope_local_base_freq: null
39
+ rotary_percentage: 1.0
40
+ scale_embeddings: false
41
+ shared_attention_norm: false
42
+ sliding_window_indices: null
43
+ sliding_window_size: null
44
+ vocab_size: 32000
out/pretrain/tinyllama_lr_plus/2503/final/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
out/pretrain/tinyllama_lr_plus/2503/final/tokenizer_config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "bos_token": {
5
+ "__type": "AddedToken",
6
+ "content": "<s>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "clean_up_tokenization_spaces": false,
13
+ "eos_token": {
14
+ "__type": "AddedToken",
15
+ "content": "</s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "legacy": false,
22
+ "model_max_length": 1000000000000000019884624838656,
23
+ "pad_token": null,
24
+ "padding_side": "right",
25
+ "sp_model_kwargs": {},
26
+ "tokenizer_class": "LlamaTokenizer",
27
+ "unk_token": {
28
+ "__type": "AddedToken",
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false
34
+ }
35
+ }
out/pretrain/tinyllama_lr_plus/2504/final/config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LlamaForCausalLM"
4
+ ],
5
+ "bos_token_id": 1,
6
+ "eos_token_id": 2,
7
+ "hidden_act": "silu",
8
+ "hidden_size": 2048,
9
+ "initializer_range": 0.02,
10
+ "intermediate_size": 5632,
11
+ "max_position_embeddings": 2048,
12
+ "model_type": "llama",
13
+ "num_attention_heads": 32,
14
+ "num_hidden_layers": 22,
15
+ "num_key_value_heads": 4,
16
+ "pretraining_tp": 1,
17
+ "rms_norm_eps": 1e-05,
18
+ "rope_scaling": null,
19
+ "tie_word_embeddings": false,
20
+ "torch_dtype": "float32",
21
+ "transformers_version": "4.31.0.dev0",
22
+ "use_cache": true,
23
+ "vocab_size": 32000
24
+ }
out/pretrain/tinyllama_lr_plus/2504/final/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 1,
3
+ "eos_token_id": 2,
4
+ "pad_token_id": 0,
5
+ "max_length": 2048,
6
+ "transformers_version": "4.31.0.dev0"
7
+ }