melihcatal commited on
Commit
07662b8
·
verified ·
1 Parent(s): 1c604b3

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. starcoder2-7b/base/adapter/README.md +207 -0
  2. starcoder2-7b/base/adapter/adapter_config.json +46 -0
  3. starcoder2-7b/base/adapter/adapter_model.safetensors +3 -0
  4. starcoder2-7b/base/audit_results.json +137 -0
  5. starcoder2-7b/base/audit_scores.npz +3 -0
  6. starcoder2-7b/base/canary_meta.json +0 -0
  7. starcoder2-7b/base/codecarbon.csv +2 -0
  8. starcoder2-7b/base/epochs/epoch_001/adapter/README.md +207 -0
  9. starcoder2-7b/base/epochs/epoch_001/adapter/adapter_config.json +46 -0
  10. starcoder2-7b/base/epochs/epoch_001/adapter/adapter_model.safetensors +3 -0
  11. starcoder2-7b/base/epochs/epoch_001/audit_results.json +137 -0
  12. starcoder2-7b/base/epochs/epoch_001/audit_scores.npz +3 -0
  13. starcoder2-7b/base/epochs/epoch_002/adapter/README.md +207 -0
  14. starcoder2-7b/base/epochs/epoch_002/adapter/adapter_config.json +46 -0
  15. starcoder2-7b/base/epochs/epoch_002/adapter/adapter_model.safetensors +3 -0
  16. starcoder2-7b/base/epochs/epoch_002/audit_results.json +137 -0
  17. starcoder2-7b/base/epochs/epoch_002/audit_scores.npz +3 -0
  18. starcoder2-7b/base/metrics.jsonl +55 -0
  19. starcoder2-7b/base/pretrain_lm_head.pt +3 -0
  20. starcoder2-7b/base/resolved_config.yaml +100 -0
  21. starcoder2-7b/base/scalars.csv +613 -0
  22. starcoder2-7b/base/summary.json +71 -0
  23. starcoder2-7b/base/tensorboard/events.out.tfevents.1774090870.364c4f8de9dd.6145.0 +3 -0
  24. starcoder2-7b/base/tokenizer/tokenizer.json +0 -0
  25. starcoder2-7b/base/tokenizer/tokenizer_config.json +516 -0
  26. starcoder2-7b/base/train.log +49 -0
  27. starcoder2-7b/dp3/adapter/README.md +207 -0
  28. starcoder2-7b/dp3/adapter/adapter_config.json +46 -0
  29. starcoder2-7b/dp3/adapter/adapter_model.safetensors +3 -0
  30. starcoder2-7b/dp3/audit_results.json +137 -0
  31. starcoder2-7b/dp3/audit_scores.npz +3 -0
  32. starcoder2-7b/dp3/canary_meta.json +0 -0
  33. starcoder2-7b/dp3/codecarbon.csv +2 -0
  34. starcoder2-7b/dp3/epochs/epoch_001/adapter/README.md +207 -0
  35. starcoder2-7b/dp3/epochs/epoch_001/adapter/adapter_config.json +46 -0
  36. starcoder2-7b/dp3/epochs/epoch_001/adapter/adapter_model.safetensors +3 -0
  37. starcoder2-7b/dp3/epochs/epoch_001/audit_results.json +137 -0
  38. starcoder2-7b/dp3/epochs/epoch_001/audit_scores.npz +3 -0
  39. starcoder2-7b/dp3/epochs/epoch_002/adapter/README.md +207 -0
  40. starcoder2-7b/dp3/epochs/epoch_002/adapter/adapter_config.json +46 -0
  41. starcoder2-7b/dp3/epochs/epoch_002/adapter/adapter_model.safetensors +3 -0
  42. starcoder2-7b/dp3/epochs/epoch_002/audit_results.json +137 -0
  43. starcoder2-7b/dp3/epochs/epoch_002/audit_scores.npz +3 -0
  44. starcoder2-7b/dp3/metrics.jsonl +30 -0
  45. starcoder2-7b/dp3/pretrain_lm_head.pt +3 -0
  46. starcoder2-7b/dp3/resolved_config.yaml +101 -0
  47. starcoder2-7b/dp3/scalars.csv +386 -0
  48. starcoder2-7b/dp3/summary.json +72 -0
  49. starcoder2-7b/dp3/tensorboard/events.out.tfevents.1774096221.364c4f8de9dd.12837.0 +3 -0
  50. starcoder2-7b/dp3/tokenizer/tokenizer.json +0 -0
starcoder2-7b/base/adapter/README.md ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: bigcode/starcoder2-7b
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:bigcode/starcoder2-7b
7
+ - lora
8
+ - transformers
9
+ ---
10
+
11
+ # Model Card for Model ID
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+
16
+
17
+ ## Model Details
18
+
19
+ ### Model Description
20
+
21
+ <!-- Provide a longer summary of what this model is. -->
22
+
23
+
24
+
25
+ - **Developed by:** [More Information Needed]
26
+ - **Funded by [optional]:** [More Information Needed]
27
+ - **Shared by [optional]:** [More Information Needed]
28
+ - **Model type:** [More Information Needed]
29
+ - **Language(s) (NLP):** [More Information Needed]
30
+ - **License:** [More Information Needed]
31
+ - **Finetuned from model [optional]:** [More Information Needed]
32
+
33
+ ### Model Sources [optional]
34
+
35
+ <!-- Provide the basic links for the model. -->
36
+
37
+ - **Repository:** [More Information Needed]
38
+ - **Paper [optional]:** [More Information Needed]
39
+ - **Demo [optional]:** [More Information Needed]
40
+
41
+ ## Uses
42
+
43
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
+
45
+ ### Direct Use
46
+
47
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
48
+
49
+ [More Information Needed]
50
+
51
+ ### Downstream Use [optional]
52
+
53
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
54
+
55
+ [More Information Needed]
56
+
57
+ ### Out-of-Scope Use
58
+
59
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
+
61
+ [More Information Needed]
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ [More Information Needed]
68
+
69
+ ### Recommendations
70
+
71
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
72
+
73
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
74
+
75
+ ## How to Get Started with the Model
76
+
77
+ Use the code below to get started with the model.
78
+
79
+ [More Information Needed]
80
+
81
+ ## Training Details
82
+
83
+ ### Training Data
84
+
85
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
86
+
87
+ [More Information Needed]
88
+
89
+ ### Training Procedure
90
+
91
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
92
+
93
+ #### Preprocessing [optional]
94
+
95
+ [More Information Needed]
96
+
97
+
98
+ #### Training Hyperparameters
99
+
100
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
101
+
102
+ #### Speeds, Sizes, Times [optional]
103
+
104
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
105
+
106
+ [More Information Needed]
107
+
108
+ ## Evaluation
109
+
110
+ <!-- This section describes the evaluation protocols and provides the results. -->
111
+
112
+ ### Testing Data, Factors & Metrics
113
+
114
+ #### Testing Data
115
+
116
+ <!-- This should link to a Dataset Card if possible. -->
117
+
118
+ [More Information Needed]
119
+
120
+ #### Factors
121
+
122
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
123
+
124
+ [More Information Needed]
125
+
126
+ #### Metrics
127
+
128
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
129
+
130
+ [More Information Needed]
131
+
132
+ ### Results
133
+
134
+ [More Information Needed]
135
+
136
+ #### Summary
137
+
138
+
139
+
140
+ ## Model Examination [optional]
141
+
142
+ <!-- Relevant interpretability work for the model goes here -->
143
+
144
+ [More Information Needed]
145
+
146
+ ## Environmental Impact
147
+
148
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
149
+
150
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
151
+
152
+ - **Hardware Type:** [More Information Needed]
153
+ - **Hours used:** [More Information Needed]
154
+ - **Cloud Provider:** [More Information Needed]
155
+ - **Compute Region:** [More Information Needed]
156
+ - **Carbon Emitted:** [More Information Needed]
157
+
158
+ ## Technical Specifications [optional]
159
+
160
+ ### Model Architecture and Objective
161
+
162
+ [More Information Needed]
163
+
164
+ ### Compute Infrastructure
165
+
166
+ [More Information Needed]
167
+
168
+ #### Hardware
169
+
170
+ [More Information Needed]
171
+
172
+ #### Software
173
+
174
+ [More Information Needed]
175
+
176
+ ## Citation [optional]
177
+
178
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
179
+
180
+ **BibTeX:**
181
+
182
+ [More Information Needed]
183
+
184
+ **APA:**
185
+
186
+ [More Information Needed]
187
+
188
+ ## Glossary [optional]
189
+
190
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
191
+
192
+ [More Information Needed]
193
+
194
+ ## More Information [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Authors [optional]
199
+
200
+ [More Information Needed]
201
+
202
+ ## Model Card Contact
203
+
204
+ [More Information Needed]
205
+ ### Framework versions
206
+
207
+ - PEFT 0.18.1
starcoder2-7b/base/adapter/adapter_config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": null,
6
+ "base_model_name_or_path": "bigcode/starcoder2-7b",
7
+ "bias": "none",
8
+ "corda_config": null,
9
+ "ensure_weight_tying": true,
10
+ "eva_config": null,
11
+ "exclude_modules": null,
12
+ "fan_in_fan_out": false,
13
+ "inference_mode": true,
14
+ "init_lora_weights": true,
15
+ "layer_replication": null,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "loftq_config": {},
19
+ "lora_alpha": 32,
20
+ "lora_bias": false,
21
+ "lora_dropout": 0.05,
22
+ "megatron_config": null,
23
+ "megatron_core": "megatron.core",
24
+ "modules_to_save": [
25
+ "lm_head",
26
+ "embed_tokens"
27
+ ],
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 16,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "v_proj",
36
+ "k_proj",
37
+ "q_proj",
38
+ "o_proj"
39
+ ],
40
+ "target_parameters": null,
41
+ "task_type": "CAUSAL_LM",
42
+ "trainable_token_indices": null,
43
+ "use_dora": false,
44
+ "use_qalora": false,
45
+ "use_rslora": false
46
+ }
starcoder2-7b/base/adapter/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8722c41c3297c10da9786673bd5f7f3556116c882e76e37035eedd5ae393938
3
+ size 2804312360
starcoder2-7b/base/audit_results.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "delta": 1e-05,
3
+ "num_canaries": 500,
4
+ "num_members": 250,
5
+ "paper_guess_fraction": 0.2,
6
+ "paper_guess_steps": 20,
7
+ "loss": {
8
+ "auc": 1.0,
9
+ "empirical_epsilon": {
10
+ "0.05": 3.4791953936219215,
11
+ "0.01": 3.023197554051876
12
+ },
13
+ "empirical_epsilon_details": {
14
+ "0.05": {
15
+ "epsilon": 3.4791953936219215,
16
+ "num_guesses": 100,
17
+ "correct_guesses": 100,
18
+ "candidate_num_guesses": [
19
+ 5,
20
+ 10,
21
+ 15,
22
+ 20,
23
+ 25,
24
+ 30,
25
+ 35,
26
+ 40,
27
+ 45,
28
+ 50,
29
+ 55,
30
+ 60,
31
+ 65,
32
+ 70,
33
+ 75,
34
+ 80,
35
+ 85,
36
+ 90,
37
+ 95,
38
+ 100
39
+ ],
40
+ "direction": "lower"
41
+ },
42
+ "0.01": {
43
+ "epsilon": 3.023197554051876,
44
+ "num_guesses": 100,
45
+ "correct_guesses": 100,
46
+ "candidate_num_guesses": [
47
+ 5,
48
+ 10,
49
+ 15,
50
+ 20,
51
+ 25,
52
+ 30,
53
+ 35,
54
+ 40,
55
+ 45,
56
+ 50,
57
+ 55,
58
+ 60,
59
+ 65,
60
+ 70,
61
+ 75,
62
+ 80,
63
+ 85,
64
+ 90,
65
+ 95,
66
+ 100
67
+ ],
68
+ "direction": "lower"
69
+ }
70
+ }
71
+ },
72
+ "embedding": {
73
+ "auc": 0.916224,
74
+ "empirical_epsilon": {
75
+ "0.05": 3.4791953936219215,
76
+ "0.01": 3.023197554051876
77
+ },
78
+ "empirical_epsilon_details": {
79
+ "0.05": {
80
+ "epsilon": 3.4791953936219215,
81
+ "num_guesses": 100,
82
+ "correct_guesses": 100,
83
+ "candidate_num_guesses": [
84
+ 5,
85
+ 10,
86
+ 15,
87
+ 20,
88
+ 25,
89
+ 30,
90
+ 35,
91
+ 40,
92
+ 45,
93
+ 50,
94
+ 55,
95
+ 60,
96
+ 65,
97
+ 70,
98
+ 75,
99
+ 80,
100
+ 85,
101
+ 90,
102
+ 95,
103
+ 100
104
+ ],
105
+ "direction": "lower"
106
+ },
107
+ "0.01": {
108
+ "epsilon": 3.023197554051876,
109
+ "num_guesses": 100,
110
+ "correct_guesses": 100,
111
+ "candidate_num_guesses": [
112
+ 5,
113
+ 10,
114
+ 15,
115
+ 20,
116
+ 25,
117
+ 30,
118
+ 35,
119
+ 40,
120
+ 45,
121
+ 50,
122
+ 55,
123
+ 60,
124
+ 65,
125
+ 70,
126
+ 75,
127
+ 80,
128
+ 85,
129
+ 90,
130
+ 95,
131
+ 100
132
+ ],
133
+ "direction": "lower"
134
+ }
135
+ }
136
+ }
137
+ }
starcoder2-7b/base/audit_scores.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df7a9ad49c78c2b7e30773bfe7d47ce312de900740e940d72ac0b3117dff09bf
3
+ size 12784
starcoder2-7b/base/canary_meta.json ADDED
The diff for this file is too large to render. See raw diff
 
starcoder2-7b/base/codecarbon.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ timestamp,project_name,run_id,experiment_id,duration,emissions,emissions_rate,cpu_power,gpu_power,ram_power,cpu_energy,gpu_energy,ram_energy,energy_consumed,water_consumed,country_name,country_iso_code,region,cloud_provider,cloud_region,os,python_version,codecarbon_version,cpu_count,cpu_model,gpu_count,gpu_model,longitude,latitude,ram_total_size,tracking_mode,cpu_utilization_percent,gpu_utilization_percent,ram_utilization_percent,ram_used_gb,on_cloud,pue,wue
2
+ 2026-03-21T12:05:04,codedp-starcoder2-7b-cpt-base,499ee9b6-ab5e-4714-be7f-0b592eb40909,5b0fa12a-3dd7-45bb-9766-cc326314d9f1,3820.65162669681,0.5568149811716148,0.0001457382236267942,179.3439937196561,2316.658494927477,70.0,0.18368298514637418,2.457235179675422,0.07169108754482326,2.7126092523666183,0.0,United States,USA,california,,,Linux-5.15.0-157-generic-x86_64-with-glibc2.39,3.12.13,3.2.5,224,Intel(R) Xeon(R) Platinum 8480C,4,4 x NVIDIA H200,-121.9552,37.3541,2015.5625190734863,machine,2.2497114375655825,96.45330535152151,2.3057974816369358,46.78091819539025,N,1.0,0.0
starcoder2-7b/base/epochs/epoch_001/adapter/README.md ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: bigcode/starcoder2-7b
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:bigcode/starcoder2-7b
7
+ - lora
8
+ - transformers
9
+ ---
10
+
11
+ # Model Card for Model ID
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+
16
+
17
+ ## Model Details
18
+
19
+ ### Model Description
20
+
21
+ <!-- Provide a longer summary of what this model is. -->
22
+
23
+
24
+
25
+ - **Developed by:** [More Information Needed]
26
+ - **Funded by [optional]:** [More Information Needed]
27
+ - **Shared by [optional]:** [More Information Needed]
28
+ - **Model type:** [More Information Needed]
29
+ - **Language(s) (NLP):** [More Information Needed]
30
+ - **License:** [More Information Needed]
31
+ - **Finetuned from model [optional]:** [More Information Needed]
32
+
33
+ ### Model Sources [optional]
34
+
35
+ <!-- Provide the basic links for the model. -->
36
+
37
+ - **Repository:** [More Information Needed]
38
+ - **Paper [optional]:** [More Information Needed]
39
+ - **Demo [optional]:** [More Information Needed]
40
+
41
+ ## Uses
42
+
43
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
+
45
+ ### Direct Use
46
+
47
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
48
+
49
+ [More Information Needed]
50
+
51
+ ### Downstream Use [optional]
52
+
53
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
54
+
55
+ [More Information Needed]
56
+
57
+ ### Out-of-Scope Use
58
+
59
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
+
61
+ [More Information Needed]
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ [More Information Needed]
68
+
69
+ ### Recommendations
70
+
71
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
72
+
73
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
74
+
75
+ ## How to Get Started with the Model
76
+
77
+ Use the code below to get started with the model.
78
+
79
+ [More Information Needed]
80
+
81
+ ## Training Details
82
+
83
+ ### Training Data
84
+
85
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
86
+
87
+ [More Information Needed]
88
+
89
+ ### Training Procedure
90
+
91
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
92
+
93
+ #### Preprocessing [optional]
94
+
95
+ [More Information Needed]
96
+
97
+
98
+ #### Training Hyperparameters
99
+
100
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
101
+
102
+ #### Speeds, Sizes, Times [optional]
103
+
104
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
105
+
106
+ [More Information Needed]
107
+
108
+ ## Evaluation
109
+
110
+ <!-- This section describes the evaluation protocols and provides the results. -->
111
+
112
+ ### Testing Data, Factors & Metrics
113
+
114
+ #### Testing Data
115
+
116
+ <!-- This should link to a Dataset Card if possible. -->
117
+
118
+ [More Information Needed]
119
+
120
+ #### Factors
121
+
122
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
123
+
124
+ [More Information Needed]
125
+
126
+ #### Metrics
127
+
128
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
129
+
130
+ [More Information Needed]
131
+
132
+ ### Results
133
+
134
+ [More Information Needed]
135
+
136
+ #### Summary
137
+
138
+
139
+
140
+ ## Model Examination [optional]
141
+
142
+ <!-- Relevant interpretability work for the model goes here -->
143
+
144
+ [More Information Needed]
145
+
146
+ ## Environmental Impact
147
+
148
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
149
+
150
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
151
+
152
+ - **Hardware Type:** [More Information Needed]
153
+ - **Hours used:** [More Information Needed]
154
+ - **Cloud Provider:** [More Information Needed]
155
+ - **Compute Region:** [More Information Needed]
156
+ - **Carbon Emitted:** [More Information Needed]
157
+
158
+ ## Technical Specifications [optional]
159
+
160
+ ### Model Architecture and Objective
161
+
162
+ [More Information Needed]
163
+
164
+ ### Compute Infrastructure
165
+
166
+ [More Information Needed]
167
+
168
+ #### Hardware
169
+
170
+ [More Information Needed]
171
+
172
+ #### Software
173
+
174
+ [More Information Needed]
175
+
176
+ ## Citation [optional]
177
+
178
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
179
+
180
+ **BibTeX:**
181
+
182
+ [More Information Needed]
183
+
184
+ **APA:**
185
+
186
+ [More Information Needed]
187
+
188
+ ## Glossary [optional]
189
+
190
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
191
+
192
+ [More Information Needed]
193
+
194
+ ## More Information [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Authors [optional]
199
+
200
+ [More Information Needed]
201
+
202
+ ## Model Card Contact
203
+
204
+ [More Information Needed]
205
+ ### Framework versions
206
+
207
+ - PEFT 0.18.1
starcoder2-7b/base/epochs/epoch_001/adapter/adapter_config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": null,
6
+ "base_model_name_or_path": "bigcode/starcoder2-7b",
7
+ "bias": "none",
8
+ "corda_config": null,
9
+ "ensure_weight_tying": true,
10
+ "eva_config": null,
11
+ "exclude_modules": null,
12
+ "fan_in_fan_out": false,
13
+ "inference_mode": true,
14
+ "init_lora_weights": true,
15
+ "layer_replication": null,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "loftq_config": {},
19
+ "lora_alpha": 32,
20
+ "lora_bias": false,
21
+ "lora_dropout": 0.05,
22
+ "megatron_config": null,
23
+ "megatron_core": "megatron.core",
24
+ "modules_to_save": [
25
+ "lm_head",
26
+ "embed_tokens"
27
+ ],
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 16,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "v_proj",
36
+ "k_proj",
37
+ "q_proj",
38
+ "o_proj"
39
+ ],
40
+ "target_parameters": null,
41
+ "task_type": "CAUSAL_LM",
42
+ "trainable_token_indices": null,
43
+ "use_dora": false,
44
+ "use_qalora": false,
45
+ "use_rslora": false
46
+ }
starcoder2-7b/base/epochs/epoch_001/adapter/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ab7c236acdfacb363251d6f1c092cab8b23cfee5381584cd586426a1ecf16d0
3
+ size 2804312360
starcoder2-7b/base/epochs/epoch_001/audit_results.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "delta": 1e-05,
3
+ "num_canaries": 500,
4
+ "num_members": 250,
5
+ "paper_guess_fraction": 0.2,
6
+ "paper_guess_steps": 20,
7
+ "loss": {
8
+ "auc": 0.998072,
9
+ "empirical_epsilon": {
10
+ "0.05": 3.4791953936219215,
11
+ "0.01": 3.023197554051876
12
+ },
13
+ "empirical_epsilon_details": {
14
+ "0.05": {
15
+ "epsilon": 3.4791953936219215,
16
+ "num_guesses": 100,
17
+ "correct_guesses": 100,
18
+ "candidate_num_guesses": [
19
+ 5,
20
+ 10,
21
+ 15,
22
+ 20,
23
+ 25,
24
+ 30,
25
+ 35,
26
+ 40,
27
+ 45,
28
+ 50,
29
+ 55,
30
+ 60,
31
+ 65,
32
+ 70,
33
+ 75,
34
+ 80,
35
+ 85,
36
+ 90,
37
+ 95,
38
+ 100
39
+ ],
40
+ "direction": "lower"
41
+ },
42
+ "0.01": {
43
+ "epsilon": 3.023197554051876,
44
+ "num_guesses": 100,
45
+ "correct_guesses": 100,
46
+ "candidate_num_guesses": [
47
+ 5,
48
+ 10,
49
+ 15,
50
+ 20,
51
+ 25,
52
+ 30,
53
+ 35,
54
+ 40,
55
+ 45,
56
+ 50,
57
+ 55,
58
+ 60,
59
+ 65,
60
+ 70,
61
+ 75,
62
+ 80,
63
+ 85,
64
+ 90,
65
+ 95,
66
+ 100
67
+ ],
68
+ "direction": "lower"
69
+ }
70
+ }
71
+ },
72
+ "embedding": {
73
+ "auc": 0.96,
74
+ "empirical_epsilon": {
75
+ "0.05": 3.4791953936219215,
76
+ "0.01": 3.023197554051876
77
+ },
78
+ "empirical_epsilon_details": {
79
+ "0.05": {
80
+ "epsilon": 3.4791953936219215,
81
+ "num_guesses": 100,
82
+ "correct_guesses": 100,
83
+ "candidate_num_guesses": [
84
+ 5,
85
+ 10,
86
+ 15,
87
+ 20,
88
+ 25,
89
+ 30,
90
+ 35,
91
+ 40,
92
+ 45,
93
+ 50,
94
+ 55,
95
+ 60,
96
+ 65,
97
+ 70,
98
+ 75,
99
+ 80,
100
+ 85,
101
+ 90,
102
+ 95,
103
+ 100
104
+ ],
105
+ "direction": "lower"
106
+ },
107
+ "0.01": {
108
+ "epsilon": 3.023197554051876,
109
+ "num_guesses": 100,
110
+ "correct_guesses": 100,
111
+ "candidate_num_guesses": [
112
+ 5,
113
+ 10,
114
+ 15,
115
+ 20,
116
+ 25,
117
+ 30,
118
+ 35,
119
+ 40,
120
+ 45,
121
+ 50,
122
+ 55,
123
+ 60,
124
+ 65,
125
+ 70,
126
+ 75,
127
+ 80,
128
+ 85,
129
+ 90,
130
+ 95,
131
+ 100
132
+ ],
133
+ "direction": "lower"
134
+ }
135
+ }
136
+ }
137
+ }
starcoder2-7b/base/epochs/epoch_001/audit_scores.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad12d1a5732cbbe84adb0a8ad0d9063e66354e84483f65f8d0ad9cbad178b42f
3
+ size 12784
starcoder2-7b/base/epochs/epoch_002/adapter/README.md ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: bigcode/starcoder2-7b
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:bigcode/starcoder2-7b
7
+ - lora
8
+ - transformers
9
+ ---
10
+
11
+ # Model Card for Model ID
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+
16
+
17
+ ## Model Details
18
+
19
+ ### Model Description
20
+
21
+ <!-- Provide a longer summary of what this model is. -->
22
+
23
+
24
+
25
+ - **Developed by:** [More Information Needed]
26
+ - **Funded by [optional]:** [More Information Needed]
27
+ - **Shared by [optional]:** [More Information Needed]
28
+ - **Model type:** [More Information Needed]
29
+ - **Language(s) (NLP):** [More Information Needed]
30
+ - **License:** [More Information Needed]
31
+ - **Finetuned from model [optional]:** [More Information Needed]
32
+
33
+ ### Model Sources [optional]
34
+
35
+ <!-- Provide the basic links for the model. -->
36
+
37
+ - **Repository:** [More Information Needed]
38
+ - **Paper [optional]:** [More Information Needed]
39
+ - **Demo [optional]:** [More Information Needed]
40
+
41
+ ## Uses
42
+
43
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
+
45
+ ### Direct Use
46
+
47
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
48
+
49
+ [More Information Needed]
50
+
51
+ ### Downstream Use [optional]
52
+
53
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
54
+
55
+ [More Information Needed]
56
+
57
+ ### Out-of-Scope Use
58
+
59
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
+
61
+ [More Information Needed]
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ [More Information Needed]
68
+
69
+ ### Recommendations
70
+
71
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
72
+
73
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
74
+
75
+ ## How to Get Started with the Model
76
+
77
+ Use the code below to get started with the model.
78
+
79
+ [More Information Needed]
80
+
81
+ ## Training Details
82
+
83
+ ### Training Data
84
+
85
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
86
+
87
+ [More Information Needed]
88
+
89
+ ### Training Procedure
90
+
91
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
92
+
93
+ #### Preprocessing [optional]
94
+
95
+ [More Information Needed]
96
+
97
+
98
+ #### Training Hyperparameters
99
+
100
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
101
+
102
+ #### Speeds, Sizes, Times [optional]
103
+
104
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
105
+
106
+ [More Information Needed]
107
+
108
+ ## Evaluation
109
+
110
+ <!-- This section describes the evaluation protocols and provides the results. -->
111
+
112
+ ### Testing Data, Factors & Metrics
113
+
114
+ #### Testing Data
115
+
116
+ <!-- This should link to a Dataset Card if possible. -->
117
+
118
+ [More Information Needed]
119
+
120
+ #### Factors
121
+
122
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
123
+
124
+ [More Information Needed]
125
+
126
+ #### Metrics
127
+
128
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
129
+
130
+ [More Information Needed]
131
+
132
+ ### Results
133
+
134
+ [More Information Needed]
135
+
136
+ #### Summary
137
+
138
+
139
+
140
+ ## Model Examination [optional]
141
+
142
+ <!-- Relevant interpretability work for the model goes here -->
143
+
144
+ [More Information Needed]
145
+
146
+ ## Environmental Impact
147
+
148
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
149
+
150
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
151
+
152
+ - **Hardware Type:** [More Information Needed]
153
+ - **Hours used:** [More Information Needed]
154
+ - **Cloud Provider:** [More Information Needed]
155
+ - **Compute Region:** [More Information Needed]
156
+ - **Carbon Emitted:** [More Information Needed]
157
+
158
+ ## Technical Specifications [optional]
159
+
160
+ ### Model Architecture and Objective
161
+
162
+ [More Information Needed]
163
+
164
+ ### Compute Infrastructure
165
+
166
+ [More Information Needed]
167
+
168
+ #### Hardware
169
+
170
+ [More Information Needed]
171
+
172
+ #### Software
173
+
174
+ [More Information Needed]
175
+
176
+ ## Citation [optional]
177
+
178
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
179
+
180
+ **BibTeX:**
181
+
182
+ [More Information Needed]
183
+
184
+ **APA:**
185
+
186
+ [More Information Needed]
187
+
188
+ ## Glossary [optional]
189
+
190
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
191
+
192
+ [More Information Needed]
193
+
194
+ ## More Information [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Authors [optional]
199
+
200
+ [More Information Needed]
201
+
202
+ ## Model Card Contact
203
+
204
+ [More Information Needed]
205
+ ### Framework versions
206
+
207
+ - PEFT 0.18.1
starcoder2-7b/base/epochs/epoch_002/adapter/adapter_config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": null,
6
+ "base_model_name_or_path": "bigcode/starcoder2-7b",
7
+ "bias": "none",
8
+ "corda_config": null,
9
+ "ensure_weight_tying": true,
10
+ "eva_config": null,
11
+ "exclude_modules": null,
12
+ "fan_in_fan_out": false,
13
+ "inference_mode": true,
14
+ "init_lora_weights": true,
15
+ "layer_replication": null,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "loftq_config": {},
19
+ "lora_alpha": 32,
20
+ "lora_bias": false,
21
+ "lora_dropout": 0.05,
22
+ "megatron_config": null,
23
+ "megatron_core": "megatron.core",
24
+ "modules_to_save": [
25
+ "lm_head",
26
+ "embed_tokens"
27
+ ],
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 16,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "v_proj",
36
+ "k_proj",
37
+ "q_proj",
38
+ "o_proj"
39
+ ],
40
+ "target_parameters": null,
41
+ "task_type": "CAUSAL_LM",
42
+ "trainable_token_indices": null,
43
+ "use_dora": false,
44
+ "use_qalora": false,
45
+ "use_rslora": false
46
+ }
starcoder2-7b/base/epochs/epoch_002/adapter/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8722c41c3297c10da9786673bd5f7f3556116c882e76e37035eedd5ae393938
3
+ size 2804312360
starcoder2-7b/base/epochs/epoch_002/audit_results.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "delta": 1e-05,
3
+ "num_canaries": 500,
4
+ "num_members": 250,
5
+ "paper_guess_fraction": 0.2,
6
+ "paper_guess_steps": 20,
7
+ "loss": {
8
+ "auc": 1.0,
9
+ "empirical_epsilon": {
10
+ "0.05": 3.4791953936219215,
11
+ "0.01": 3.023197554051876
12
+ },
13
+ "empirical_epsilon_details": {
14
+ "0.05": {
15
+ "epsilon": 3.4791953936219215,
16
+ "num_guesses": 100,
17
+ "correct_guesses": 100,
18
+ "candidate_num_guesses": [
19
+ 5,
20
+ 10,
21
+ 15,
22
+ 20,
23
+ 25,
24
+ 30,
25
+ 35,
26
+ 40,
27
+ 45,
28
+ 50,
29
+ 55,
30
+ 60,
31
+ 65,
32
+ 70,
33
+ 75,
34
+ 80,
35
+ 85,
36
+ 90,
37
+ 95,
38
+ 100
39
+ ],
40
+ "direction": "lower"
41
+ },
42
+ "0.01": {
43
+ "epsilon": 3.023197554051876,
44
+ "num_guesses": 100,
45
+ "correct_guesses": 100,
46
+ "candidate_num_guesses": [
47
+ 5,
48
+ 10,
49
+ 15,
50
+ 20,
51
+ 25,
52
+ 30,
53
+ 35,
54
+ 40,
55
+ 45,
56
+ 50,
57
+ 55,
58
+ 60,
59
+ 65,
60
+ 70,
61
+ 75,
62
+ 80,
63
+ 85,
64
+ 90,
65
+ 95,
66
+ 100
67
+ ],
68
+ "direction": "lower"
69
+ }
70
+ }
71
+ },
72
+ "embedding": {
73
+ "auc": 0.916224,
74
+ "empirical_epsilon": {
75
+ "0.05": 3.4791953936219215,
76
+ "0.01": 3.023197554051876
77
+ },
78
+ "empirical_epsilon_details": {
79
+ "0.05": {
80
+ "epsilon": 3.4791953936219215,
81
+ "num_guesses": 100,
82
+ "correct_guesses": 100,
83
+ "candidate_num_guesses": [
84
+ 5,
85
+ 10,
86
+ 15,
87
+ 20,
88
+ 25,
89
+ 30,
90
+ 35,
91
+ 40,
92
+ 45,
93
+ 50,
94
+ 55,
95
+ 60,
96
+ 65,
97
+ 70,
98
+ 75,
99
+ 80,
100
+ 85,
101
+ 90,
102
+ 95,
103
+ 100
104
+ ],
105
+ "direction": "lower"
106
+ },
107
+ "0.01": {
108
+ "epsilon": 3.023197554051876,
109
+ "num_guesses": 100,
110
+ "correct_guesses": 100,
111
+ "candidate_num_guesses": [
112
+ 5,
113
+ 10,
114
+ 15,
115
+ 20,
116
+ 25,
117
+ 30,
118
+ 35,
119
+ 40,
120
+ 45,
121
+ 50,
122
+ 55,
123
+ 60,
124
+ 65,
125
+ 70,
126
+ 75,
127
+ 80,
128
+ 85,
129
+ 90,
130
+ 95,
131
+ 100
132
+ ],
133
+ "direction": "lower"
134
+ }
135
+ }
136
+ }
137
+ }
starcoder2-7b/base/epochs/epoch_002/audit_scores.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df7a9ad49c78c2b7e30773bfe7d47ce312de900740e940d72ac0b3117dff09bf
3
+ size 12784
starcoder2-7b/base/metrics.jsonl ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"timestamp": 1774091071.9870658, "event": "train_step", "step": 10, "epoch": 1, "metrics": {"train/step_loss": 1.5507386380975896, "train/step_real_loss": 1.2866992354393005, "train/lr": 4.761904761904762e-05, "train/step_canary_loss": 10.0, "perf/step_duration_sec": 8.244300998747349, "perf/samples_per_sec": 8.00553012438873, "perf/tokens_per_sec": 6440.812872803661, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 53100.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.16860580444336}}
2
+ {"timestamp": 1774091154.380765, "event": "train_step", "step": 20, "epoch": 1, "metrics": {"train/step_loss": 1.4504926845208923, "train/step_real_loss": 1.2006134390830994, "train/lr": 9.523809523809524e-05, "train/step_canary_loss": 6.78125, "perf/step_duration_sec": 8.244972918182611, "perf/samples_per_sec": 8.126163744242886, "perf/tokens_per_sec": 6422.095078472549, "perf/logical_batch_size": 67.0, "perf/logical_token_count": 52950.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.911073684692383, "system/cuda_max_memory_allocated_gb": 82.16860580444336}}
3
+ {"timestamp": 1774091235.6533117, "event": "train_step", "step": 30, "epoch": 1, "metrics": {"train/step_loss": 0.9861254319548607, "train/step_real_loss": 0.9861254319548607, "train/lr": 9.98706541985615e-05, "perf/step_duration_sec": 7.973596462979913, "perf/samples_per_sec": 8.026490968929942, "perf/tokens_per_sec": 7148.9697609673985, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 57003.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.16860580444336}}
4
+ {"timestamp": 1774091318.662035, "event": "train_step", "step": 40, "epoch": 1, "metrics": {"train/step_loss": 1.0557814985513687, "train/step_real_loss": 1.0557814985513687, "train/lr": 9.942439201095397e-05, "perf/step_duration_sec": 8.119743634015322, "perf/samples_per_sec": 7.882022251526572, "perf/tokens_per_sec": 6340.347961767047, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 51482.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
5
+ {"timestamp": 1774091402.0365882, "event": "train_step", "step": 50, "epoch": 1, "metrics": {"train/step_loss": 1.017115435233483, "train/step_real_loss": 0.9500000476837158, "train/lr": 9.866246608261724e-05, "train/step_canary_loss": 5.3125, "perf/step_duration_sec": 8.119036318734288, "perf/samples_per_sec": 8.005876245437603, "perf/tokens_per_sec": 6307.645142790008, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 51212.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
6
+ {"timestamp": 1774091429.1958544, "event": "eval_step", "step": 50, "epoch": 1, "metrics": {"eval/loss": 0.8033588177513802, "eval/duration_sec": 27.156014366075397}}
7
+ {"timestamp": 1774091511.9092546, "event": "train_step", "step": 60, "epoch": 1, "metrics": {"train/step_loss": 1.2237938749256418, "train/step_real_loss": 0.9667061120271683, "train/lr": 9.7589742682592e-05, "train/step_canary_loss": 6.708333333333333, "perf/step_duration_sec": 8.495171755552292, "perf/samples_per_sec": 7.886832889071372, "perf/tokens_per_sec": 6228.008276044599, "perf/logical_batch_size": 67.0, "perf/logical_token_count": 52908.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
8
+ {"timestamp": 1774091593.5344725, "event": "train_step", "step": 70, "epoch": 1, "metrics": {"train/step_loss": 1.0743033794256358, "train/step_real_loss": 0.9387456253170967, "train/lr": 9.621307308142384e-05, "train/step_canary_loss": 9.75, "perf/step_duration_sec": 8.257235972210765, "perf/samples_per_sec": 7.8718835478062665, "perf/tokens_per_sec": 6568.057420488078, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 54234.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
9
+ {"timestamp": 1774091676.5485113, "event": "train_step", "step": 80, "epoch": 1, "metrics": {"train/step_loss": 0.8626528829336166, "train/step_real_loss": 0.8626528829336166, "train/lr": 9.454124979346391e-05, "perf/step_duration_sec": 8.386640733107924, "perf/samples_per_sec": 7.631184169765056, "perf/tokens_per_sec": 6658.446662625317, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 55842.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
10
+ {"timestamp": 1774091757.879373, "event": "train_step", "step": 90, "epoch": 1, "metrics": {"train/step_loss": 0.9340124875307083, "train/step_real_loss": 0.9340124875307083, "train/lr": 9.258495042083221e-05, "perf/step_duration_sec": 8.493970146402717, "perf/samples_per_sec": 7.534756880103312, "perf/tokens_per_sec": 5785.162786428072, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 49139.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
11
+ {"timestamp": 1774091839.6918592, "event": "train_step", "step": 100, "epoch": 1, "metrics": {"train/step_loss": 0.9577739440477812, "train/step_real_loss": 0.880453996360302, "train/lr": 9.035666945770107e-05, "train/step_canary_loss": 5.90625, "perf/step_duration_sec": 8.234961070120335, "perf/samples_per_sec": 7.893176354633353, "perf/tokens_per_sec": 6771.616711381143, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 55764.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
12
+ {"timestamp": 1774091866.8768253, "event": "eval_step", "step": 100, "epoch": 1, "metrics": {"eval/loss": 0.7754902612717671, "eval/duration_sec": 27.182465851306915}}
13
+ {"timestamp": 1774091950.1670244, "event": "train_step", "step": 110, "epoch": 1, "metrics": {"train/step_loss": 0.9567192720644402, "train/step_real_loss": 0.860762245953083, "train/lr": 8.787063849045118e-05, "train/step_canary_loss": 4.02734375, "perf/step_duration_sec": 8.036363879218698, "perf/samples_per_sec": 8.21266943507498, "perf/tokens_per_sec": 6478.054102878841, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 52060.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
14
+ {"timestamp": 1774092031.8751383, "event": "train_step", "step": 120, "epoch": 1, "metrics": {"train/step_loss": 0.9170611575245857, "train/step_real_loss": 0.9170611575245857, "train/lr": 8.5142735303366e-05, "perf/step_duration_sec": 7.79720464348793, "perf/samples_per_sec": 8.208069805305357, "perf/tokens_per_sec": 6600.4423832818775, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 51465.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
15
+ {"timestamp": 1774092114.4179726, "event": "train_step", "step": 130, "epoch": 1, "metrics": {"train/step_loss": 0.9158061426697355, "train/step_real_loss": 0.8711828961968422, "train/lr": 8.219038247038819e-05, "train/step_canary_loss": 2.34375, "perf/step_duration_sec": 8.22744246572256, "perf/samples_per_sec": 8.021933945447975, "perf/tokens_per_sec": 6342.311139506375, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 52181.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
16
+ {"timestamp": 1774092196.3902268, "event": "train_step", "step": 140, "epoch": 1, "metrics": {"train/step_loss": 0.9856266433542425, "train/step_real_loss": 0.8489470109343529, "train/lr": 7.903243608061246e-05, "train/step_canary_loss": 5.359375, "perf/step_duration_sec": 8.333684524521232, "perf/samples_per_sec": 7.9196662419605675, "perf/tokens_per_sec": 6461.007714123103, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 53844.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
17
+ {"timestamp": 1774092278.982639, "event": "train_step", "step": 150, "epoch": 1, "metrics": {"train/step_loss": 0.94221530854702, "train/step_real_loss": 0.94221530854702, "train/lr": 7.568906530820282e-05, "perf/step_duration_sec": 8.388992108404636, "perf/samples_per_sec": 7.629045202686584, "perf/tokens_per_sec": 6081.421861022835, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 51017.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
18
+ {"timestamp": 1774092306.1516783, "event": "eval_step", "step": 150, "epoch": 1, "metrics": {"eval/loss": 0.7622188641422096, "eval/duration_sec": 27.166858203709126}}
19
+ {"timestamp": 1774092388.1349308, "event": "train_step", "step": 160, "epoch": 1, "metrics": {"train/step_loss": 0.9667639210820198, "train/step_real_loss": 0.9667639210820198, "train/lr": 7.21816235958972e-05, "perf/step_duration_sec": 8.235355000942945, "perf/samples_per_sec": 7.771371117902267, "perf/tokens_per_sec": 6008.24129552819, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 49480.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
20
+ {"timestamp": 1774092469.9282126, "event": "train_step", "step": 170, "epoch": 1, "metrics": {"train/step_loss": 0.9262635037302971, "train/step_real_loss": 0.9262635037302971, "train/lr": 6.853251227482479e-05, "perf/step_duration_sec": 7.9527543764561415, "perf/samples_per_sec": 8.047526299752175, "perf/tokens_per_sec": 6280.842791759705, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 49950.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
21
+ {"timestamp": 1774092551.943189, "event": "train_step", "step": 180, "epoch": 1, "metrics": {"train/step_loss": 0.8705217910535408, "train/step_real_loss": 0.8467002063989639, "train/lr": 6.476503749166904e-05, "train/step_canary_loss": 1.6328125, "perf/step_duration_sec": 8.540382800623775, "perf/samples_per_sec": 7.7279908337574135, "perf/tokens_per_sec": 6385.076790236751, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 54531.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
22
+ {"timestamp": 1774092634.1860914, "event": "train_step", "step": 190, "epoch": 1, "metrics": {"train/step_loss": 0.8329041105050307, "train/step_real_loss": 0.830781526863575, "train/lr": 6.090326135695403e-05, "train/step_canary_loss": 0.96875, "perf/step_duration_sec": 7.938739079982042, "perf/samples_per_sec": 8.18769824088324, "perf/tokens_per_sec": 6963.07051322375, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 55278.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
23
+ {"timestamp": 1774092717.69117, "event": "train_step", "step": 200, "epoch": 1, "metrics": {"train/step_loss": 0.9225348509274997, "train/step_real_loss": 0.8622424304485321, "train/lr": 5.697184826514057e-05, "train/step_canary_loss": 4.78125, "perf/step_duration_sec": 8.257816776633263, "perf/samples_per_sec": 7.871329887571168, "perf/tokens_per_sec": 6772.4922352662325, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 55926.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
24
+ {"timestamp": 1774092744.8800561, "event": "eval_step", "step": 200, "epoch": 1, "metrics": {"eval/loss": 0.7524948388166811, "eval/duration_sec": 27.18629032932222}}
25
+ {"timestamp": 1774092830.2611096, "event": "train_epoch", "step": 207, "epoch": 1, "metrics": {"train/epoch_loss": 1.036294524908722, "train/epoch_real_loss": 0.957591299003593, "train/epoch_canary_loss": 5.154934052784185, "perf/epoch_duration_sec": 1814.8832574244589, "perf/epoch_samples_per_sec": 29.762244915220545, "perf/epoch_tokens_per_sec": 24175.732417227537, "perf/epoch_samples": 54015.0, "perf/epoch_tokens": 43876132.0, "system/cuda_epoch_peak_memory_gb": 88.38552379608154, "eval/loss": 0.7516984458284817, "eval/duration_sec": 27.281295781955123}}
26
+ {"timestamp": 1774092840.935869, "event": "audit_epoch", "step": 207, "epoch": 1, "metrics": {"audit/delta": 1e-05, "audit/num_canaries": 500.0, "audit/num_members": 250.0, "audit/paper_guess_fraction": 0.2, "audit/paper_guess_steps": 20.0, "audit/loss/auc": 0.998072, "audit/loss/empirical_epsilon/0.05": 3.4791953936219215, "audit/loss/empirical_epsilon/0.01": 3.023197554051876, "audit/loss/empirical_epsilon_details/0.05/epsilon": 3.4791953936219215, "audit/loss/empirical_epsilon_details/0.05/num_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.05/correct_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.01/epsilon": 3.023197554051876, "audit/loss/empirical_epsilon_details/0.01/num_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.01/correct_guesses": 100.0, "audit/embedding/auc": 0.96, "audit/embedding/empirical_epsilon/0.05": 3.4791953936219215, "audit/embedding/empirical_epsilon/0.01": 3.023197554051876, "audit/embedding/empirical_epsilon_details/0.05/epsilon": 3.4791953936219215, "audit/embedding/empirical_epsilon_details/0.05/num_guesses": 100.0, "audit/embedding/empirical_epsilon_details/0.05/correct_guesses": 100.0, "audit/embedding/empirical_epsilon_details/0.01/epsilon": 3.023197554051876, "audit/embedding/empirical_epsilon_details/0.01/num_guesses": 100.0, "audit/embedding/empirical_epsilon_details/0.01/correct_guesses": 100.0, "perf/audit_duration_sec": 6.3129072319716215}}
27
+ {"timestamp": 1774092866.0795767, "event": "train_step", "step": 210, "epoch": 2, "metrics": {"train/step_loss": 0.8786039505944108, "train/step_real_loss": 0.8291560411453247, "train/lr": 5.29959073680547e-05, "train/step_canary_loss": 2.4609375, "perf/step_duration_sec": 8.105620637536049, "perf/samples_per_sec": 8.142498020985931, "perf/tokens_per_sec": 6755.805933775599, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 54760.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.911073684692383, "system/cuda_max_memory_allocated_gb": 75.98184061050415}}
28
+ {"timestamp": 1774092947.7881637, "event": "train_step", "step": 220, "epoch": 2, "metrics": {"train/step_loss": 0.8338463682394761, "train/step_real_loss": 0.8286867365241051, "train/lr": 4.9000832207739676e-05, "train/step_canary_loss": 1.1640625, "perf/step_duration_sec": 8.100939376279712, "perf/samples_per_sec": 8.023760823383757, "perf/tokens_per_sec": 6522.083124668936, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 52835.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
29
+ {"timestamp": 1774093029.3670137, "event": "train_step", "step": 230, "epoch": 2, "metrics": {"train/step_loss": 0.7926744809517494, "train/step_real_loss": 0.8034120723605156, "train/lr": 4.501213853296425e-05, "train/step_canary_loss": 0.10546875, "perf/step_duration_sec": 8.09845926053822, "perf/samples_per_sec": 8.026218063073905, "perf/tokens_per_sec": 6708.683497950849, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 54330.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
30
+ {"timestamp": 1774093111.4853563, "event": "train_step", "step": 240, "epoch": 2, "metrics": {"train/step_loss": 0.8129148510369387, "train/step_real_loss": 0.7906041964888573, "train/lr": 4.1055301335220955e-05, "train/step_canary_loss": 1.52685546875, "perf/step_duration_sec": 8.505769751966, "perf/samples_per_sec": 7.75943881913156, "perf/tokens_per_sec": 6383.901937558238, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 54300.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
31
+ {"timestamp": 1774093192.88336, "event": "train_step", "step": 250, "epoch": 2, "metrics": {"train/step_loss": 0.8410789425556476, "train/step_real_loss": 0.850802831351757, "train/lr": 3.715559214503298e-05, "train/step_canary_loss": 0.21875, "perf/step_duration_sec": 8.232771871611476, "perf/samples_per_sec": 7.895275250385015, "perf/tokens_per_sec": 6654.745309889904, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 54787.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
32
+ {"timestamp": 1774093220.0732572, "event": "eval_step", "step": 250, "epoch": 2, "metrics": {"eval/loss": 0.748522541389383, "eval/duration_sec": 27.187080297619104}}
33
+ {"timestamp": 1774093302.2765276, "event": "train_step", "step": 260, "epoch": 2, "metrics": {"train/step_loss": 0.8724462010643699, "train/step_real_loss": 0.8488068357110023, "train/lr": 3.33379176277258e-05, "train/step_canary_loss": 1.62890625, "perf/step_duration_sec": 8.470736568793654, "perf/samples_per_sec": 7.7915302245551095, "perf/tokens_per_sec": 6277.49423773815, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 53175.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.911073684692383, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
34
+ {"timestamp": 1774093384.391679, "event": "train_step", "step": 270, "epoch": 2, "metrics": {"train/step_loss": 0.8640658855438232, "train/step_real_loss": 0.7926059737801552, "train/lr": 2.962666050951997e-05, "train/step_canary_loss": 5.4375, "perf/step_duration_sec": 7.947357261553407, "perf/samples_per_sec": 8.178819431516908, "perf/tokens_per_sec": 6407.161314659091, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 50920.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
35
+ {"timestamp": 1774093468.0585487, "event": "train_step", "step": 280, "epoch": 2, "metrics": {"train/step_loss": 0.7910019425245431, "train/step_real_loss": 0.8018507286906242, "train/lr": 2.604552384991855e-05, "train/step_canary_loss": 0.0966796875, "perf/step_duration_sec": 8.62639987282455, "perf/samples_per_sec": 7.53500892124967, "perf/tokens_per_sec": 6275.039506402552, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 54131.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
36
+ {"timestamp": 1774093550.7132335, "event": "train_step", "step": 290, "epoch": 2, "metrics": {"train/step_loss": 0.8430270507480159, "train/step_real_loss": 0.8613760396838188, "train/lr": 2.2617379654990623e-05, "train/step_canary_loss": 0.255859375, "perf/step_duration_sec": 8.335046991705894, "perf/samples_per_sec": 7.9183716739300705, "perf/tokens_per_sec": 6261.392413496007, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 52189.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
37
+ {"timestamp": 1774093632.5524132, "event": "train_step", "step": 300, "epoch": 2, "metrics": {"train/step_loss": 0.7691241015087474, "train/step_real_loss": 0.784599058330059, "train/lr": 1.936412279842705e-05, "train/step_canary_loss": 0.27392578125, "perf/step_duration_sec": 8.091827008873224, "perf/samples_per_sec": 8.156378025336753, "perf/tokens_per_sec": 5984.062677922065, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 48422.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.911073684692383, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
38
+ {"timestamp": 1774093659.719083, "event": "eval_step", "step": 300, "epoch": 2, "metrics": {"eval/loss": 0.7462242173737493, "eval/duration_sec": 27.16416385397315}}
39
+ {"timestamp": 1774093742.5612013, "event": "train_step", "step": 310, "epoch": 2, "metrics": {"train/step_loss": 0.7514625552928809, "train/step_real_loss": 0.7710395082831383, "train/lr": 1.6306531183346385e-05, "train/step_canary_loss": 0.125, "perf/step_duration_sec": 8.240195842459798, "perf/samples_per_sec": 8.009518373327666, "perf/tokens_per_sec": 6369.751520897312, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 52488.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
40
+ {"timestamp": 1774093824.76255, "event": "train_step", "step": 320, "epoch": 2, "metrics": {"train/step_loss": 0.8317307600608239, "train/step_real_loss": 0.8416290208697319, "train/lr": 1.3464133037968912e-05, "train/step_canary_loss": 0.1982421875, "perf/step_duration_sec": 7.953533122316003, "perf/samples_per_sec": 8.172468637569782, "perf/tokens_per_sec": 6226.666720107783, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 49524.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
41
+ {"timestamp": 1774093907.919859, "event": "train_step", "step": 330, "epoch": 2, "metrics": {"train/step_loss": 0.8719267258277306, "train/step_real_loss": 0.8635779246687889, "train/lr": 1.0855082192715294e-05, "train/step_canary_loss": 1.40625, "perf/step_duration_sec": 8.101611092686653, "perf/samples_per_sec": 8.023095561656334, "perf/tokens_per_sec": 5623.572827524039, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 45560.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
42
+ {"timestamp": 1774093989.7239172, "event": "train_step", "step": 340, "epoch": 2, "metrics": {"train/step_loss": 0.8476224361043988, "train/step_real_loss": 0.8228105828166008, "train/lr": 8.49604213531004e-06, "train/step_canary_loss": 1.6416015625, "perf/step_duration_sec": 8.274557610973716, "perf/samples_per_sec": 7.976257233676254, "perf/tokens_per_sec": 6674.435371234426, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 55228.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
43
+ {"timestamp": 1774094071.9116077, "event": "train_step", "step": 350, "epoch": 2, "metrics": {"train/step_loss": 0.8573094416355741, "train/step_real_loss": 0.816773310303688, "train/lr": 6.402079584406673e-06, "train/step_canary_loss": 1.376171875, "perf/step_duration_sec": 8.637357847765088, "perf/samples_per_sec": 7.988554048140279, "perf/tokens_per_sec": 6212.547974249382, "perf/logical_batch_size": 69.0, "perf/logical_token_count": 53660.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
44
+ {"timestamp": 1774094099.091823, "event": "eval_step", "step": 350, "epoch": 2, "metrics": {"eval/loss": 0.7456388631942629, "eval/duration_sec": 27.17722495831549}}
45
+ {"timestamp": 1774094181.5654633, "event": "train_step", "step": 360, "epoch": 2, "metrics": {"train/step_loss": 0.8446246770712046, "train/step_real_loss": 0.8453097268939018, "train/lr": 4.586568261458729e-06, "train/step_canary_loss": 0.80078125, "perf/step_duration_sec": 7.976690696552396, "perf/samples_per_sec": 8.148742689508275, "perf/tokens_per_sec": 6948.119478163339, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 55423.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
46
+ {"timestamp": 1774094263.5117958, "event": "train_step", "step": 370, "epoch": 2, "metrics": {"train/step_loss": 0.8721330687403679, "train/step_real_loss": 0.8721330687403679, "train/lr": 3.06110347542643e-06, "perf/step_duration_sec": 8.37641635723412, "perf/samples_per_sec": 7.640498904370688, "perf/tokens_per_sec": 5830.058812420977, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 48835.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
47
+ {"timestamp": 1774094346.1630964, "event": "train_step", "step": 380, "epoch": 2, "metrics": {"train/step_loss": 0.81929341004445, "train/step_real_loss": 0.8280055150389671, "train/lr": 1.8354280658494649e-06, "train/step_canary_loss": 0.26171875, "perf/step_duration_sec": 7.9715049508959055, "perf/samples_per_sec": 8.154043734576712, "perf/tokens_per_sec": 6787.676898315027, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 54108.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
48
+ {"timestamp": 1774094429.2324018, "event": "train_step", "step": 390, "epoch": 2, "metrics": {"train/step_loss": 0.8554338663816452, "train/step_real_loss": 0.8554338663816452, "train/lr": 9.17370177272775e-07, "perf/step_duration_sec": 8.24701151996851, "perf/samples_per_sec": 7.760386880148844, "perf/tokens_per_sec": 6142.34621563781, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 50656.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
49
+ {"timestamp": 1774094512.2904398, "event": "train_step", "step": 400, "epoch": 2, "metrics": {"train/step_loss": 0.7998756021261215, "train/step_real_loss": 0.7998756021261215, "train/lr": 3.127932624475638e-07, "perf/step_duration_sec": 7.943094424903393, "perf/samples_per_sec": 8.0573132555677, "perf/tokens_per_sec": 6869.86671452841, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 54568.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
50
+ {"timestamp": 1774094539.4831598, "event": "eval_step", "step": 400, "epoch": 2, "metrics": {"eval/loss": 0.7455017728538349, "eval/duration_sec": 27.190383709967136}}
51
+ {"timestamp": 1774094621.4459202, "event": "train_step", "step": 410, "epoch": 2, "metrics": {"train/step_loss": 0.854436350591255, "train/step_real_loss": 0.8517185375094414, "train/lr": 2.5558633627303928e-08, "train/step_canary_loss": 0.94140625, "perf/step_duration_sec": 8.378037076443434, "perf/samples_per_sec": 7.877740262760654, "perf/tokens_per_sec": 6647.619184760483, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 55694.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
52
+ {"timestamp": 1774094682.1799755, "event": "train_epoch", "step": 414, "epoch": 2, "metrics": {"train/epoch_loss": 0.840315752633972, "train/epoch_real_loss": 0.8362452851041504, "train/epoch_canary_loss": 1.0832555509625612, "perf/epoch_duration_sec": 1814.1021996028721, "perf/epoch_samples_per_sec": 29.765136722628178, "perf/epoch_tokens_per_sec": 24186.19579955584, "perf/epoch_samples": 53997.0, "perf/epoch_tokens": 43876231.0, "system/cuda_epoch_peak_memory_gb": 88.38552379608154, "eval/loss": 0.7454980848164394, "eval/duration_sec": 27.13153049722314}}
53
+ {"timestamp": 1774094693.161264, "event": "audit_epoch", "step": 414, "epoch": 2, "metrics": {"audit/delta": 1e-05, "audit/num_canaries": 500.0, "audit/num_members": 250.0, "audit/paper_guess_fraction": 0.2, "audit/paper_guess_steps": 20.0, "audit/loss/auc": 1.0, "audit/loss/empirical_epsilon/0.05": 3.4791953936219215, "audit/loss/empirical_epsilon/0.01": 3.023197554051876, "audit/loss/empirical_epsilon_details/0.05/epsilon": 3.4791953936219215, "audit/loss/empirical_epsilon_details/0.05/num_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.05/correct_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.01/epsilon": 3.023197554051876, "audit/loss/empirical_epsilon_details/0.01/num_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.01/correct_guesses": 100.0, "audit/embedding/auc": 0.916224, "audit/embedding/empirical_epsilon/0.05": 3.4791953936219215, "audit/embedding/empirical_epsilon/0.01": 3.023197554051876, "audit/embedding/empirical_epsilon_details/0.05/epsilon": 3.4791953936219215, "audit/embedding/empirical_epsilon_details/0.05/num_guesses": 100.0, "audit/embedding/empirical_epsilon_details/0.05/correct_guesses": 100.0, "audit/embedding/empirical_epsilon_details/0.01/epsilon": 3.023197554051876, "audit/embedding/empirical_epsilon_details/0.01/num_guesses": 100.0, "audit/embedding/empirical_epsilon_details/0.01/correct_guesses": 100.0, "perf/audit_duration_sec": 6.722778998315334}}
54
+ {"timestamp": 1774094704.3854914, "event": "audit_final", "step": 414, "epoch": 2, "metrics": {"audit/delta": 1e-05, "audit/num_canaries": 500.0, "audit/num_members": 250.0, "audit/paper_guess_fraction": 0.2, "audit/paper_guess_steps": 20.0, "audit/loss/auc": 1.0, "audit/loss/empirical_epsilon/0.05": 3.4791953936219215, "audit/loss/empirical_epsilon/0.01": 3.023197554051876, "audit/loss/empirical_epsilon_details/0.05/epsilon": 3.4791953936219215, "audit/loss/empirical_epsilon_details/0.05/num_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.05/correct_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.01/epsilon": 3.023197554051876, "audit/loss/empirical_epsilon_details/0.01/num_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.01/correct_guesses": 100.0, "audit/embedding/auc": 0.916224, "audit/embedding/empirical_epsilon/0.05": 3.4791953936219215, "audit/embedding/empirical_epsilon/0.01": 3.023197554051876, "audit/embedding/empirical_epsilon_details/0.05/epsilon": 3.4791953936219215, "audit/embedding/empirical_epsilon_details/0.05/num_guesses": 100.0, "audit/embedding/empirical_epsilon_details/0.05/correct_guesses": 100.0, "audit/embedding/empirical_epsilon_details/0.01/epsilon": 3.023197554051876, "audit/embedding/empirical_epsilon_details/0.01/num_guesses": 100.0, "audit/embedding/empirical_epsilon_details/0.01/correct_guesses": 100.0}}
55
+ {"timestamp": 1774094704.914049, "event": "energy_final", "step": 414, "epoch": null, "metrics": {"energy/codecarbon/duration": 3820.65162669681, "energy/codecarbon/emissions": 0.5568149811716148, "energy/codecarbon/emissions_rate": 0.0001457382236267942, "energy/codecarbon/cpu_power": 179.3439937196561, "energy/codecarbon/gpu_power": 2316.658494927477, "energy/codecarbon/ram_power": 70.0, "energy/codecarbon/cpu_energy": 0.18368298514637418, "energy/codecarbon/gpu_energy": 2.457235179675422, "energy/codecarbon/ram_energy": 0.07169108754482326, "energy/codecarbon/energy_consumed": 2.7126092523666183, "energy/codecarbon/water_consumed": 0.0, "energy/codecarbon/cpu_count": 224.0, "energy/codecarbon/gpu_count": 4.0, "energy/codecarbon/longitude": -121.9552, "energy/codecarbon/latitude": 37.3541, "energy/codecarbon/ram_total_size": 2015.5625190734863, "energy/codecarbon/cpu_utilization_percent": 2.2497114375655825, "energy/codecarbon/gpu_utilization_percent": 96.45330535152151, "energy/codecarbon/ram_utilization_percent": 2.3057974816369358, "energy/codecarbon/ram_used_gb": 46.78091819539025, "energy/codecarbon/pue": 1.0, "energy/codecarbon/wue": 0.0}}
starcoder2-7b/base/pretrain_lm_head.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9a9b40bb4ecc60cfc3a7e2f7cd31be2121d8f32f92d1a4ecdac554172fb4e5b
3
+ size 457594472
starcoder2-7b/base/resolved_config.yaml ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ name: bigcode/starcoder2-7b
3
+ tokenizer_name: bigcode/starcoder2-7b
4
+ max_length: 1024
5
+ dtype: bfloat16
6
+ trust_remote_code: true
7
+ use_fast_tokenizer: true
8
+ cache_dir: null
9
+ local_files_only: false
10
+ low_cpu_mem_usage: true
11
+ tie_word_embeddings: true
12
+ gradient_checkpointing: false
13
+ use_chat_template: false
14
+ dataset:
15
+ name: melihcatal/codedp-cpt
16
+ split: train
17
+ mode: cpt
18
+ text_column: text
19
+ validation_ratio: 0.05
20
+ max_samples: -1
21
+ lora:
22
+ enabled: true
23
+ r: 16
24
+ alpha: 32
25
+ dropout: 0.05
26
+ target_modules:
27
+ - q_proj
28
+ - k_proj
29
+ - v_proj
30
+ - o_proj
31
+ modules_to_save:
32
+ - lm_head
33
+ bias: none
34
+ training:
35
+ seed: 42
36
+ epochs: 2
37
+ warmup_steps: null
38
+ warmup_ratio: 0.05
39
+ mixed_precision: false
40
+ mixed_precision_dtype: bfloat16
41
+ batch_size: 8
42
+ eval_batch_size: 8
43
+ eval_every_steps: 50
44
+ eval_every_epochs: 1
45
+ learning_rate: 0.0001
46
+ optimizer: adamw
47
+ lr_scheduler: cosine
48
+ adam_beta1: 0.9
49
+ adam_beta2: 0.999
50
+ adam_epsilon: 1.0e-08
51
+ sgd_momentum: 0.9
52
+ weight_decay: 0.01
53
+ max_grad_norm: 1.0
54
+ log_every: 10
55
+ gradient_accumulation_steps: 8
56
+ num_workers: 4
57
+ output_dir: runs/cpt/starcoder2-7b/base
58
+ distributed:
59
+ strategy: dpddp
60
+ backend: nccl
61
+ devices: null
62
+ dp:
63
+ module_validator: auto
64
+ target_delta: 1.0e-05
65
+ noise_multiplier: null
66
+ max_grad_norm: 1.0
67
+ grad_sample_mode: ghost
68
+ secure_mode: false
69
+ enabled: false
70
+ target_epsilon: 8.0
71
+ audit:
72
+ enabled: true
73
+ run_every_epoch: true
74
+ epoch_device: cuda
75
+ q_canary: auto
76
+ num_canaries: 500
77
+ prefix_length: 49
78
+ num_digits: 12
79
+ batch_size: 32
80
+ delta: 1.0e-05
81
+ p_values:
82
+ - 0.05
83
+ - 0.01
84
+ paper_guess_fraction: 0.2
85
+ paper_guess_steps: 20
86
+ enable_holdout_empirical_epsilon: false
87
+ holdout_seed: 42
88
+ tie_seed: 42
89
+ tracking:
90
+ enabled: true
91
+ tensorboard: true
92
+ wandb: false
93
+ wandb_project: codedp-finetune-h200-audit
94
+ wandb_run_name: starcoder2-7b-cpt-base
95
+ wandb_mode: online
96
+ codecarbon: true
97
+ codecarbon_output_file: codecarbon.csv
98
+ codecarbon_measure_power_secs: 15
99
+ codecarbon_country_iso_code: null
100
+ codecarbon_project_name: codedp-starcoder2-7b-cpt-base
starcoder2-7b/base/scalars.csv ADDED
@@ -0,0 +1,613 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ timestamp,event,step,epoch,key,value
2
+ 1774091071.9870658,train_step,10,1,train/step_loss,1.5507386380975896
3
+ 1774091071.9870658,train_step,10,1,train/step_real_loss,1.2866992354393005
4
+ 1774091071.9870658,train_step,10,1,train/lr,4.761904761904762e-05
5
+ 1774091071.9870658,train_step,10,1,train/step_canary_loss,10.0
6
+ 1774091071.9870658,train_step,10,1,perf/step_duration_sec,8.244300998747349
7
+ 1774091071.9870658,train_step,10,1,perf/samples_per_sec,8.00553012438873
8
+ 1774091071.9870658,train_step,10,1,perf/tokens_per_sec,6440.812872803661
9
+ 1774091071.9870658,train_step,10,1,perf/logical_batch_size,66.0
10
+ 1774091071.9870658,train_step,10,1,perf/logical_token_count,53100.0
11
+ 1774091071.9870658,train_step,10,1,perf/gradient_accumulation_steps,8.0
12
+ 1774091071.9870658,train_step,10,1,system/cuda_memory_allocated_gb,17.816345691680908
13
+ 1774091071.9870658,train_step,10,1,system/cuda_max_memory_allocated_gb,82.16860580444336
14
+ 1774091154.380765,train_step,20,1,train/step_loss,1.4504926845208923
15
+ 1774091154.380765,train_step,20,1,train/step_real_loss,1.2006134390830994
16
+ 1774091154.380765,train_step,20,1,train/lr,9.523809523809524e-05
17
+ 1774091154.380765,train_step,20,1,train/step_canary_loss,6.78125
18
+ 1774091154.380765,train_step,20,1,perf/step_duration_sec,8.244972918182611
19
+ 1774091154.380765,train_step,20,1,perf/samples_per_sec,8.126163744242886
20
+ 1774091154.380765,train_step,20,1,perf/tokens_per_sec,6422.095078472549
21
+ 1774091154.380765,train_step,20,1,perf/logical_batch_size,67.0
22
+ 1774091154.380765,train_step,20,1,perf/logical_token_count,52950.0
23
+ 1774091154.380765,train_step,20,1,perf/gradient_accumulation_steps,8.0
24
+ 1774091154.380765,train_step,20,1,system/cuda_memory_allocated_gb,17.911073684692383
25
+ 1774091154.380765,train_step,20,1,system/cuda_max_memory_allocated_gb,82.16860580444336
26
+ 1774091235.6533117,train_step,30,1,train/step_loss,0.9861254319548607
27
+ 1774091235.6533117,train_step,30,1,train/step_real_loss,0.9861254319548607
28
+ 1774091235.6533117,train_step,30,1,train/lr,9.98706541985615e-05
29
+ 1774091235.6533117,train_step,30,1,perf/step_duration_sec,7.973596462979913
30
+ 1774091235.6533117,train_step,30,1,perf/samples_per_sec,8.026490968929942
31
+ 1774091235.6533117,train_step,30,1,perf/tokens_per_sec,7148.9697609673985
32
+ 1774091235.6533117,train_step,30,1,perf/logical_batch_size,64.0
33
+ 1774091235.6533117,train_step,30,1,perf/logical_token_count,57003.0
34
+ 1774091235.6533117,train_step,30,1,perf/gradient_accumulation_steps,8.0
35
+ 1774091235.6533117,train_step,30,1,system/cuda_memory_allocated_gb,17.816345691680908
36
+ 1774091235.6533117,train_step,30,1,system/cuda_max_memory_allocated_gb,82.16860580444336
37
+ 1774091318.662035,train_step,40,1,train/step_loss,1.0557814985513687
38
+ 1774091318.662035,train_step,40,1,train/step_real_loss,1.0557814985513687
39
+ 1774091318.662035,train_step,40,1,train/lr,9.942439201095397e-05
40
+ 1774091318.662035,train_step,40,1,perf/step_duration_sec,8.119743634015322
41
+ 1774091318.662035,train_step,40,1,perf/samples_per_sec,7.882022251526572
42
+ 1774091318.662035,train_step,40,1,perf/tokens_per_sec,6340.347961767047
43
+ 1774091318.662035,train_step,40,1,perf/logical_batch_size,64.0
44
+ 1774091318.662035,train_step,40,1,perf/logical_token_count,51482.0
45
+ 1774091318.662035,train_step,40,1,perf/gradient_accumulation_steps,8.0
46
+ 1774091318.662035,train_step,40,1,system/cuda_memory_allocated_gb,17.816345691680908
47
+ 1774091318.662035,train_step,40,1,system/cuda_max_memory_allocated_gb,88.38552379608154
48
+ 1774091402.0365882,train_step,50,1,train/step_loss,1.017115435233483
49
+ 1774091402.0365882,train_step,50,1,train/step_real_loss,0.9500000476837158
50
+ 1774091402.0365882,train_step,50,1,train/lr,9.866246608261724e-05
51
+ 1774091402.0365882,train_step,50,1,train/step_canary_loss,5.3125
52
+ 1774091402.0365882,train_step,50,1,perf/step_duration_sec,8.119036318734288
53
+ 1774091402.0365882,train_step,50,1,perf/samples_per_sec,8.005876245437603
54
+ 1774091402.0365882,train_step,50,1,perf/tokens_per_sec,6307.645142790008
55
+ 1774091402.0365882,train_step,50,1,perf/logical_batch_size,65.0
56
+ 1774091402.0365882,train_step,50,1,perf/logical_token_count,51212.0
57
+ 1774091402.0365882,train_step,50,1,perf/gradient_accumulation_steps,8.0
58
+ 1774091402.0365882,train_step,50,1,system/cuda_memory_allocated_gb,17.816345691680908
59
+ 1774091402.0365882,train_step,50,1,system/cuda_max_memory_allocated_gb,88.38552379608154
60
+ 1774091429.1958544,eval_step,50,1,eval/loss,0.8033588177513802
61
+ 1774091429.1958544,eval_step,50,1,eval/duration_sec,27.156014366075397
62
+ 1774091511.9092546,train_step,60,1,train/step_loss,1.2237938749256418
63
+ 1774091511.9092546,train_step,60,1,train/step_real_loss,0.9667061120271683
64
+ 1774091511.9092546,train_step,60,1,train/lr,9.7589742682592e-05
65
+ 1774091511.9092546,train_step,60,1,train/step_canary_loss,6.708333333333333
66
+ 1774091511.9092546,train_step,60,1,perf/step_duration_sec,8.495171755552292
67
+ 1774091511.9092546,train_step,60,1,perf/samples_per_sec,7.886832889071372
68
+ 1774091511.9092546,train_step,60,1,perf/tokens_per_sec,6228.008276044599
69
+ 1774091511.9092546,train_step,60,1,perf/logical_batch_size,67.0
70
+ 1774091511.9092546,train_step,60,1,perf/logical_token_count,52908.0
71
+ 1774091511.9092546,train_step,60,1,perf/gradient_accumulation_steps,8.0
72
+ 1774091511.9092546,train_step,60,1,system/cuda_memory_allocated_gb,17.816345691680908
73
+ 1774091511.9092546,train_step,60,1,system/cuda_max_memory_allocated_gb,88.38552379608154
74
+ 1774091593.5344725,train_step,70,1,train/step_loss,1.0743033794256358
75
+ 1774091593.5344725,train_step,70,1,train/step_real_loss,0.9387456253170967
76
+ 1774091593.5344725,train_step,70,1,train/lr,9.621307308142384e-05
77
+ 1774091593.5344725,train_step,70,1,train/step_canary_loss,9.75
78
+ 1774091593.5344725,train_step,70,1,perf/step_duration_sec,8.257235972210765
79
+ 1774091593.5344725,train_step,70,1,perf/samples_per_sec,7.8718835478062665
80
+ 1774091593.5344725,train_step,70,1,perf/tokens_per_sec,6568.057420488078
81
+ 1774091593.5344725,train_step,70,1,perf/logical_batch_size,65.0
82
+ 1774091593.5344725,train_step,70,1,perf/logical_token_count,54234.0
83
+ 1774091593.5344725,train_step,70,1,perf/gradient_accumulation_steps,8.0
84
+ 1774091593.5344725,train_step,70,1,system/cuda_memory_allocated_gb,17.816345691680908
85
+ 1774091593.5344725,train_step,70,1,system/cuda_max_memory_allocated_gb,88.38552379608154
86
+ 1774091676.5485113,train_step,80,1,train/step_loss,0.8626528829336166
87
+ 1774091676.5485113,train_step,80,1,train/step_real_loss,0.8626528829336166
88
+ 1774091676.5485113,train_step,80,1,train/lr,9.454124979346391e-05
89
+ 1774091676.5485113,train_step,80,1,perf/step_duration_sec,8.386640733107924
90
+ 1774091676.5485113,train_step,80,1,perf/samples_per_sec,7.631184169765056
91
+ 1774091676.5485113,train_step,80,1,perf/tokens_per_sec,6658.446662625317
92
+ 1774091676.5485113,train_step,80,1,perf/logical_batch_size,64.0
93
+ 1774091676.5485113,train_step,80,1,perf/logical_token_count,55842.0
94
+ 1774091676.5485113,train_step,80,1,perf/gradient_accumulation_steps,8.0
95
+ 1774091676.5485113,train_step,80,1,system/cuda_memory_allocated_gb,17.816345691680908
96
+ 1774091676.5485113,train_step,80,1,system/cuda_max_memory_allocated_gb,88.38552379608154
97
+ 1774091757.879373,train_step,90,1,train/step_loss,0.9340124875307083
98
+ 1774091757.879373,train_step,90,1,train/step_real_loss,0.9340124875307083
99
+ 1774091757.879373,train_step,90,1,train/lr,9.258495042083221e-05
100
+ 1774091757.879373,train_step,90,1,perf/step_duration_sec,8.493970146402717
101
+ 1774091757.879373,train_step,90,1,perf/samples_per_sec,7.534756880103312
102
+ 1774091757.879373,train_step,90,1,perf/tokens_per_sec,5785.162786428072
103
+ 1774091757.879373,train_step,90,1,perf/logical_batch_size,64.0
104
+ 1774091757.879373,train_step,90,1,perf/logical_token_count,49139.0
105
+ 1774091757.879373,train_step,90,1,perf/gradient_accumulation_steps,8.0
106
+ 1774091757.879373,train_step,90,1,system/cuda_memory_allocated_gb,17.816345691680908
107
+ 1774091757.879373,train_step,90,1,system/cuda_max_memory_allocated_gb,88.38552379608154
108
+ 1774091839.6918592,train_step,100,1,train/step_loss,0.9577739440477812
109
+ 1774091839.6918592,train_step,100,1,train/step_real_loss,0.880453996360302
110
+ 1774091839.6918592,train_step,100,1,train/lr,9.035666945770107e-05
111
+ 1774091839.6918592,train_step,100,1,train/step_canary_loss,5.90625
112
+ 1774091839.6918592,train_step,100,1,perf/step_duration_sec,8.234961070120335
113
+ 1774091839.6918592,train_step,100,1,perf/samples_per_sec,7.893176354633353
114
+ 1774091839.6918592,train_step,100,1,perf/tokens_per_sec,6771.616711381143
115
+ 1774091839.6918592,train_step,100,1,perf/logical_batch_size,65.0
116
+ 1774091839.6918592,train_step,100,1,perf/logical_token_count,55764.0
117
+ 1774091839.6918592,train_step,100,1,perf/gradient_accumulation_steps,8.0
118
+ 1774091839.6918592,train_step,100,1,system/cuda_memory_allocated_gb,17.816345691680908
119
+ 1774091839.6918592,train_step,100,1,system/cuda_max_memory_allocated_gb,88.38552379608154
120
+ 1774091866.8768253,eval_step,100,1,eval/loss,0.7754902612717671
121
+ 1774091866.8768253,eval_step,100,1,eval/duration_sec,27.182465851306915
122
+ 1774091950.1670244,train_step,110,1,train/step_loss,0.9567192720644402
123
+ 1774091950.1670244,train_step,110,1,train/step_real_loss,0.860762245953083
124
+ 1774091950.1670244,train_step,110,1,train/lr,8.787063849045118e-05
125
+ 1774091950.1670244,train_step,110,1,train/step_canary_loss,4.02734375
126
+ 1774091950.1670244,train_step,110,1,perf/step_duration_sec,8.036363879218698
127
+ 1774091950.1670244,train_step,110,1,perf/samples_per_sec,8.21266943507498
128
+ 1774091950.1670244,train_step,110,1,perf/tokens_per_sec,6478.054102878841
129
+ 1774091950.1670244,train_step,110,1,perf/logical_batch_size,66.0
130
+ 1774091950.1670244,train_step,110,1,perf/logical_token_count,52060.0
131
+ 1774091950.1670244,train_step,110,1,perf/gradient_accumulation_steps,8.0
132
+ 1774091950.1670244,train_step,110,1,system/cuda_memory_allocated_gb,17.816345691680908
133
+ 1774091950.1670244,train_step,110,1,system/cuda_max_memory_allocated_gb,88.38552379608154
134
+ 1774092031.8751383,train_step,120,1,train/step_loss,0.9170611575245857
135
+ 1774092031.8751383,train_step,120,1,train/step_real_loss,0.9170611575245857
136
+ 1774092031.8751383,train_step,120,1,train/lr,8.5142735303366e-05
137
+ 1774092031.8751383,train_step,120,1,perf/step_duration_sec,7.79720464348793
138
+ 1774092031.8751383,train_step,120,1,perf/samples_per_sec,8.208069805305357
139
+ 1774092031.8751383,train_step,120,1,perf/tokens_per_sec,6600.4423832818775
140
+ 1774092031.8751383,train_step,120,1,perf/logical_batch_size,64.0
141
+ 1774092031.8751383,train_step,120,1,perf/logical_token_count,51465.0
142
+ 1774092031.8751383,train_step,120,1,perf/gradient_accumulation_steps,8.0
143
+ 1774092031.8751383,train_step,120,1,system/cuda_memory_allocated_gb,17.816345691680908
144
+ 1774092031.8751383,train_step,120,1,system/cuda_max_memory_allocated_gb,88.38552379608154
145
+ 1774092114.4179726,train_step,130,1,train/step_loss,0.9158061426697355
146
+ 1774092114.4179726,train_step,130,1,train/step_real_loss,0.8711828961968422
147
+ 1774092114.4179726,train_step,130,1,train/lr,8.219038247038819e-05
148
+ 1774092114.4179726,train_step,130,1,train/step_canary_loss,2.34375
149
+ 1774092114.4179726,train_step,130,1,perf/step_duration_sec,8.22744246572256
150
+ 1774092114.4179726,train_step,130,1,perf/samples_per_sec,8.021933945447975
151
+ 1774092114.4179726,train_step,130,1,perf/tokens_per_sec,6342.311139506375
152
+ 1774092114.4179726,train_step,130,1,perf/logical_batch_size,66.0
153
+ 1774092114.4179726,train_step,130,1,perf/logical_token_count,52181.0
154
+ 1774092114.4179726,train_step,130,1,perf/gradient_accumulation_steps,8.0
155
+ 1774092114.4179726,train_step,130,1,system/cuda_memory_allocated_gb,17.816345691680908
156
+ 1774092114.4179726,train_step,130,1,system/cuda_max_memory_allocated_gb,88.38552379608154
157
+ 1774092196.3902268,train_step,140,1,train/step_loss,0.9856266433542425
158
+ 1774092196.3902268,train_step,140,1,train/step_real_loss,0.8489470109343529
159
+ 1774092196.3902268,train_step,140,1,train/lr,7.903243608061246e-05
160
+ 1774092196.3902268,train_step,140,1,train/step_canary_loss,5.359375
161
+ 1774092196.3902268,train_step,140,1,perf/step_duration_sec,8.333684524521232
162
+ 1774092196.3902268,train_step,140,1,perf/samples_per_sec,7.9196662419605675
163
+ 1774092196.3902268,train_step,140,1,perf/tokens_per_sec,6461.007714123103
164
+ 1774092196.3902268,train_step,140,1,perf/logical_batch_size,66.0
165
+ 1774092196.3902268,train_step,140,1,perf/logical_token_count,53844.0
166
+ 1774092196.3902268,train_step,140,1,perf/gradient_accumulation_steps,8.0
167
+ 1774092196.3902268,train_step,140,1,system/cuda_memory_allocated_gb,17.816345691680908
168
+ 1774092196.3902268,train_step,140,1,system/cuda_max_memory_allocated_gb,88.38552379608154
169
+ 1774092278.982639,train_step,150,1,train/step_loss,0.94221530854702
170
+ 1774092278.982639,train_step,150,1,train/step_real_loss,0.94221530854702
171
+ 1774092278.982639,train_step,150,1,train/lr,7.568906530820282e-05
172
+ 1774092278.982639,train_step,150,1,perf/step_duration_sec,8.388992108404636
173
+ 1774092278.982639,train_step,150,1,perf/samples_per_sec,7.629045202686584
174
+ 1774092278.982639,train_step,150,1,perf/tokens_per_sec,6081.421861022835
175
+ 1774092278.982639,train_step,150,1,perf/logical_batch_size,64.0
176
+ 1774092278.982639,train_step,150,1,perf/logical_token_count,51017.0
177
+ 1774092278.982639,train_step,150,1,perf/gradient_accumulation_steps,8.0
178
+ 1774092278.982639,train_step,150,1,system/cuda_memory_allocated_gb,17.816345691680908
179
+ 1774092278.982639,train_step,150,1,system/cuda_max_memory_allocated_gb,88.38552379608154
180
+ 1774092306.1516783,eval_step,150,1,eval/loss,0.7622188641422096
181
+ 1774092306.1516783,eval_step,150,1,eval/duration_sec,27.166858203709126
182
+ 1774092388.1349308,train_step,160,1,train/step_loss,0.9667639210820198
183
+ 1774092388.1349308,train_step,160,1,train/step_real_loss,0.9667639210820198
184
+ 1774092388.1349308,train_step,160,1,train/lr,7.21816235958972e-05
185
+ 1774092388.1349308,train_step,160,1,perf/step_duration_sec,8.235355000942945
186
+ 1774092388.1349308,train_step,160,1,perf/samples_per_sec,7.771371117902267
187
+ 1774092388.1349308,train_step,160,1,perf/tokens_per_sec,6008.24129552819
188
+ 1774092388.1349308,train_step,160,1,perf/logical_batch_size,64.0
189
+ 1774092388.1349308,train_step,160,1,perf/logical_token_count,49480.0
190
+ 1774092388.1349308,train_step,160,1,perf/gradient_accumulation_steps,8.0
191
+ 1774092388.1349308,train_step,160,1,system/cuda_memory_allocated_gb,17.816345691680908
192
+ 1774092388.1349308,train_step,160,1,system/cuda_max_memory_allocated_gb,88.38552379608154
193
+ 1774092469.9282126,train_step,170,1,train/step_loss,0.9262635037302971
194
+ 1774092469.9282126,train_step,170,1,train/step_real_loss,0.9262635037302971
195
+ 1774092469.9282126,train_step,170,1,train/lr,6.853251227482479e-05
196
+ 1774092469.9282126,train_step,170,1,perf/step_duration_sec,7.9527543764561415
197
+ 1774092469.9282126,train_step,170,1,perf/samples_per_sec,8.047526299752175
198
+ 1774092469.9282126,train_step,170,1,perf/tokens_per_sec,6280.842791759705
199
+ 1774092469.9282126,train_step,170,1,perf/logical_batch_size,64.0
200
+ 1774092469.9282126,train_step,170,1,perf/logical_token_count,49950.0
201
+ 1774092469.9282126,train_step,170,1,perf/gradient_accumulation_steps,8.0
202
+ 1774092469.9282126,train_step,170,1,system/cuda_memory_allocated_gb,17.816345691680908
203
+ 1774092469.9282126,train_step,170,1,system/cuda_max_memory_allocated_gb,88.38552379608154
204
+ 1774092551.943189,train_step,180,1,train/step_loss,0.8705217910535408
205
+ 1774092551.943189,train_step,180,1,train/step_real_loss,0.8467002063989639
206
+ 1774092551.943189,train_step,180,1,train/lr,6.476503749166904e-05
207
+ 1774092551.943189,train_step,180,1,train/step_canary_loss,1.6328125
208
+ 1774092551.943189,train_step,180,1,perf/step_duration_sec,8.540382800623775
209
+ 1774092551.943189,train_step,180,1,perf/samples_per_sec,7.7279908337574135
210
+ 1774092551.943189,train_step,180,1,perf/tokens_per_sec,6385.076790236751
211
+ 1774092551.943189,train_step,180,1,perf/logical_batch_size,66.0
212
+ 1774092551.943189,train_step,180,1,perf/logical_token_count,54531.0
213
+ 1774092551.943189,train_step,180,1,perf/gradient_accumulation_steps,8.0
214
+ 1774092551.943189,train_step,180,1,system/cuda_memory_allocated_gb,17.816345691680908
215
+ 1774092551.943189,train_step,180,1,system/cuda_max_memory_allocated_gb,88.38552379608154
216
+ 1774092634.1860914,train_step,190,1,train/step_loss,0.8329041105050307
217
+ 1774092634.1860914,train_step,190,1,train/step_real_loss,0.830781526863575
218
+ 1774092634.1860914,train_step,190,1,train/lr,6.090326135695403e-05
219
+ 1774092634.1860914,train_step,190,1,train/step_canary_loss,0.96875
220
+ 1774092634.1860914,train_step,190,1,perf/step_duration_sec,7.938739079982042
221
+ 1774092634.1860914,train_step,190,1,perf/samples_per_sec,8.18769824088324
222
+ 1774092634.1860914,train_step,190,1,perf/tokens_per_sec,6963.07051322375
223
+ 1774092634.1860914,train_step,190,1,perf/logical_batch_size,65.0
224
+ 1774092634.1860914,train_step,190,1,perf/logical_token_count,55278.0
225
+ 1774092634.1860914,train_step,190,1,perf/gradient_accumulation_steps,8.0
226
+ 1774092634.1860914,train_step,190,1,system/cuda_memory_allocated_gb,17.816345691680908
227
+ 1774092634.1860914,train_step,190,1,system/cuda_max_memory_allocated_gb,88.38552379608154
228
+ 1774092717.69117,train_step,200,1,train/step_loss,0.9225348509274997
229
+ 1774092717.69117,train_step,200,1,train/step_real_loss,0.8622424304485321
230
+ 1774092717.69117,train_step,200,1,train/lr,5.697184826514057e-05
231
+ 1774092717.69117,train_step,200,1,train/step_canary_loss,4.78125
232
+ 1774092717.69117,train_step,200,1,perf/step_duration_sec,8.257816776633263
233
+ 1774092717.69117,train_step,200,1,perf/samples_per_sec,7.871329887571168
234
+ 1774092717.69117,train_step,200,1,perf/tokens_per_sec,6772.4922352662325
235
+ 1774092717.69117,train_step,200,1,perf/logical_batch_size,65.0
236
+ 1774092717.69117,train_step,200,1,perf/logical_token_count,55926.0
237
+ 1774092717.69117,train_step,200,1,perf/gradient_accumulation_steps,8.0
238
+ 1774092717.69117,train_step,200,1,system/cuda_memory_allocated_gb,17.816345691680908
239
+ 1774092717.69117,train_step,200,1,system/cuda_max_memory_allocated_gb,88.38552379608154
240
+ 1774092744.8800561,eval_step,200,1,eval/loss,0.7524948388166811
241
+ 1774092744.8800561,eval_step,200,1,eval/duration_sec,27.18629032932222
242
+ 1774092830.2611096,train_epoch,207,1,train/epoch_loss,1.036294524908722
243
+ 1774092830.2611096,train_epoch,207,1,train/epoch_real_loss,0.957591299003593
244
+ 1774092830.2611096,train_epoch,207,1,train/epoch_canary_loss,5.154934052784185
245
+ 1774092830.2611096,train_epoch,207,1,perf/epoch_duration_sec,1814.8832574244589
246
+ 1774092830.2611096,train_epoch,207,1,perf/epoch_samples_per_sec,29.762244915220545
247
+ 1774092830.2611096,train_epoch,207,1,perf/epoch_tokens_per_sec,24175.732417227537
248
+ 1774092830.2611096,train_epoch,207,1,perf/epoch_samples,54015.0
249
+ 1774092830.2611096,train_epoch,207,1,perf/epoch_tokens,43876132.0
250
+ 1774092830.2611096,train_epoch,207,1,system/cuda_epoch_peak_memory_gb,88.38552379608154
251
+ 1774092830.2611096,train_epoch,207,1,eval/loss,0.7516984458284817
252
+ 1774092830.2611096,train_epoch,207,1,eval/duration_sec,27.281295781955123
253
+ 1774092840.935869,audit_epoch,207,1,audit/delta,1e-05
254
+ 1774092840.935869,audit_epoch,207,1,audit/num_canaries,500.0
255
+ 1774092840.935869,audit_epoch,207,1,audit/num_members,250.0
256
+ 1774092840.935869,audit_epoch,207,1,audit/paper_guess_fraction,0.2
257
+ 1774092840.935869,audit_epoch,207,1,audit/paper_guess_steps,20.0
258
+ 1774092840.935869,audit_epoch,207,1,audit/loss/auc,0.998072
259
+ 1774092840.935869,audit_epoch,207,1,audit/loss/empirical_epsilon/0.05,3.4791953936219215
260
+ 1774092840.935869,audit_epoch,207,1,audit/loss/empirical_epsilon/0.01,3.023197554051876
261
+ 1774092840.935869,audit_epoch,207,1,audit/loss/empirical_epsilon_details/0.05/epsilon,3.4791953936219215
262
+ 1774092840.935869,audit_epoch,207,1,audit/loss/empirical_epsilon_details/0.05/num_guesses,100.0
263
+ 1774092840.935869,audit_epoch,207,1,audit/loss/empirical_epsilon_details/0.05/correct_guesses,100.0
264
+ 1774092840.935869,audit_epoch,207,1,audit/loss/empirical_epsilon_details/0.01/epsilon,3.023197554051876
265
+ 1774092840.935869,audit_epoch,207,1,audit/loss/empirical_epsilon_details/0.01/num_guesses,100.0
266
+ 1774092840.935869,audit_epoch,207,1,audit/loss/empirical_epsilon_details/0.01/correct_guesses,100.0
267
+ 1774092840.935869,audit_epoch,207,1,audit/embedding/auc,0.96
268
+ 1774092840.935869,audit_epoch,207,1,audit/embedding/empirical_epsilon/0.05,3.4791953936219215
269
+ 1774092840.935869,audit_epoch,207,1,audit/embedding/empirical_epsilon/0.01,3.023197554051876
270
+ 1774092840.935869,audit_epoch,207,1,audit/embedding/empirical_epsilon_details/0.05/epsilon,3.4791953936219215
271
+ 1774092840.935869,audit_epoch,207,1,audit/embedding/empirical_epsilon_details/0.05/num_guesses,100.0
272
+ 1774092840.935869,audit_epoch,207,1,audit/embedding/empirical_epsilon_details/0.05/correct_guesses,100.0
273
+ 1774092840.935869,audit_epoch,207,1,audit/embedding/empirical_epsilon_details/0.01/epsilon,3.023197554051876
274
+ 1774092840.935869,audit_epoch,207,1,audit/embedding/empirical_epsilon_details/0.01/num_guesses,100.0
275
+ 1774092840.935869,audit_epoch,207,1,audit/embedding/empirical_epsilon_details/0.01/correct_guesses,100.0
276
+ 1774092840.935869,audit_epoch,207,1,perf/audit_duration_sec,6.3129072319716215
277
+ 1774092866.0795767,train_step,210,2,train/step_loss,0.8786039505944108
278
+ 1774092866.0795767,train_step,210,2,train/step_real_loss,0.8291560411453247
279
+ 1774092866.0795767,train_step,210,2,train/lr,5.29959073680547e-05
280
+ 1774092866.0795767,train_step,210,2,train/step_canary_loss,2.4609375
281
+ 1774092866.0795767,train_step,210,2,perf/step_duration_sec,8.105620637536049
282
+ 1774092866.0795767,train_step,210,2,perf/samples_per_sec,8.142498020985931
283
+ 1774092866.0795767,train_step,210,2,perf/tokens_per_sec,6755.805933775599
284
+ 1774092866.0795767,train_step,210,2,perf/logical_batch_size,66.0
285
+ 1774092866.0795767,train_step,210,2,perf/logical_token_count,54760.0
286
+ 1774092866.0795767,train_step,210,2,perf/gradient_accumulation_steps,8.0
287
+ 1774092866.0795767,train_step,210,2,system/cuda_memory_allocated_gb,17.911073684692383
288
+ 1774092866.0795767,train_step,210,2,system/cuda_max_memory_allocated_gb,75.98184061050415
289
+ 1774092947.7881637,train_step,220,2,train/step_loss,0.8338463682394761
290
+ 1774092947.7881637,train_step,220,2,train/step_real_loss,0.8286867365241051
291
+ 1774092947.7881637,train_step,220,2,train/lr,4.9000832207739676e-05
292
+ 1774092947.7881637,train_step,220,2,train/step_canary_loss,1.1640625
293
+ 1774092947.7881637,train_step,220,2,perf/step_duration_sec,8.100939376279712
294
+ 1774092947.7881637,train_step,220,2,perf/samples_per_sec,8.023760823383757
295
+ 1774092947.7881637,train_step,220,2,perf/tokens_per_sec,6522.083124668936
296
+ 1774092947.7881637,train_step,220,2,perf/logical_batch_size,65.0
297
+ 1774092947.7881637,train_step,220,2,perf/logical_token_count,52835.0
298
+ 1774092947.7881637,train_step,220,2,perf/gradient_accumulation_steps,8.0
299
+ 1774092947.7881637,train_step,220,2,system/cuda_memory_allocated_gb,17.816345691680908
300
+ 1774092947.7881637,train_step,220,2,system/cuda_max_memory_allocated_gb,82.1704511642456
301
+ 1774093029.3670137,train_step,230,2,train/step_loss,0.7926744809517494
302
+ 1774093029.3670137,train_step,230,2,train/step_real_loss,0.8034120723605156
303
+ 1774093029.3670137,train_step,230,2,train/lr,4.501213853296425e-05
304
+ 1774093029.3670137,train_step,230,2,train/step_canary_loss,0.10546875
305
+ 1774093029.3670137,train_step,230,2,perf/step_duration_sec,8.09845926053822
306
+ 1774093029.3670137,train_step,230,2,perf/samples_per_sec,8.026218063073905
307
+ 1774093029.3670137,train_step,230,2,perf/tokens_per_sec,6708.683497950849
308
+ 1774093029.3670137,train_step,230,2,perf/logical_batch_size,65.0
309
+ 1774093029.3670137,train_step,230,2,perf/logical_token_count,54330.0
310
+ 1774093029.3670137,train_step,230,2,perf/gradient_accumulation_steps,8.0
311
+ 1774093029.3670137,train_step,230,2,system/cuda_memory_allocated_gb,17.816345691680908
312
+ 1774093029.3670137,train_step,230,2,system/cuda_max_memory_allocated_gb,82.1704511642456
313
+ 1774093111.4853563,train_step,240,2,train/step_loss,0.8129148510369387
314
+ 1774093111.4853563,train_step,240,2,train/step_real_loss,0.7906041964888573
315
+ 1774093111.4853563,train_step,240,2,train/lr,4.1055301335220955e-05
316
+ 1774093111.4853563,train_step,240,2,train/step_canary_loss,1.52685546875
317
+ 1774093111.4853563,train_step,240,2,perf/step_duration_sec,8.505769751966
318
+ 1774093111.4853563,train_step,240,2,perf/samples_per_sec,7.75943881913156
319
+ 1774093111.4853563,train_step,240,2,perf/tokens_per_sec,6383.901937558238
320
+ 1774093111.4853563,train_step,240,2,perf/logical_batch_size,66.0
321
+ 1774093111.4853563,train_step,240,2,perf/logical_token_count,54300.0
322
+ 1774093111.4853563,train_step,240,2,perf/gradient_accumulation_steps,8.0
323
+ 1774093111.4853563,train_step,240,2,system/cuda_memory_allocated_gb,17.816345691680908
324
+ 1774093111.4853563,train_step,240,2,system/cuda_max_memory_allocated_gb,82.1704511642456
325
+ 1774093192.88336,train_step,250,2,train/step_loss,0.8410789425556476
326
+ 1774093192.88336,train_step,250,2,train/step_real_loss,0.850802831351757
327
+ 1774093192.88336,train_step,250,2,train/lr,3.715559214503298e-05
328
+ 1774093192.88336,train_step,250,2,train/step_canary_loss,0.21875
329
+ 1774093192.88336,train_step,250,2,perf/step_duration_sec,8.232771871611476
330
+ 1774093192.88336,train_step,250,2,perf/samples_per_sec,7.895275250385015
331
+ 1774093192.88336,train_step,250,2,perf/tokens_per_sec,6654.745309889904
332
+ 1774093192.88336,train_step,250,2,perf/logical_batch_size,65.0
333
+ 1774093192.88336,train_step,250,2,perf/logical_token_count,54787.0
334
+ 1774093192.88336,train_step,250,2,perf/gradient_accumulation_steps,8.0
335
+ 1774093192.88336,train_step,250,2,system/cuda_memory_allocated_gb,17.816345691680908
336
+ 1774093192.88336,train_step,250,2,system/cuda_max_memory_allocated_gb,82.1704511642456
337
+ 1774093220.0732572,eval_step,250,2,eval/loss,0.748522541389383
338
+ 1774093220.0732572,eval_step,250,2,eval/duration_sec,27.187080297619104
339
+ 1774093302.2765276,train_step,260,2,train/step_loss,0.8724462010643699
340
+ 1774093302.2765276,train_step,260,2,train/step_real_loss,0.8488068357110023
341
+ 1774093302.2765276,train_step,260,2,train/lr,3.33379176277258e-05
342
+ 1774093302.2765276,train_step,260,2,train/step_canary_loss,1.62890625
343
+ 1774093302.2765276,train_step,260,2,perf/step_duration_sec,8.470736568793654
344
+ 1774093302.2765276,train_step,260,2,perf/samples_per_sec,7.7915302245551095
345
+ 1774093302.2765276,train_step,260,2,perf/tokens_per_sec,6277.49423773815
346
+ 1774093302.2765276,train_step,260,2,perf/logical_batch_size,66.0
347
+ 1774093302.2765276,train_step,260,2,perf/logical_token_count,53175.0
348
+ 1774093302.2765276,train_step,260,2,perf/gradient_accumulation_steps,8.0
349
+ 1774093302.2765276,train_step,260,2,system/cuda_memory_allocated_gb,17.911073684692383
350
+ 1774093302.2765276,train_step,260,2,system/cuda_max_memory_allocated_gb,82.1704511642456
351
+ 1774093384.391679,train_step,270,2,train/step_loss,0.8640658855438232
352
+ 1774093384.391679,train_step,270,2,train/step_real_loss,0.7926059737801552
353
+ 1774093384.391679,train_step,270,2,train/lr,2.962666050951997e-05
354
+ 1774093384.391679,train_step,270,2,train/step_canary_loss,5.4375
355
+ 1774093384.391679,train_step,270,2,perf/step_duration_sec,7.947357261553407
356
+ 1774093384.391679,train_step,270,2,perf/samples_per_sec,8.178819431516908
357
+ 1774093384.391679,train_step,270,2,perf/tokens_per_sec,6407.161314659091
358
+ 1774093384.391679,train_step,270,2,perf/logical_batch_size,65.0
359
+ 1774093384.391679,train_step,270,2,perf/logical_token_count,50920.0
360
+ 1774093384.391679,train_step,270,2,perf/gradient_accumulation_steps,8.0
361
+ 1774093384.391679,train_step,270,2,system/cuda_memory_allocated_gb,17.816345691680908
362
+ 1774093384.391679,train_step,270,2,system/cuda_max_memory_allocated_gb,82.1704511642456
363
+ 1774093468.0585487,train_step,280,2,train/step_loss,0.7910019425245431
364
+ 1774093468.0585487,train_step,280,2,train/step_real_loss,0.8018507286906242
365
+ 1774093468.0585487,train_step,280,2,train/lr,2.604552384991855e-05
366
+ 1774093468.0585487,train_step,280,2,train/step_canary_loss,0.0966796875
367
+ 1774093468.0585487,train_step,280,2,perf/step_duration_sec,8.62639987282455
368
+ 1774093468.0585487,train_step,280,2,perf/samples_per_sec,7.53500892124967
369
+ 1774093468.0585487,train_step,280,2,perf/tokens_per_sec,6275.039506402552
370
+ 1774093468.0585487,train_step,280,2,perf/logical_batch_size,65.0
371
+ 1774093468.0585487,train_step,280,2,perf/logical_token_count,54131.0
372
+ 1774093468.0585487,train_step,280,2,perf/gradient_accumulation_steps,8.0
373
+ 1774093468.0585487,train_step,280,2,system/cuda_memory_allocated_gb,17.816345691680908
374
+ 1774093468.0585487,train_step,280,2,system/cuda_max_memory_allocated_gb,82.1704511642456
375
+ 1774093550.7132335,train_step,290,2,train/step_loss,0.8430270507480159
376
+ 1774093550.7132335,train_step,290,2,train/step_real_loss,0.8613760396838188
377
+ 1774093550.7132335,train_step,290,2,train/lr,2.2617379654990623e-05
378
+ 1774093550.7132335,train_step,290,2,train/step_canary_loss,0.255859375
379
+ 1774093550.7132335,train_step,290,2,perf/step_duration_sec,8.335046991705894
380
+ 1774093550.7132335,train_step,290,2,perf/samples_per_sec,7.9183716739300705
381
+ 1774093550.7132335,train_step,290,2,perf/tokens_per_sec,6261.392413496007
382
+ 1774093550.7132335,train_step,290,2,perf/logical_batch_size,66.0
383
+ 1774093550.7132335,train_step,290,2,perf/logical_token_count,52189.0
384
+ 1774093550.7132335,train_step,290,2,perf/gradient_accumulation_steps,8.0
385
+ 1774093550.7132335,train_step,290,2,system/cuda_memory_allocated_gb,17.816345691680908
386
+ 1774093550.7132335,train_step,290,2,system/cuda_max_memory_allocated_gb,82.1704511642456
387
+ 1774093632.5524132,train_step,300,2,train/step_loss,0.7691241015087474
388
+ 1774093632.5524132,train_step,300,2,train/step_real_loss,0.784599058330059
389
+ 1774093632.5524132,train_step,300,2,train/lr,1.936412279842705e-05
390
+ 1774093632.5524132,train_step,300,2,train/step_canary_loss,0.27392578125
391
+ 1774093632.5524132,train_step,300,2,perf/step_duration_sec,8.091827008873224
392
+ 1774093632.5524132,train_step,300,2,perf/samples_per_sec,8.156378025336753
393
+ 1774093632.5524132,train_step,300,2,perf/tokens_per_sec,5984.062677922065
394
+ 1774093632.5524132,train_step,300,2,perf/logical_batch_size,66.0
395
+ 1774093632.5524132,train_step,300,2,perf/logical_token_count,48422.0
396
+ 1774093632.5524132,train_step,300,2,perf/gradient_accumulation_steps,8.0
397
+ 1774093632.5524132,train_step,300,2,system/cuda_memory_allocated_gb,17.911073684692383
398
+ 1774093632.5524132,train_step,300,2,system/cuda_max_memory_allocated_gb,82.1704511642456
399
+ 1774093659.719083,eval_step,300,2,eval/loss,0.7462242173737493
400
+ 1774093659.719083,eval_step,300,2,eval/duration_sec,27.16416385397315
401
+ 1774093742.5612013,train_step,310,2,train/step_loss,0.7514625552928809
402
+ 1774093742.5612013,train_step,310,2,train/step_real_loss,0.7710395082831383
403
+ 1774093742.5612013,train_step,310,2,train/lr,1.6306531183346385e-05
404
+ 1774093742.5612013,train_step,310,2,train/step_canary_loss,0.125
405
+ 1774093742.5612013,train_step,310,2,perf/step_duration_sec,8.240195842459798
406
+ 1774093742.5612013,train_step,310,2,perf/samples_per_sec,8.009518373327666
407
+ 1774093742.5612013,train_step,310,2,perf/tokens_per_sec,6369.751520897312
408
+ 1774093742.5612013,train_step,310,2,perf/logical_batch_size,66.0
409
+ 1774093742.5612013,train_step,310,2,perf/logical_token_count,52488.0
410
+ 1774093742.5612013,train_step,310,2,perf/gradient_accumulation_steps,8.0
411
+ 1774093742.5612013,train_step,310,2,system/cuda_memory_allocated_gb,17.816345691680908
412
+ 1774093742.5612013,train_step,310,2,system/cuda_max_memory_allocated_gb,82.1704511642456
413
+ 1774093824.76255,train_step,320,2,train/step_loss,0.8317307600608239
414
+ 1774093824.76255,train_step,320,2,train/step_real_loss,0.8416290208697319
415
+ 1774093824.76255,train_step,320,2,train/lr,1.3464133037968912e-05
416
+ 1774093824.76255,train_step,320,2,train/step_canary_loss,0.1982421875
417
+ 1774093824.76255,train_step,320,2,perf/step_duration_sec,7.953533122316003
418
+ 1774093824.76255,train_step,320,2,perf/samples_per_sec,8.172468637569782
419
+ 1774093824.76255,train_step,320,2,perf/tokens_per_sec,6226.666720107783
420
+ 1774093824.76255,train_step,320,2,perf/logical_batch_size,65.0
421
+ 1774093824.76255,train_step,320,2,perf/logical_token_count,49524.0
422
+ 1774093824.76255,train_step,320,2,perf/gradient_accumulation_steps,8.0
423
+ 1774093824.76255,train_step,320,2,system/cuda_memory_allocated_gb,17.816345691680908
424
+ 1774093824.76255,train_step,320,2,system/cuda_max_memory_allocated_gb,82.1704511642456
425
+ 1774093907.919859,train_step,330,2,train/step_loss,0.8719267258277306
426
+ 1774093907.919859,train_step,330,2,train/step_real_loss,0.8635779246687889
427
+ 1774093907.919859,train_step,330,2,train/lr,1.0855082192715294e-05
428
+ 1774093907.919859,train_step,330,2,train/step_canary_loss,1.40625
429
+ 1774093907.919859,train_step,330,2,perf/step_duration_sec,8.101611092686653
430
+ 1774093907.919859,train_step,330,2,perf/samples_per_sec,8.023095561656334
431
+ 1774093907.919859,train_step,330,2,perf/tokens_per_sec,5623.572827524039
432
+ 1774093907.919859,train_step,330,2,perf/logical_batch_size,65.0
433
+ 1774093907.919859,train_step,330,2,perf/logical_token_count,45560.0
434
+ 1774093907.919859,train_step,330,2,perf/gradient_accumulation_steps,8.0
435
+ 1774093907.919859,train_step,330,2,system/cuda_memory_allocated_gb,17.816345691680908
436
+ 1774093907.919859,train_step,330,2,system/cuda_max_memory_allocated_gb,82.1704511642456
437
+ 1774093989.7239172,train_step,340,2,train/step_loss,0.8476224361043988
438
+ 1774093989.7239172,train_step,340,2,train/step_real_loss,0.8228105828166008
439
+ 1774093989.7239172,train_step,340,2,train/lr,8.49604213531004e-06
440
+ 1774093989.7239172,train_step,340,2,train/step_canary_loss,1.6416015625
441
+ 1774093989.7239172,train_step,340,2,perf/step_duration_sec,8.274557610973716
442
+ 1774093989.7239172,train_step,340,2,perf/samples_per_sec,7.976257233676254
443
+ 1774093989.7239172,train_step,340,2,perf/tokens_per_sec,6674.435371234426
444
+ 1774093989.7239172,train_step,340,2,perf/logical_batch_size,66.0
445
+ 1774093989.7239172,train_step,340,2,perf/logical_token_count,55228.0
446
+ 1774093989.7239172,train_step,340,2,perf/gradient_accumulation_steps,8.0
447
+ 1774093989.7239172,train_step,340,2,system/cuda_memory_allocated_gb,17.816345691680908
448
+ 1774093989.7239172,train_step,340,2,system/cuda_max_memory_allocated_gb,82.1704511642456
449
+ 1774094071.9116077,train_step,350,2,train/step_loss,0.8573094416355741
450
+ 1774094071.9116077,train_step,350,2,train/step_real_loss,0.816773310303688
451
+ 1774094071.9116077,train_step,350,2,train/lr,6.402079584406673e-06
452
+ 1774094071.9116077,train_step,350,2,train/step_canary_loss,1.376171875
453
+ 1774094071.9116077,train_step,350,2,perf/step_duration_sec,8.637357847765088
454
+ 1774094071.9116077,train_step,350,2,perf/samples_per_sec,7.988554048140279
455
+ 1774094071.9116077,train_step,350,2,perf/tokens_per_sec,6212.547974249382
456
+ 1774094071.9116077,train_step,350,2,perf/logical_batch_size,69.0
457
+ 1774094071.9116077,train_step,350,2,perf/logical_token_count,53660.0
458
+ 1774094071.9116077,train_step,350,2,perf/gradient_accumulation_steps,8.0
459
+ 1774094071.9116077,train_step,350,2,system/cuda_memory_allocated_gb,17.816345691680908
460
+ 1774094071.9116077,train_step,350,2,system/cuda_max_memory_allocated_gb,82.1704511642456
461
+ 1774094099.091823,eval_step,350,2,eval/loss,0.7456388631942629
462
+ 1774094099.091823,eval_step,350,2,eval/duration_sec,27.17722495831549
463
+ 1774094181.5654633,train_step,360,2,train/step_loss,0.8446246770712046
464
+ 1774094181.5654633,train_step,360,2,train/step_real_loss,0.8453097268939018
465
+ 1774094181.5654633,train_step,360,2,train/lr,4.586568261458729e-06
466
+ 1774094181.5654633,train_step,360,2,train/step_canary_loss,0.80078125
467
+ 1774094181.5654633,train_step,360,2,perf/step_duration_sec,7.976690696552396
468
+ 1774094181.5654633,train_step,360,2,perf/samples_per_sec,8.148742689508275
469
+ 1774094181.5654633,train_step,360,2,perf/tokens_per_sec,6948.119478163339
470
+ 1774094181.5654633,train_step,360,2,perf/logical_batch_size,65.0
471
+ 1774094181.5654633,train_step,360,2,perf/logical_token_count,55423.0
472
+ 1774094181.5654633,train_step,360,2,perf/gradient_accumulation_steps,8.0
473
+ 1774094181.5654633,train_step,360,2,system/cuda_memory_allocated_gb,17.816345691680908
474
+ 1774094181.5654633,train_step,360,2,system/cuda_max_memory_allocated_gb,82.1704511642456
475
+ 1774094263.5117958,train_step,370,2,train/step_loss,0.8721330687403679
476
+ 1774094263.5117958,train_step,370,2,train/step_real_loss,0.8721330687403679
477
+ 1774094263.5117958,train_step,370,2,train/lr,3.06110347542643e-06
478
+ 1774094263.5117958,train_step,370,2,perf/step_duration_sec,8.37641635723412
479
+ 1774094263.5117958,train_step,370,2,perf/samples_per_sec,7.640498904370688
480
+ 1774094263.5117958,train_step,370,2,perf/tokens_per_sec,5830.058812420977
481
+ 1774094263.5117958,train_step,370,2,perf/logical_batch_size,64.0
482
+ 1774094263.5117958,train_step,370,2,perf/logical_token_count,48835.0
483
+ 1774094263.5117958,train_step,370,2,perf/gradient_accumulation_steps,8.0
484
+ 1774094263.5117958,train_step,370,2,system/cuda_memory_allocated_gb,17.816345691680908
485
+ 1774094263.5117958,train_step,370,2,system/cuda_max_memory_allocated_gb,82.1704511642456
486
+ 1774094346.1630964,train_step,380,2,train/step_loss,0.81929341004445
487
+ 1774094346.1630964,train_step,380,2,train/step_real_loss,0.8280055150389671
488
+ 1774094346.1630964,train_step,380,2,train/lr,1.8354280658494649e-06
489
+ 1774094346.1630964,train_step,380,2,train/step_canary_loss,0.26171875
490
+ 1774094346.1630964,train_step,380,2,perf/step_duration_sec,7.9715049508959055
491
+ 1774094346.1630964,train_step,380,2,perf/samples_per_sec,8.154043734576712
492
+ 1774094346.1630964,train_step,380,2,perf/tokens_per_sec,6787.676898315027
493
+ 1774094346.1630964,train_step,380,2,perf/logical_batch_size,65.0
494
+ 1774094346.1630964,train_step,380,2,perf/logical_token_count,54108.0
495
+ 1774094346.1630964,train_step,380,2,perf/gradient_accumulation_steps,8.0
496
+ 1774094346.1630964,train_step,380,2,system/cuda_memory_allocated_gb,17.816345691680908
497
+ 1774094346.1630964,train_step,380,2,system/cuda_max_memory_allocated_gb,82.1704511642456
498
+ 1774094429.2324018,train_step,390,2,train/step_loss,0.8554338663816452
499
+ 1774094429.2324018,train_step,390,2,train/step_real_loss,0.8554338663816452
500
+ 1774094429.2324018,train_step,390,2,train/lr,9.17370177272775e-07
501
+ 1774094429.2324018,train_step,390,2,perf/step_duration_sec,8.24701151996851
502
+ 1774094429.2324018,train_step,390,2,perf/samples_per_sec,7.760386880148844
503
+ 1774094429.2324018,train_step,390,2,perf/tokens_per_sec,6142.34621563781
504
+ 1774094429.2324018,train_step,390,2,perf/logical_batch_size,64.0
505
+ 1774094429.2324018,train_step,390,2,perf/logical_token_count,50656.0
506
+ 1774094429.2324018,train_step,390,2,perf/gradient_accumulation_steps,8.0
507
+ 1774094429.2324018,train_step,390,2,system/cuda_memory_allocated_gb,17.816345691680908
508
+ 1774094429.2324018,train_step,390,2,system/cuda_max_memory_allocated_gb,82.1704511642456
509
+ 1774094512.2904398,train_step,400,2,train/step_loss,0.7998756021261215
510
+ 1774094512.2904398,train_step,400,2,train/step_real_loss,0.7998756021261215
511
+ 1774094512.2904398,train_step,400,2,train/lr,3.127932624475638e-07
512
+ 1774094512.2904398,train_step,400,2,perf/step_duration_sec,7.943094424903393
513
+ 1774094512.2904398,train_step,400,2,perf/samples_per_sec,8.0573132555677
514
+ 1774094512.2904398,train_step,400,2,perf/tokens_per_sec,6869.86671452841
515
+ 1774094512.2904398,train_step,400,2,perf/logical_batch_size,64.0
516
+ 1774094512.2904398,train_step,400,2,perf/logical_token_count,54568.0
517
+ 1774094512.2904398,train_step,400,2,perf/gradient_accumulation_steps,8.0
518
+ 1774094512.2904398,train_step,400,2,system/cuda_memory_allocated_gb,17.816345691680908
519
+ 1774094512.2904398,train_step,400,2,system/cuda_max_memory_allocated_gb,82.1704511642456
520
+ 1774094539.4831598,eval_step,400,2,eval/loss,0.7455017728538349
521
+ 1774094539.4831598,eval_step,400,2,eval/duration_sec,27.190383709967136
522
+ 1774094621.4459202,train_step,410,2,train/step_loss,0.854436350591255
523
+ 1774094621.4459202,train_step,410,2,train/step_real_loss,0.8517185375094414
524
+ 1774094621.4459202,train_step,410,2,train/lr,2.5558633627303928e-08
525
+ 1774094621.4459202,train_step,410,2,train/step_canary_loss,0.94140625
526
+ 1774094621.4459202,train_step,410,2,perf/step_duration_sec,8.378037076443434
527
+ 1774094621.4459202,train_step,410,2,perf/samples_per_sec,7.877740262760654
528
+ 1774094621.4459202,train_step,410,2,perf/tokens_per_sec,6647.619184760483
529
+ 1774094621.4459202,train_step,410,2,perf/logical_batch_size,66.0
530
+ 1774094621.4459202,train_step,410,2,perf/logical_token_count,55694.0
531
+ 1774094621.4459202,train_step,410,2,perf/gradient_accumulation_steps,8.0
532
+ 1774094621.4459202,train_step,410,2,system/cuda_memory_allocated_gb,17.816345691680908
533
+ 1774094621.4459202,train_step,410,2,system/cuda_max_memory_allocated_gb,88.38552379608154
534
+ 1774094682.1799755,train_epoch,414,2,train/epoch_loss,0.840315752633972
535
+ 1774094682.1799755,train_epoch,414,2,train/epoch_real_loss,0.8362452851041504
536
+ 1774094682.1799755,train_epoch,414,2,train/epoch_canary_loss,1.0832555509625612
537
+ 1774094682.1799755,train_epoch,414,2,perf/epoch_duration_sec,1814.1021996028721
538
+ 1774094682.1799755,train_epoch,414,2,perf/epoch_samples_per_sec,29.765136722628178
539
+ 1774094682.1799755,train_epoch,414,2,perf/epoch_tokens_per_sec,24186.19579955584
540
+ 1774094682.1799755,train_epoch,414,2,perf/epoch_samples,53997.0
541
+ 1774094682.1799755,train_epoch,414,2,perf/epoch_tokens,43876231.0
542
+ 1774094682.1799755,train_epoch,414,2,system/cuda_epoch_peak_memory_gb,88.38552379608154
543
+ 1774094682.1799755,train_epoch,414,2,eval/loss,0.7454980848164394
544
+ 1774094682.1799755,train_epoch,414,2,eval/duration_sec,27.13153049722314
545
+ 1774094693.161264,audit_epoch,414,2,audit/delta,1e-05
546
+ 1774094693.161264,audit_epoch,414,2,audit/num_canaries,500.0
547
+ 1774094693.161264,audit_epoch,414,2,audit/num_members,250.0
548
+ 1774094693.161264,audit_epoch,414,2,audit/paper_guess_fraction,0.2
549
+ 1774094693.161264,audit_epoch,414,2,audit/paper_guess_steps,20.0
550
+ 1774094693.161264,audit_epoch,414,2,audit/loss/auc,1.0
551
+ 1774094693.161264,audit_epoch,414,2,audit/loss/empirical_epsilon/0.05,3.4791953936219215
552
+ 1774094693.161264,audit_epoch,414,2,audit/loss/empirical_epsilon/0.01,3.023197554051876
553
+ 1774094693.161264,audit_epoch,414,2,audit/loss/empirical_epsilon_details/0.05/epsilon,3.4791953936219215
554
+ 1774094693.161264,audit_epoch,414,2,audit/loss/empirical_epsilon_details/0.05/num_guesses,100.0
555
+ 1774094693.161264,audit_epoch,414,2,audit/loss/empirical_epsilon_details/0.05/correct_guesses,100.0
556
+ 1774094693.161264,audit_epoch,414,2,audit/loss/empirical_epsilon_details/0.01/epsilon,3.023197554051876
557
+ 1774094693.161264,audit_epoch,414,2,audit/loss/empirical_epsilon_details/0.01/num_guesses,100.0
558
+ 1774094693.161264,audit_epoch,414,2,audit/loss/empirical_epsilon_details/0.01/correct_guesses,100.0
559
+ 1774094693.161264,audit_epoch,414,2,audit/embedding/auc,0.916224
560
+ 1774094693.161264,audit_epoch,414,2,audit/embedding/empirical_epsilon/0.05,3.4791953936219215
561
+ 1774094693.161264,audit_epoch,414,2,audit/embedding/empirical_epsilon/0.01,3.023197554051876
562
+ 1774094693.161264,audit_epoch,414,2,audit/embedding/empirical_epsilon_details/0.05/epsilon,3.4791953936219215
563
+ 1774094693.161264,audit_epoch,414,2,audit/embedding/empirical_epsilon_details/0.05/num_guesses,100.0
564
+ 1774094693.161264,audit_epoch,414,2,audit/embedding/empirical_epsilon_details/0.05/correct_guesses,100.0
565
+ 1774094693.161264,audit_epoch,414,2,audit/embedding/empirical_epsilon_details/0.01/epsilon,3.023197554051876
566
+ 1774094693.161264,audit_epoch,414,2,audit/embedding/empirical_epsilon_details/0.01/num_guesses,100.0
567
+ 1774094693.161264,audit_epoch,414,2,audit/embedding/empirical_epsilon_details/0.01/correct_guesses,100.0
568
+ 1774094693.161264,audit_epoch,414,2,perf/audit_duration_sec,6.722778998315334
569
+ 1774094704.3854914,audit_final,414,2,audit/delta,1e-05
570
+ 1774094704.3854914,audit_final,414,2,audit/num_canaries,500.0
571
+ 1774094704.3854914,audit_final,414,2,audit/num_members,250.0
572
+ 1774094704.3854914,audit_final,414,2,audit/paper_guess_fraction,0.2
573
+ 1774094704.3854914,audit_final,414,2,audit/paper_guess_steps,20.0
574
+ 1774094704.3854914,audit_final,414,2,audit/loss/auc,1.0
575
+ 1774094704.3854914,audit_final,414,2,audit/loss/empirical_epsilon/0.05,3.4791953936219215
576
+ 1774094704.3854914,audit_final,414,2,audit/loss/empirical_epsilon/0.01,3.023197554051876
577
+ 1774094704.3854914,audit_final,414,2,audit/loss/empirical_epsilon_details/0.05/epsilon,3.4791953936219215
578
+ 1774094704.3854914,audit_final,414,2,audit/loss/empirical_epsilon_details/0.05/num_guesses,100.0
579
+ 1774094704.3854914,audit_final,414,2,audit/loss/empirical_epsilon_details/0.05/correct_guesses,100.0
580
+ 1774094704.3854914,audit_final,414,2,audit/loss/empirical_epsilon_details/0.01/epsilon,3.023197554051876
581
+ 1774094704.3854914,audit_final,414,2,audit/loss/empirical_epsilon_details/0.01/num_guesses,100.0
582
+ 1774094704.3854914,audit_final,414,2,audit/loss/empirical_epsilon_details/0.01/correct_guesses,100.0
583
+ 1774094704.3854914,audit_final,414,2,audit/embedding/auc,0.916224
584
+ 1774094704.3854914,audit_final,414,2,audit/embedding/empirical_epsilon/0.05,3.4791953936219215
585
+ 1774094704.3854914,audit_final,414,2,audit/embedding/empirical_epsilon/0.01,3.023197554051876
586
+ 1774094704.3854914,audit_final,414,2,audit/embedding/empirical_epsilon_details/0.05/epsilon,3.4791953936219215
587
+ 1774094704.3854914,audit_final,414,2,audit/embedding/empirical_epsilon_details/0.05/num_guesses,100.0
588
+ 1774094704.3854914,audit_final,414,2,audit/embedding/empirical_epsilon_details/0.05/correct_guesses,100.0
589
+ 1774094704.3854914,audit_final,414,2,audit/embedding/empirical_epsilon_details/0.01/epsilon,3.023197554051876
590
+ 1774094704.3854914,audit_final,414,2,audit/embedding/empirical_epsilon_details/0.01/num_guesses,100.0
591
+ 1774094704.3854914,audit_final,414,2,audit/embedding/empirical_epsilon_details/0.01/correct_guesses,100.0
592
+ 1774094704.914049,energy_final,414,,energy/codecarbon/duration,3820.65162669681
593
+ 1774094704.914049,energy_final,414,,energy/codecarbon/emissions,0.5568149811716148
594
+ 1774094704.914049,energy_final,414,,energy/codecarbon/emissions_rate,0.0001457382236267942
595
+ 1774094704.914049,energy_final,414,,energy/codecarbon/cpu_power,179.3439937196561
596
+ 1774094704.914049,energy_final,414,,energy/codecarbon/gpu_power,2316.658494927477
597
+ 1774094704.914049,energy_final,414,,energy/codecarbon/ram_power,70.0
598
+ 1774094704.914049,energy_final,414,,energy/codecarbon/cpu_energy,0.18368298514637418
599
+ 1774094704.914049,energy_final,414,,energy/codecarbon/gpu_energy,2.457235179675422
600
+ 1774094704.914049,energy_final,414,,energy/codecarbon/ram_energy,0.07169108754482326
601
+ 1774094704.914049,energy_final,414,,energy/codecarbon/energy_consumed,2.7126092523666183
602
+ 1774094704.914049,energy_final,414,,energy/codecarbon/water_consumed,0.0
603
+ 1774094704.914049,energy_final,414,,energy/codecarbon/cpu_count,224.0
604
+ 1774094704.914049,energy_final,414,,energy/codecarbon/gpu_count,4.0
605
+ 1774094704.914049,energy_final,414,,energy/codecarbon/longitude,-121.9552
606
+ 1774094704.914049,energy_final,414,,energy/codecarbon/latitude,37.3541
607
+ 1774094704.914049,energy_final,414,,energy/codecarbon/ram_total_size,2015.5625190734863
608
+ 1774094704.914049,energy_final,414,,energy/codecarbon/cpu_utilization_percent,2.2497114375655825
609
+ 1774094704.914049,energy_final,414,,energy/codecarbon/gpu_utilization_percent,96.45330535152151
610
+ 1774094704.914049,energy_final,414,,energy/codecarbon/ram_utilization_percent,2.3057974816369358
611
+ 1774094704.914049,energy_final,414,,energy/codecarbon/ram_used_gb,46.78091819539025
612
+ 1774094704.914049,energy_final,414,,energy/codecarbon/pue,1.0
613
+ 1774094704.914049,energy_final,414,,energy/codecarbon/wue,0.0
starcoder2-7b/base/summary.json ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "audit/delta": 1e-05,
3
+ "audit/embedding/auc": 0.916224,
4
+ "audit/embedding/empirical_epsilon/0.01": 3.023197554051876,
5
+ "audit/embedding/empirical_epsilon/0.05": 3.4791953936219215,
6
+ "audit/embedding/empirical_epsilon_details/0.01/correct_guesses": 100.0,
7
+ "audit/embedding/empirical_epsilon_details/0.01/epsilon": 3.023197554051876,
8
+ "audit/embedding/empirical_epsilon_details/0.01/num_guesses": 100.0,
9
+ "audit/embedding/empirical_epsilon_details/0.05/correct_guesses": 100.0,
10
+ "audit/embedding/empirical_epsilon_details/0.05/epsilon": 3.4791953936219215,
11
+ "audit/embedding/empirical_epsilon_details/0.05/num_guesses": 100.0,
12
+ "audit/loss/auc": 1.0,
13
+ "audit/loss/empirical_epsilon/0.01": 3.023197554051876,
14
+ "audit/loss/empirical_epsilon/0.05": 3.4791953936219215,
15
+ "audit/loss/empirical_epsilon_details/0.01/correct_guesses": 100.0,
16
+ "audit/loss/empirical_epsilon_details/0.01/epsilon": 3.023197554051876,
17
+ "audit/loss/empirical_epsilon_details/0.01/num_guesses": 100.0,
18
+ "audit/loss/empirical_epsilon_details/0.05/correct_guesses": 100.0,
19
+ "audit/loss/empirical_epsilon_details/0.05/epsilon": 3.4791953936219215,
20
+ "audit/loss/empirical_epsilon_details/0.05/num_guesses": 100.0,
21
+ "audit/num_canaries": 500.0,
22
+ "audit/num_members": 250.0,
23
+ "audit/paper_guess_fraction": 0.2,
24
+ "audit/paper_guess_steps": 20.0,
25
+ "energy/codecarbon/cpu_count": 224.0,
26
+ "energy/codecarbon/cpu_energy": 0.18368298514637418,
27
+ "energy/codecarbon/cpu_power": 179.3439937196561,
28
+ "energy/codecarbon/cpu_utilization_percent": 2.2497114375655825,
29
+ "energy/codecarbon/duration": 3820.65162669681,
30
+ "energy/codecarbon/emissions": 0.5568149811716148,
31
+ "energy/codecarbon/emissions_rate": 0.0001457382236267942,
32
+ "energy/codecarbon/energy_consumed": 2.7126092523666183,
33
+ "energy/codecarbon/gpu_count": 4.0,
34
+ "energy/codecarbon/gpu_energy": 2.457235179675422,
35
+ "energy/codecarbon/gpu_power": 2316.658494927477,
36
+ "energy/codecarbon/gpu_utilization_percent": 96.45330535152151,
37
+ "energy/codecarbon/latitude": 37.3541,
38
+ "energy/codecarbon/longitude": -121.9552,
39
+ "energy/codecarbon/pue": 1.0,
40
+ "energy/codecarbon/ram_energy": 0.07169108754482326,
41
+ "energy/codecarbon/ram_power": 70.0,
42
+ "energy/codecarbon/ram_total_size": 2015.5625190734863,
43
+ "energy/codecarbon/ram_used_gb": 46.78091819539025,
44
+ "energy/codecarbon/ram_utilization_percent": 2.3057974816369358,
45
+ "energy/codecarbon/water_consumed": 0.0,
46
+ "energy/codecarbon/wue": 0.0,
47
+ "eval/duration_sec": 27.13153049722314,
48
+ "eval/loss": 0.7454980848164394,
49
+ "perf/audit_duration_sec": 6.722778998315334,
50
+ "perf/epoch_duration_sec": 1814.1021996028721,
51
+ "perf/epoch_samples": 53997.0,
52
+ "perf/epoch_samples_per_sec": 29.765136722628178,
53
+ "perf/epoch_tokens": 43876231.0,
54
+ "perf/epoch_tokens_per_sec": 24186.19579955584,
55
+ "perf/gradient_accumulation_steps": 8.0,
56
+ "perf/logical_batch_size": 66.0,
57
+ "perf/logical_token_count": 55694.0,
58
+ "perf/samples_per_sec": 7.877740262760654,
59
+ "perf/step_duration_sec": 8.378037076443434,
60
+ "perf/tokens_per_sec": 6647.619184760483,
61
+ "system/cuda_epoch_peak_memory_gb": 88.38552379608154,
62
+ "system/cuda_max_memory_allocated_gb": 88.38552379608154,
63
+ "system/cuda_memory_allocated_gb": 17.816345691680908,
64
+ "train/epoch_canary_loss": 1.0832555509625612,
65
+ "train/epoch_loss": 0.840315752633972,
66
+ "train/epoch_real_loss": 0.8362452851041504,
67
+ "train/lr": 2.5558633627303928e-08,
68
+ "train/step_canary_loss": 0.94140625,
69
+ "train/step_loss": 0.854436350591255,
70
+ "train/step_real_loss": 0.8517185375094414
71
+ }
starcoder2-7b/base/tensorboard/events.out.tfevents.1774090870.364c4f8de9dd.6145.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da29212bc469bf3917a2aa87b8ff1081b72b125dbbaefea65a979c0832e71373
3
+ size 41006
starcoder2-7b/base/tokenizer/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
starcoder2-7b/base/tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,516 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "backend": "tokenizers",
4
+ "bos_token": "<|endoftext|>",
5
+ "clean_up_tokenization_spaces": true,
6
+ "eos_token": "<|endoftext|>",
7
+ "errors": "replace",
8
+ "extra_special_tokens": {
9
+ "extra_special_token_0": "865331112869",
10
+ "extra_special_token_1": "569765693871",
11
+ "extra_special_token_2": "485177821815",
12
+ "extra_special_token_3": "135441121756",
13
+ "extra_special_token_4": "367459894796",
14
+ "extra_special_token_5": "877482678543",
15
+ "extra_special_token_6": "457919547633",
16
+ "extra_special_token_7": "765474393376",
17
+ "extra_special_token_8": "114848338811",
18
+ "extra_special_token_9": "746285987371",
19
+ "extra_special_token_10": "649291669397",
20
+ "extra_special_token_11": "927914615679",
21
+ "extra_special_token_12": "445925149649",
22
+ "extra_special_token_13": "691587454538",
23
+ "extra_special_token_14": "143777992227",
24
+ "extra_special_token_15": "997981281989",
25
+ "extra_special_token_16": "425949483533",
26
+ "extra_special_token_17": "982993456429",
27
+ "extra_special_token_18": "718726519731",
28
+ "extra_special_token_19": "172599315861",
29
+ "extra_special_token_20": "643489267333",
30
+ "extra_special_token_21": "282322838685",
31
+ "extra_special_token_22": "781653545886",
32
+ "extra_special_token_23": "796415361892",
33
+ "extra_special_token_24": "841991688488",
34
+ "extra_special_token_25": "211411365397",
35
+ "extra_special_token_26": "698218415444",
36
+ "extra_special_token_27": "355977139358",
37
+ "extra_special_token_28": "682564697312",
38
+ "extra_special_token_29": "383837596997",
39
+ "extra_special_token_30": "689362171782",
40
+ "extra_special_token_31": "749966767285",
41
+ "extra_special_token_32": "753159165157",
42
+ "extra_special_token_33": "795693824762",
43
+ "extra_special_token_34": "669689115557",
44
+ "extra_special_token_35": "327491773134",
45
+ "extra_special_token_36": "983569279932",
46
+ "extra_special_token_37": "612128769512",
47
+ "extra_special_token_38": "374327157578",
48
+ "extra_special_token_39": "311632789559",
49
+ "extra_special_token_40": "523918658846",
50
+ "extra_special_token_41": "765981581453",
51
+ "extra_special_token_42": "794825141891",
52
+ "extra_special_token_43": "873898736873",
53
+ "extra_special_token_44": "447445629421",
54
+ "extra_special_token_45": "473822473819",
55
+ "extra_special_token_46": "181439694557",
56
+ "extra_special_token_47": "592538279337",
57
+ "extra_special_token_48": "668134915514",
58
+ "extra_special_token_49": "643692393748",
59
+ "extra_special_token_50": "696651276628",
60
+ "extra_special_token_51": "853859348234",
61
+ "extra_special_token_52": "778466723723",
62
+ "extra_special_token_53": "929826356991",
63
+ "extra_special_token_54": "272362973463",
64
+ "extra_special_token_55": "694235616268",
65
+ "extra_special_token_56": "281673864127",
66
+ "extra_special_token_57": "479676316326",
67
+ "extra_special_token_58": "646979124677",
68
+ "extra_special_token_59": "922327493433",
69
+ "extra_special_token_60": "883685933161",
70
+ "extra_special_token_61": "264259917554",
71
+ "extra_special_token_62": "836746273134",
72
+ "extra_special_token_63": "658481324922",
73
+ "extra_special_token_64": "481884157827",
74
+ "extra_special_token_65": "587787496812",
75
+ "extra_special_token_66": "579184949249",
76
+ "extra_special_token_67": "912193598348",
77
+ "extra_special_token_68": "529679678956",
78
+ "extra_special_token_69": "795838284624",
79
+ "extra_special_token_70": "159337222655",
80
+ "extra_special_token_71": "173781362446",
81
+ "extra_special_token_72": "773687856563",
82
+ "extra_special_token_73": "535787224917",
83
+ "extra_special_token_74": "351885857332",
84
+ "extra_special_token_75": "578827344666",
85
+ "extra_special_token_76": "198462689911",
86
+ "extra_special_token_77": "722618266242",
87
+ "extra_special_token_78": "952872416512",
88
+ "extra_special_token_79": "517778845323",
89
+ "extra_special_token_80": "749665846687",
90
+ "extra_special_token_81": "661436365453",
91
+ "extra_special_token_82": "259666844669",
92
+ "extra_special_token_83": "242851284913",
93
+ "extra_special_token_84": "514532995959",
94
+ "extra_special_token_85": "161588262349",
95
+ "extra_special_token_86": "742765629356",
96
+ "extra_special_token_87": "225164373623",
97
+ "extra_special_token_88": "676539973863",
98
+ "extra_special_token_89": "826214551218",
99
+ "extra_special_token_90": "182345464792",
100
+ "extra_special_token_91": "232776999554",
101
+ "extra_special_token_92": "337326533813",
102
+ "extra_special_token_93": "676676697292",
103
+ "extra_special_token_94": "929185622831",
104
+ "extra_special_token_95": "545512344383",
105
+ "extra_special_token_96": "499444466686",
106
+ "extra_special_token_97": "314697386682",
107
+ "extra_special_token_98": "517379856925",
108
+ "extra_special_token_99": "379557332953",
109
+ "extra_special_token_100": "614797267726",
110
+ "extra_special_token_101": "429781429464",
111
+ "extra_special_token_102": "922466849763",
112
+ "extra_special_token_103": "721737645236",
113
+ "extra_special_token_104": "479227349997",
114
+ "extra_special_token_105": "136931728327",
115
+ "extra_special_token_106": "259533577263",
116
+ "extra_special_token_107": "488538864842",
117
+ "extra_special_token_108": "937495658852",
118
+ "extra_special_token_109": "489991411364",
119
+ "extra_special_token_110": "499148455254",
120
+ "extra_special_token_111": "441373944925",
121
+ "extra_special_token_112": "899151413682",
122
+ "extra_special_token_113": "467893531755",
123
+ "extra_special_token_114": "527117488925",
124
+ "extra_special_token_115": "928335588653",
125
+ "extra_special_token_116": "374439448821",
126
+ "extra_special_token_117": "879425227932",
127
+ "extra_special_token_118": "867678158885",
128
+ "extra_special_token_119": "399749397872",
129
+ "extra_special_token_120": "129693547287",
130
+ "extra_special_token_121": "689285841825",
131
+ "extra_special_token_122": "771619544974",
132
+ "extra_special_token_123": "724883568652",
133
+ "extra_special_token_124": "516968424863",
134
+ "extra_special_token_125": "733737988257",
135
+ "extra_special_token_126": "852347289392",
136
+ "extra_special_token_127": "296953381169",
137
+ "extra_special_token_128": "377273562477",
138
+ "extra_special_token_129": "262296912232",
139
+ "extra_special_token_130": "547149832394",
140
+ "extra_special_token_131": "298464134954",
141
+ "extra_special_token_132": "216667245274",
142
+ "extra_special_token_133": "843998562287",
143
+ "extra_special_token_134": "572154333646",
144
+ "extra_special_token_135": "124589118494",
145
+ "extra_special_token_136": "841824384614",
146
+ "extra_special_token_137": "232896526252",
147
+ "extra_special_token_138": "295448593321",
148
+ "extra_special_token_139": "123741461297",
149
+ "extra_special_token_140": "653573457168",
150
+ "extra_special_token_141": "196735786156",
151
+ "extra_special_token_142": "377338713663",
152
+ "extra_special_token_143": "964342468552",
153
+ "extra_special_token_144": "586855179568",
154
+ "extra_special_token_145": "484773717614",
155
+ "extra_special_token_146": "894885246797",
156
+ "extra_special_token_147": "677896358599",
157
+ "extra_special_token_148": "848845611563",
158
+ "extra_special_token_149": "851852651677",
159
+ "extra_special_token_150": "398549545767",
160
+ "extra_special_token_151": "454244839926",
161
+ "extra_special_token_152": "799364566435",
162
+ "extra_special_token_153": "967114116556",
163
+ "extra_special_token_154": "817378986438",
164
+ "extra_special_token_155": "233795848681",
165
+ "extra_special_token_156": "824387273757",
166
+ "extra_special_token_157": "916198946615",
167
+ "extra_special_token_158": "563117729724",
168
+ "extra_special_token_159": "951794811935",
169
+ "extra_special_token_160": "374598961236",
170
+ "extra_special_token_161": "922867396683",
171
+ "extra_special_token_162": "765737843639",
172
+ "extra_special_token_163": "175469284871",
173
+ "extra_special_token_164": "231853711778",
174
+ "extra_special_token_165": "662426712668",
175
+ "extra_special_token_166": "711412347158",
176
+ "extra_special_token_167": "753466987363",
177
+ "extra_special_token_168": "513361312532",
178
+ "extra_special_token_169": "712992815957",
179
+ "extra_special_token_170": "971621888444",
180
+ "extra_special_token_171": "829235161526",
181
+ "extra_special_token_172": "585544633356",
182
+ "extra_special_token_173": "582471228164",
183
+ "extra_special_token_174": "678666359123",
184
+ "extra_special_token_175": "557533689478",
185
+ "extra_special_token_176": "632962475133",
186
+ "extra_special_token_177": "484489193824",
187
+ "extra_special_token_178": "489562189822",
188
+ "extra_special_token_179": "589547936288",
189
+ "extra_special_token_180": "363214487524",
190
+ "extra_special_token_181": "244885399387",
191
+ "extra_special_token_182": "431751228368",
192
+ "extra_special_token_183": "433581868192",
193
+ "extra_special_token_184": "486391569221",
194
+ "extra_special_token_185": "185438575221",
195
+ "extra_special_token_186": "126574388585",
196
+ "extra_special_token_187": "741757479784",
197
+ "extra_special_token_188": "529854679937",
198
+ "extra_special_token_189": "996116119839",
199
+ "extra_special_token_190": "616248973917",
200
+ "extra_special_token_191": "763531783491",
201
+ "extra_special_token_192": "955456118295",
202
+ "extra_special_token_193": "364196983365",
203
+ "extra_special_token_194": "195792996468",
204
+ "extra_special_token_195": "151859598873",
205
+ "extra_special_token_196": "399223169721",
206
+ "extra_special_token_197": "938488813964",
207
+ "extra_special_token_198": "961981959227",
208
+ "extra_special_token_199": "183368827562",
209
+ "extra_special_token_200": "533417736566",
210
+ "extra_special_token_201": "786391632558",
211
+ "extra_special_token_202": "665661658354",
212
+ "extra_special_token_203": "693281533643",
213
+ "extra_special_token_204": "475794684356",
214
+ "extra_special_token_205": "652154162978",
215
+ "extra_special_token_206": "753233719644",
216
+ "extra_special_token_207": "668514843129",
217
+ "extra_special_token_208": "819162623892",
218
+ "extra_special_token_209": "941169431859",
219
+ "extra_special_token_210": "877385381798",
220
+ "extra_special_token_211": "752644929761",
221
+ "extra_special_token_212": "881136466196",
222
+ "extra_special_token_213": "275597777299",
223
+ "extra_special_token_214": "731681792655",
224
+ "extra_special_token_215": "961133895172",
225
+ "extra_special_token_216": "864718285734",
226
+ "extra_special_token_217": "963852916563",
227
+ "extra_special_token_218": "319584985416",
228
+ "extra_special_token_219": "563365646341",
229
+ "extra_special_token_220": "811371928234",
230
+ "extra_special_token_221": "837131396371",
231
+ "extra_special_token_222": "267514771964",
232
+ "extra_special_token_223": "944513428457",
233
+ "extra_special_token_224": "117298239631",
234
+ "extra_special_token_225": "158142752582",
235
+ "extra_special_token_226": "252867443568",
236
+ "extra_special_token_227": "839269684865",
237
+ "extra_special_token_228": "612788593128",
238
+ "extra_special_token_229": "145669731981",
239
+ "extra_special_token_230": "121557291859",
240
+ "extra_special_token_231": "245416776926",
241
+ "extra_special_token_232": "799417897197",
242
+ "extra_special_token_233": "997958836435",
243
+ "extra_special_token_234": "892336777248",
244
+ "extra_special_token_235": "158929292238",
245
+ "extra_special_token_236": "581976444672",
246
+ "extra_special_token_237": "897784492783",
247
+ "extra_special_token_238": "492373714791",
248
+ "extra_special_token_239": "512659818733",
249
+ "extra_special_token_240": "881112998642",
250
+ "extra_special_token_241": "619454958782",
251
+ "extra_special_token_242": "431149748713",
252
+ "extra_special_token_243": "624221476921",
253
+ "extra_special_token_244": "125866399464",
254
+ "extra_special_token_245": "339882449689",
255
+ "extra_special_token_246": "186198784585",
256
+ "extra_special_token_247": "943193294691",
257
+ "extra_special_token_248": "955668961269",
258
+ "extra_special_token_249": "232787996724",
259
+ "extra_special_token_250": "215671314196",
260
+ "extra_special_token_251": "286173241916",
261
+ "extra_special_token_252": "745977673725",
262
+ "extra_special_token_253": "556976448182",
263
+ "extra_special_token_254": "599961512792",
264
+ "extra_special_token_255": "766294538337",
265
+ "extra_special_token_256": "934912591213",
266
+ "extra_special_token_257": "295118729589",
267
+ "extra_special_token_258": "529455466433",
268
+ "extra_special_token_259": "196119929397",
269
+ "extra_special_token_260": "379571934299",
270
+ "extra_special_token_261": "251789649997",
271
+ "extra_special_token_262": "564544131355",
272
+ "extra_special_token_263": "244371196654",
273
+ "extra_special_token_264": "384598329253",
274
+ "extra_special_token_265": "887753195844",
275
+ "extra_special_token_266": "364947325679",
276
+ "extra_special_token_267": "655517954651",
277
+ "extra_special_token_268": "673948786567",
278
+ "extra_special_token_269": "857231548835",
279
+ "extra_special_token_270": "816115936673",
280
+ "extra_special_token_271": "644234165531",
281
+ "extra_special_token_272": "182782912224",
282
+ "extra_special_token_273": "234316622259",
283
+ "extra_special_token_274": "421369185549",
284
+ "extra_special_token_275": "434632855397",
285
+ "extra_special_token_276": "921889371893",
286
+ "extra_special_token_277": "415956914763",
287
+ "extra_special_token_278": "598916996413",
288
+ "extra_special_token_279": "773671349113",
289
+ "extra_special_token_280": "952465217972",
290
+ "extra_special_token_281": "117657531962",
291
+ "extra_special_token_282": "729825168745",
292
+ "extra_special_token_283": "691315125346",
293
+ "extra_special_token_284": "768461952319",
294
+ "extra_special_token_285": "664847713559",
295
+ "extra_special_token_286": "953267689786",
296
+ "extra_special_token_287": "886464195129",
297
+ "extra_special_token_288": "824488329416",
298
+ "extra_special_token_289": "837873762491",
299
+ "extra_special_token_290": "532833541879",
300
+ "extra_special_token_291": "669183782449",
301
+ "extra_special_token_292": "941976537588",
302
+ "extra_special_token_293": "739394546916",
303
+ "extra_special_token_294": "267954879268",
304
+ "extra_special_token_295": "637551427887",
305
+ "extra_special_token_296": "217756494954",
306
+ "extra_special_token_297": "524444658383",
307
+ "extra_special_token_298": "117783274348",
308
+ "extra_special_token_299": "138218735276",
309
+ "extra_special_token_300": "814611949491",
310
+ "extra_special_token_301": "711641973413",
311
+ "extra_special_token_302": "499156317423",
312
+ "extra_special_token_303": "515856611931",
313
+ "extra_special_token_304": "454164859837",
314
+ "extra_special_token_305": "345271433112",
315
+ "extra_special_token_306": "462294118988",
316
+ "extra_special_token_307": "511785788222",
317
+ "extra_special_token_308": "497294727353",
318
+ "extra_special_token_309": "866519986723",
319
+ "extra_special_token_310": "334513529294",
320
+ "extra_special_token_311": "549946382131",
321
+ "extra_special_token_312": "284445431422",
322
+ "extra_special_token_313": "396521188476",
323
+ "extra_special_token_314": "421435255895",
324
+ "extra_special_token_315": "133373659361",
325
+ "extra_special_token_316": "322683334381",
326
+ "extra_special_token_317": "228358422847",
327
+ "extra_special_token_318": "291762694874",
328
+ "extra_special_token_319": "143182978129",
329
+ "extra_special_token_320": "511923256573",
330
+ "extra_special_token_321": "327158398268",
331
+ "extra_special_token_322": "879764613759",
332
+ "extra_special_token_323": "564395222747",
333
+ "extra_special_token_324": "451161679736",
334
+ "extra_special_token_325": "538631466654",
335
+ "extra_special_token_326": "221762325616",
336
+ "extra_special_token_327": "218391991184",
337
+ "extra_special_token_328": "322589379462",
338
+ "extra_special_token_329": "876537814263",
339
+ "extra_special_token_330": "152676556624",
340
+ "extra_special_token_331": "332522971941",
341
+ "extra_special_token_332": "884354318946",
342
+ "extra_special_token_333": "513349618943",
343
+ "extra_special_token_334": "116639746413",
344
+ "extra_special_token_335": "635185846287",
345
+ "extra_special_token_336": "993832498489",
346
+ "extra_special_token_337": "813981174797",
347
+ "extra_special_token_338": "438745114173",
348
+ "extra_special_token_339": "983493951323",
349
+ "extra_special_token_340": "724492262421",
350
+ "extra_special_token_341": "622553389126",
351
+ "extra_special_token_342": "889965243135",
352
+ "extra_special_token_343": "364492359246",
353
+ "extra_special_token_344": "154962668224",
354
+ "extra_special_token_345": "179564995814",
355
+ "extra_special_token_346": "418412875665",
356
+ "extra_special_token_347": "718951851413",
357
+ "extra_special_token_348": "699446724178",
358
+ "extra_special_token_349": "624266421831",
359
+ "extra_special_token_350": "815458725125",
360
+ "extra_special_token_351": "455423278865",
361
+ "extra_special_token_352": "393741199486",
362
+ "extra_special_token_353": "328552864359",
363
+ "extra_special_token_354": "211662639865",
364
+ "extra_special_token_355": "218784516525",
365
+ "extra_special_token_356": "762486672996",
366
+ "extra_special_token_357": "142799718159",
367
+ "extra_special_token_358": "858146415154",
368
+ "extra_special_token_359": "767858144912",
369
+ "extra_special_token_360": "571317457151",
370
+ "extra_special_token_361": "635127952696",
371
+ "extra_special_token_362": "116427191984",
372
+ "extra_special_token_363": "268921994538",
373
+ "extra_special_token_364": "523937669294",
374
+ "extra_special_token_365": "165429152138",
375
+ "extra_special_token_366": "739246183345",
376
+ "extra_special_token_367": "591464355756",
377
+ "extra_special_token_368": "212985874612",
378
+ "extra_special_token_369": "191887635211",
379
+ "extra_special_token_370": "967214577653",
380
+ "extra_special_token_371": "119342152414",
381
+ "extra_special_token_372": "946444632795",
382
+ "extra_special_token_373": "618423867817",
383
+ "extra_special_token_374": "228565148417",
384
+ "extra_special_token_375": "729116422489",
385
+ "extra_special_token_376": "527874729936",
386
+ "extra_special_token_377": "739784153482",
387
+ "extra_special_token_378": "387763951128",
388
+ "extra_special_token_379": "331369926711",
389
+ "extra_special_token_380": "562716493614",
390
+ "extra_special_token_381": "739667844957",
391
+ "extra_special_token_382": "562389434565",
392
+ "extra_special_token_383": "256497188281",
393
+ "extra_special_token_384": "859927364588",
394
+ "extra_special_token_385": "417668946583",
395
+ "extra_special_token_386": "357621613582",
396
+ "extra_special_token_387": "438435178228",
397
+ "extra_special_token_388": "485692541169",
398
+ "extra_special_token_389": "825815739116",
399
+ "extra_special_token_390": "342221452223",
400
+ "extra_special_token_391": "697747991249",
401
+ "extra_special_token_392": "716763689965",
402
+ "extra_special_token_393": "141499982867",
403
+ "extra_special_token_394": "818479319499",
404
+ "extra_special_token_395": "336813343298",
405
+ "extra_special_token_396": "594688742928",
406
+ "extra_special_token_397": "472129283475",
407
+ "extra_special_token_398": "514354144759",
408
+ "extra_special_token_399": "349249721685",
409
+ "extra_special_token_400": "546276298359",
410
+ "extra_special_token_401": "353755529131",
411
+ "extra_special_token_402": "315534574435",
412
+ "extra_special_token_403": "523723475786",
413
+ "extra_special_token_404": "215826764872",
414
+ "extra_special_token_405": "367968398551",
415
+ "extra_special_token_406": "569853653352",
416
+ "extra_special_token_407": "389715484387",
417
+ "extra_special_token_408": "293847485454",
418
+ "extra_special_token_409": "714738141818",
419
+ "extra_special_token_410": "178478368922",
420
+ "extra_special_token_411": "581493616981",
421
+ "extra_special_token_412": "589439538674",
422
+ "extra_special_token_413": "846657726193",
423
+ "extra_special_token_414": "722339992679",
424
+ "extra_special_token_415": "138154781148",
425
+ "extra_special_token_416": "757785319772",
426
+ "extra_special_token_417": "492516914298",
427
+ "extra_special_token_418": "919181521716",
428
+ "extra_special_token_419": "985781138935",
429
+ "extra_special_token_420": "476969195485",
430
+ "extra_special_token_421": "313145133463",
431
+ "extra_special_token_422": "758963111966",
432
+ "extra_special_token_423": "147541537162",
433
+ "extra_special_token_424": "557163366873",
434
+ "extra_special_token_425": "144373897488",
435
+ "extra_special_token_426": "522515164754",
436
+ "extra_special_token_427": "724964923582",
437
+ "extra_special_token_428": "284776712475",
438
+ "extra_special_token_429": "375429755114",
439
+ "extra_special_token_430": "181233596124",
440
+ "extra_special_token_431": "948585673431",
441
+ "extra_special_token_432": "243165586174",
442
+ "extra_special_token_433": "396847976144",
443
+ "extra_special_token_434": "997724962668",
444
+ "extra_special_token_435": "558837194455",
445
+ "extra_special_token_436": "163165456396",
446
+ "extra_special_token_437": "378749551722",
447
+ "extra_special_token_438": "161238482259",
448
+ "extra_special_token_439": "754978243758",
449
+ "extra_special_token_440": "195388849133",
450
+ "extra_special_token_441": "229775525672",
451
+ "extra_special_token_442": "262437452884",
452
+ "extra_special_token_443": "441377892146",
453
+ "extra_special_token_444": "451885565366",
454
+ "extra_special_token_445": "981277526855",
455
+ "extra_special_token_446": "762495822823",
456
+ "extra_special_token_447": "368763327262",
457
+ "extra_special_token_448": "757422791351",
458
+ "extra_special_token_449": "636324136426",
459
+ "extra_special_token_450": "214193645583",
460
+ "extra_special_token_451": "412843856172",
461
+ "extra_special_token_452": "179386156569",
462
+ "extra_special_token_453": "756916173536",
463
+ "extra_special_token_454": "892697125149",
464
+ "extra_special_token_455": "625334487352",
465
+ "extra_special_token_456": "941861857715",
466
+ "extra_special_token_457": "887417525236",
467
+ "extra_special_token_458": "649516938598",
468
+ "extra_special_token_459": "717628619782",
469
+ "extra_special_token_460": "438124184139",
470
+ "extra_special_token_461": "547563892268",
471
+ "extra_special_token_462": "856317483891",
472
+ "extra_special_token_463": "313313831273",
473
+ "extra_special_token_464": "371496153876",
474
+ "extra_special_token_465": "587541149322",
475
+ "extra_special_token_466": "265847332563",
476
+ "extra_special_token_467": "449549215429",
477
+ "extra_special_token_468": "163497196769",
478
+ "extra_special_token_469": "861342291298",
479
+ "extra_special_token_470": "268433315926",
480
+ "extra_special_token_471": "774679513717",
481
+ "extra_special_token_472": "851254219729",
482
+ "extra_special_token_473": "583527834464",
483
+ "extra_special_token_474": "488496781997",
484
+ "extra_special_token_475": "556814553861",
485
+ "extra_special_token_476": "482829231639",
486
+ "extra_special_token_477": "618878266619",
487
+ "extra_special_token_478": "147444452794",
488
+ "extra_special_token_479": "949235426629",
489
+ "extra_special_token_480": "357299947518",
490
+ "extra_special_token_481": "175528632226",
491
+ "extra_special_token_482": "645527857972",
492
+ "extra_special_token_483": "186872457894",
493
+ "extra_special_token_484": "552738847828",
494
+ "extra_special_token_485": "626748382482",
495
+ "extra_special_token_486": "921894985642",
496
+ "extra_special_token_487": "943878645871",
497
+ "extra_special_token_488": "859289776479",
498
+ "extra_special_token_489": "614583493135",
499
+ "extra_special_token_490": "933775286797",
500
+ "extra_special_token_491": "332234613346",
501
+ "extra_special_token_492": "325196781219",
502
+ "extra_special_token_493": "142526557681",
503
+ "extra_special_token_494": "356722692178",
504
+ "extra_special_token_495": "449318681694",
505
+ "extra_special_token_496": "687284547244",
506
+ "extra_special_token_497": "947262995132",
507
+ "extra_special_token_498": "893974619684",
508
+ "extra_special_token_499": "797238311233"
509
+ },
510
+ "is_local": false,
511
+ "model_max_length": 1000000000000000019884624838656,
512
+ "pad_token": "<|endoftext|>",
513
+ "tokenizer_class": "GPT2Tokenizer",
514
+ "unk_token": "<|endoftext|>",
515
+ "vocab_size": 49152
516
+ }
starcoder2-7b/base/train.log ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-03-21 11:04:31,986 [INFO] new_opacus_codex.train_steps: epoch=1 step=10 loss=1.5025
2
+ 2026-03-21 11:05:54,380 [INFO] new_opacus_codex.train_steps: epoch=1 step=20 loss=1.3712
3
+ 2026-03-21 11:07:15,652 [INFO] new_opacus_codex.train_steps: epoch=1 step=30 loss=1.1814
4
+ 2026-03-21 11:08:38,661 [INFO] new_opacus_codex.train_steps: epoch=1 step=40 loss=1.1834
5
+ 2026-03-21 11:10:02,035 [INFO] new_opacus_codex.train_steps: epoch=1 step=50 loss=1.1336
6
+ 2026-03-21 11:10:29,195 [INFO] new_opacus_codex.train_steps: eval event=eval_step epoch=1 step=50 eval_loss=0.8034 duration_sec=27.16
7
+ 2026-03-21 11:11:51,908 [INFO] new_opacus_codex.train_steps: epoch=1 step=60 loss=1.1219
8
+ 2026-03-21 11:13:13,533 [INFO] new_opacus_codex.train_steps: epoch=1 step=70 loss=1.0407
9
+ 2026-03-21 11:14:36,548 [INFO] new_opacus_codex.train_steps: epoch=1 step=80 loss=1.0129
10
+ 2026-03-21 11:15:57,878 [INFO] new_opacus_codex.train_steps: epoch=1 step=90 loss=1.0161
11
+ 2026-03-21 11:17:19,691 [INFO] new_opacus_codex.train_steps: epoch=1 step=100 loss=1.0200
12
+ 2026-03-21 11:17:46,876 [INFO] new_opacus_codex.train_steps: eval event=eval_step epoch=1 step=100 eval_loss=0.7755 duration_sec=27.18
13
+ 2026-03-21 11:19:10,166 [INFO] new_opacus_codex.train_steps: epoch=1 step=110 loss=1.0177
14
+ 2026-03-21 11:20:31,874 [INFO] new_opacus_codex.train_steps: epoch=1 step=120 loss=0.9631
15
+ 2026-03-21 11:21:54,417 [INFO] new_opacus_codex.train_steps: epoch=1 step=130 loss=0.9569
16
+ 2026-03-21 11:23:16,389 [INFO] new_opacus_codex.train_steps: epoch=1 step=140 loss=0.9588
17
+ 2026-03-21 11:24:38,982 [INFO] new_opacus_codex.train_steps: epoch=1 step=150 loss=0.9273
18
+ 2026-03-21 11:25:06,151 [INFO] new_opacus_codex.train_steps: eval event=eval_step epoch=1 step=150 eval_loss=0.7622 duration_sec=27.17
19
+ 2026-03-21 11:26:28,134 [INFO] new_opacus_codex.train_steps: epoch=1 step=160 loss=0.9226
20
+ 2026-03-21 11:27:49,927 [INFO] new_opacus_codex.train_steps: epoch=1 step=170 loss=0.9244
21
+ 2026-03-21 11:29:11,942 [INFO] new_opacus_codex.train_steps: epoch=1 step=180 loss=0.9250
22
+ 2026-03-21 11:30:34,185 [INFO] new_opacus_codex.train_steps: epoch=1 step=190 loss=0.8691
23
+ 2026-03-21 11:31:57,690 [INFO] new_opacus_codex.train_steps: epoch=1 step=200 loss=0.8916
24
+ 2026-03-21 11:32:24,879 [INFO] new_opacus_codex.train_steps: eval event=eval_step epoch=1 step=200 eval_loss=0.7525 duration_sec=27.19
25
+ 2026-03-21 11:34:26,079 [INFO] new_opacus_codex.train_steps: epoch=2 step=210 loss=0.9168
26
+ 2026-03-21 11:35:47,787 [INFO] new_opacus_codex.train_steps: epoch=2 step=220 loss=0.8558
27
+ 2026-03-21 11:37:09,366 [INFO] new_opacus_codex.train_steps: epoch=2 step=230 loss=0.8524
28
+ 2026-03-21 11:38:31,484 [INFO] new_opacus_codex.train_steps: epoch=2 step=240 loss=0.8580
29
+ 2026-03-21 11:39:52,882 [INFO] new_opacus_codex.train_steps: epoch=2 step=250 loss=0.8491
30
+ 2026-03-21 11:40:20,073 [INFO] new_opacus_codex.train_steps: eval event=eval_step epoch=2 step=250 eval_loss=0.7485 duration_sec=27.19
31
+ 2026-03-21 11:41:42,276 [INFO] new_opacus_codex.train_steps: epoch=2 step=260 loss=0.8442
32
+ 2026-03-21 11:43:04,391 [INFO] new_opacus_codex.train_steps: epoch=2 step=270 loss=0.8571
33
+ 2026-03-21 11:44:28,058 [INFO] new_opacus_codex.train_steps: epoch=2 step=280 loss=0.8479
34
+ 2026-03-21 11:45:50,712 [INFO] new_opacus_codex.train_steps: epoch=2 step=290 loss=0.8434
35
+ 2026-03-21 11:47:12,551 [INFO] new_opacus_codex.train_steps: epoch=2 step=300 loss=0.8405
36
+ 2026-03-21 11:47:39,718 [INFO] new_opacus_codex.train_steps: eval event=eval_step epoch=2 step=300 eval_loss=0.7462 duration_sec=27.16
37
+ 2026-03-21 11:49:02,560 [INFO] new_opacus_codex.train_steps: epoch=2 step=310 loss=0.8507
38
+ 2026-03-21 11:50:24,762 [INFO] new_opacus_codex.train_steps: epoch=2 step=320 loss=0.8479
39
+ 2026-03-21 11:51:47,919 [INFO] new_opacus_codex.train_steps: epoch=2 step=330 loss=0.8466
40
+ 2026-03-21 11:53:09,723 [INFO] new_opacus_codex.train_steps: epoch=2 step=340 loss=0.8369
41
+ 2026-03-21 11:54:31,911 [INFO] new_opacus_codex.train_steps: epoch=2 step=350 loss=0.8474
42
+ 2026-03-21 11:54:59,091 [INFO] new_opacus_codex.train_steps: eval event=eval_step epoch=2 step=350 eval_loss=0.7456 duration_sec=27.18
43
+ 2026-03-21 11:56:21,565 [INFO] new_opacus_codex.train_steps: epoch=2 step=360 loss=0.8568
44
+ 2026-03-21 11:57:43,511 [INFO] new_opacus_codex.train_steps: epoch=2 step=370 loss=0.8227
45
+ 2026-03-21 11:59:06,162 [INFO] new_opacus_codex.train_steps: epoch=2 step=380 loss=0.8342
46
+ 2026-03-21 12:00:29,231 [INFO] new_opacus_codex.train_steps: epoch=2 step=390 loss=0.8265
47
+ 2026-03-21 12:01:52,289 [INFO] new_opacus_codex.train_steps: epoch=2 step=400 loss=0.8197
48
+ 2026-03-21 12:02:19,482 [INFO] new_opacus_codex.train_steps: eval event=eval_step epoch=2 step=400 eval_loss=0.7455 duration_sec=27.19
49
+ 2026-03-21 12:03:41,445 [INFO] new_opacus_codex.train_steps: epoch=2 step=410 loss=0.8286
starcoder2-7b/dp3/adapter/README.md ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: bigcode/starcoder2-7b
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:bigcode/starcoder2-7b
7
+ - lora
8
+ - transformers
9
+ ---
10
+
11
+ # Model Card for Model ID
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+
16
+
17
+ ## Model Details
18
+
19
+ ### Model Description
20
+
21
+ <!-- Provide a longer summary of what this model is. -->
22
+
23
+
24
+
25
+ - **Developed by:** [More Information Needed]
26
+ - **Funded by [optional]:** [More Information Needed]
27
+ - **Shared by [optional]:** [More Information Needed]
28
+ - **Model type:** [More Information Needed]
29
+ - **Language(s) (NLP):** [More Information Needed]
30
+ - **License:** [More Information Needed]
31
+ - **Finetuned from model [optional]:** [More Information Needed]
32
+
33
+ ### Model Sources [optional]
34
+
35
+ <!-- Provide the basic links for the model. -->
36
+
37
+ - **Repository:** [More Information Needed]
38
+ - **Paper [optional]:** [More Information Needed]
39
+ - **Demo [optional]:** [More Information Needed]
40
+
41
+ ## Uses
42
+
43
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
+
45
+ ### Direct Use
46
+
47
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
48
+
49
+ [More Information Needed]
50
+
51
+ ### Downstream Use [optional]
52
+
53
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
54
+
55
+ [More Information Needed]
56
+
57
+ ### Out-of-Scope Use
58
+
59
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
+
61
+ [More Information Needed]
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ [More Information Needed]
68
+
69
+ ### Recommendations
70
+
71
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
72
+
73
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
74
+
75
+ ## How to Get Started with the Model
76
+
77
+ Use the code below to get started with the model.
78
+
79
+ [More Information Needed]
80
+
81
+ ## Training Details
82
+
83
+ ### Training Data
84
+
85
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
86
+
87
+ [More Information Needed]
88
+
89
+ ### Training Procedure
90
+
91
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
92
+
93
+ #### Preprocessing [optional]
94
+
95
+ [More Information Needed]
96
+
97
+
98
+ #### Training Hyperparameters
99
+
100
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
101
+
102
+ #### Speeds, Sizes, Times [optional]
103
+
104
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
105
+
106
+ [More Information Needed]
107
+
108
+ ## Evaluation
109
+
110
+ <!-- This section describes the evaluation protocols and provides the results. -->
111
+
112
+ ### Testing Data, Factors & Metrics
113
+
114
+ #### Testing Data
115
+
116
+ <!-- This should link to a Dataset Card if possible. -->
117
+
118
+ [More Information Needed]
119
+
120
+ #### Factors
121
+
122
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
123
+
124
+ [More Information Needed]
125
+
126
+ #### Metrics
127
+
128
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
129
+
130
+ [More Information Needed]
131
+
132
+ ### Results
133
+
134
+ [More Information Needed]
135
+
136
+ #### Summary
137
+
138
+
139
+
140
+ ## Model Examination [optional]
141
+
142
+ <!-- Relevant interpretability work for the model goes here -->
143
+
144
+ [More Information Needed]
145
+
146
+ ## Environmental Impact
147
+
148
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
149
+
150
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
151
+
152
+ - **Hardware Type:** [More Information Needed]
153
+ - **Hours used:** [More Information Needed]
154
+ - **Cloud Provider:** [More Information Needed]
155
+ - **Compute Region:** [More Information Needed]
156
+ - **Carbon Emitted:** [More Information Needed]
157
+
158
+ ## Technical Specifications [optional]
159
+
160
+ ### Model Architecture and Objective
161
+
162
+ [More Information Needed]
163
+
164
+ ### Compute Infrastructure
165
+
166
+ [More Information Needed]
167
+
168
+ #### Hardware
169
+
170
+ [More Information Needed]
171
+
172
+ #### Software
173
+
174
+ [More Information Needed]
175
+
176
+ ## Citation [optional]
177
+
178
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
179
+
180
+ **BibTeX:**
181
+
182
+ [More Information Needed]
183
+
184
+ **APA:**
185
+
186
+ [More Information Needed]
187
+
188
+ ## Glossary [optional]
189
+
190
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
191
+
192
+ [More Information Needed]
193
+
194
+ ## More Information [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Authors [optional]
199
+
200
+ [More Information Needed]
201
+
202
+ ## Model Card Contact
203
+
204
+ [More Information Needed]
205
+ ### Framework versions
206
+
207
+ - PEFT 0.18.1
starcoder2-7b/dp3/adapter/adapter_config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": null,
6
+ "base_model_name_or_path": "bigcode/starcoder2-7b",
7
+ "bias": "none",
8
+ "corda_config": null,
9
+ "ensure_weight_tying": true,
10
+ "eva_config": null,
11
+ "exclude_modules": null,
12
+ "fan_in_fan_out": false,
13
+ "inference_mode": true,
14
+ "init_lora_weights": true,
15
+ "layer_replication": null,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "loftq_config": {},
19
+ "lora_alpha": 32,
20
+ "lora_bias": false,
21
+ "lora_dropout": 0.05,
22
+ "megatron_config": null,
23
+ "megatron_core": "megatron.core",
24
+ "modules_to_save": [
25
+ "lm_head",
26
+ "embed_tokens"
27
+ ],
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 16,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "k_proj",
36
+ "q_proj",
37
+ "v_proj",
38
+ "o_proj"
39
+ ],
40
+ "target_parameters": null,
41
+ "task_type": "CAUSAL_LM",
42
+ "trainable_token_indices": null,
43
+ "use_dora": false,
44
+ "use_qalora": false,
45
+ "use_rslora": false
46
+ }
starcoder2-7b/dp3/adapter/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:853fac2b8cce7de6abb5b5d69b51e3a5f349fcbb8c2008b2dfeca74d3f065d9a
3
+ size 2804312360
starcoder2-7b/dp3/audit_results.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "delta": 1e-05,
3
+ "num_canaries": 500,
4
+ "num_members": 250,
5
+ "paper_guess_fraction": 0.2,
6
+ "paper_guess_steps": 20,
7
+ "loss": {
8
+ "auc": 0.525808,
9
+ "empirical_epsilon": {
10
+ "0.05": 0.05073561053723097,
11
+ "0.01": 0.0
12
+ },
13
+ "empirical_epsilon_details": {
14
+ "0.05": {
15
+ "epsilon": 0.05073561053723097,
16
+ "num_guesses": 100,
17
+ "correct_guesses": 60,
18
+ "candidate_num_guesses": [
19
+ 5,
20
+ 10,
21
+ 15,
22
+ 20,
23
+ 25,
24
+ 30,
25
+ 35,
26
+ 40,
27
+ 45,
28
+ 50,
29
+ 55,
30
+ 60,
31
+ 65,
32
+ 70,
33
+ 75,
34
+ 80,
35
+ 85,
36
+ 90,
37
+ 95,
38
+ 100
39
+ ],
40
+ "direction": "higher"
41
+ },
42
+ "0.01": {
43
+ "epsilon": 0.0,
44
+ "num_guesses": 0,
45
+ "correct_guesses": 0,
46
+ "candidate_num_guesses": [
47
+ 5,
48
+ 10,
49
+ 15,
50
+ 20,
51
+ 25,
52
+ 30,
53
+ 35,
54
+ 40,
55
+ 45,
56
+ 50,
57
+ 55,
58
+ 60,
59
+ 65,
60
+ 70,
61
+ 75,
62
+ 80,
63
+ 85,
64
+ 90,
65
+ 95,
66
+ 100
67
+ ],
68
+ "direction": "lower"
69
+ }
70
+ }
71
+ },
72
+ "embedding": {
73
+ "auc": 0.520928,
74
+ "empirical_epsilon": {
75
+ "0.05": 0.0,
76
+ "0.01": 0.0
77
+ },
78
+ "empirical_epsilon_details": {
79
+ "0.05": {
80
+ "epsilon": 0.0,
81
+ "num_guesses": 0,
82
+ "correct_guesses": 0,
83
+ "candidate_num_guesses": [
84
+ 5,
85
+ 10,
86
+ 15,
87
+ 20,
88
+ 25,
89
+ 30,
90
+ 35,
91
+ 40,
92
+ 45,
93
+ 50,
94
+ 55,
95
+ 60,
96
+ 65,
97
+ 70,
98
+ 75,
99
+ 80,
100
+ 85,
101
+ 90,
102
+ 95,
103
+ 100
104
+ ],
105
+ "direction": "lower"
106
+ },
107
+ "0.01": {
108
+ "epsilon": 0.0,
109
+ "num_guesses": 0,
110
+ "correct_guesses": 0,
111
+ "candidate_num_guesses": [
112
+ 5,
113
+ 10,
114
+ 15,
115
+ 20,
116
+ 25,
117
+ 30,
118
+ 35,
119
+ 40,
120
+ 45,
121
+ 50,
122
+ 55,
123
+ 60,
124
+ 65,
125
+ 70,
126
+ 75,
127
+ 80,
128
+ 85,
129
+ 90,
130
+ 95,
131
+ 100
132
+ ],
133
+ "direction": "lower"
134
+ }
135
+ }
136
+ }
137
+ }
starcoder2-7b/dp3/audit_scores.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fe821d6aec2e4b9779349695c9909f92bfbca3b25bd240d418bb85f362550a5
3
+ size 12784
starcoder2-7b/dp3/canary_meta.json ADDED
The diff for this file is too large to render. See raw diff
 
starcoder2-7b/dp3/codecarbon.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ timestamp,project_name,run_id,experiment_id,duration,emissions,emissions_rate,cpu_power,gpu_power,ram_power,cpu_energy,gpu_energy,ram_energy,energy_consumed,water_consumed,country_name,country_iso_code,region,cloud_provider,cloud_region,os,python_version,codecarbon_version,cpu_count,cpu_model,gpu_count,gpu_model,longitude,latitude,ram_total_size,tracking_mode,cpu_utilization_percent,gpu_utilization_percent,ram_utilization_percent,ram_used_gb,on_cloud,pue,wue
2
+ 2026-03-21T13:36:12,codedp-starcoder2-7b-cpt-dp3,457cc088-8edf-418e-8c6d-170a67e73d63,5b0fa12a-3dd7-45bb-9766-cc326314d9f1,3938.0607158355415,0.5702959606126596,0.00014481644691749236,179.92904019353898,2298.550082608512,70.0,0.18975838183590477,2.514705143984827,0.07382037755078118,2.778283903371513,0.0,United States,USA,california,,,Linux-5.15.0-157-generic-x86_64-with-glibc2.39,3.12.13,3.2.5,224,Intel(R) Xeon(R) Platinum 8480C,4,4 x NVIDIA H200,-121.9552,37.3541,2015.5625190734863,machine,2.4580389144905275,95.87682458386683,2.36410650281618,47.93522234837092,N,1.0,0.0
starcoder2-7b/dp3/epochs/epoch_001/adapter/README.md ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: bigcode/starcoder2-7b
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:bigcode/starcoder2-7b
7
+ - lora
8
+ - transformers
9
+ ---
10
+
11
+ # Model Card for Model ID
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+
16
+
17
+ ## Model Details
18
+
19
+ ### Model Description
20
+
21
+ <!-- Provide a longer summary of what this model is. -->
22
+
23
+
24
+
25
+ - **Developed by:** [More Information Needed]
26
+ - **Funded by [optional]:** [More Information Needed]
27
+ - **Shared by [optional]:** [More Information Needed]
28
+ - **Model type:** [More Information Needed]
29
+ - **Language(s) (NLP):** [More Information Needed]
30
+ - **License:** [More Information Needed]
31
+ - **Finetuned from model [optional]:** [More Information Needed]
32
+
33
+ ### Model Sources [optional]
34
+
35
+ <!-- Provide the basic links for the model. -->
36
+
37
+ - **Repository:** [More Information Needed]
38
+ - **Paper [optional]:** [More Information Needed]
39
+ - **Demo [optional]:** [More Information Needed]
40
+
41
+ ## Uses
42
+
43
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
+
45
+ ### Direct Use
46
+
47
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
48
+
49
+ [More Information Needed]
50
+
51
+ ### Downstream Use [optional]
52
+
53
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
54
+
55
+ [More Information Needed]
56
+
57
+ ### Out-of-Scope Use
58
+
59
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
+
61
+ [More Information Needed]
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ [More Information Needed]
68
+
69
+ ### Recommendations
70
+
71
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
72
+
73
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
74
+
75
+ ## How to Get Started with the Model
76
+
77
+ Use the code below to get started with the model.
78
+
79
+ [More Information Needed]
80
+
81
+ ## Training Details
82
+
83
+ ### Training Data
84
+
85
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
86
+
87
+ [More Information Needed]
88
+
89
+ ### Training Procedure
90
+
91
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
92
+
93
+ #### Preprocessing [optional]
94
+
95
+ [More Information Needed]
96
+
97
+
98
+ #### Training Hyperparameters
99
+
100
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
101
+
102
+ #### Speeds, Sizes, Times [optional]
103
+
104
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
105
+
106
+ [More Information Needed]
107
+
108
+ ## Evaluation
109
+
110
+ <!-- This section describes the evaluation protocols and provides the results. -->
111
+
112
+ ### Testing Data, Factors & Metrics
113
+
114
+ #### Testing Data
115
+
116
+ <!-- This should link to a Dataset Card if possible. -->
117
+
118
+ [More Information Needed]
119
+
120
+ #### Factors
121
+
122
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
123
+
124
+ [More Information Needed]
125
+
126
+ #### Metrics
127
+
128
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
129
+
130
+ [More Information Needed]
131
+
132
+ ### Results
133
+
134
+ [More Information Needed]
135
+
136
+ #### Summary
137
+
138
+
139
+
140
+ ## Model Examination [optional]
141
+
142
+ <!-- Relevant interpretability work for the model goes here -->
143
+
144
+ [More Information Needed]
145
+
146
+ ## Environmental Impact
147
+
148
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
149
+
150
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
151
+
152
+ - **Hardware Type:** [More Information Needed]
153
+ - **Hours used:** [More Information Needed]
154
+ - **Cloud Provider:** [More Information Needed]
155
+ - **Compute Region:** [More Information Needed]
156
+ - **Carbon Emitted:** [More Information Needed]
157
+
158
+ ## Technical Specifications [optional]
159
+
160
+ ### Model Architecture and Objective
161
+
162
+ [More Information Needed]
163
+
164
+ ### Compute Infrastructure
165
+
166
+ [More Information Needed]
167
+
168
+ #### Hardware
169
+
170
+ [More Information Needed]
171
+
172
+ #### Software
173
+
174
+ [More Information Needed]
175
+
176
+ ## Citation [optional]
177
+
178
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
179
+
180
+ **BibTeX:**
181
+
182
+ [More Information Needed]
183
+
184
+ **APA:**
185
+
186
+ [More Information Needed]
187
+
188
+ ## Glossary [optional]
189
+
190
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
191
+
192
+ [More Information Needed]
193
+
194
+ ## More Information [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Authors [optional]
199
+
200
+ [More Information Needed]
201
+
202
+ ## Model Card Contact
203
+
204
+ [More Information Needed]
205
+ ### Framework versions
206
+
207
+ - PEFT 0.18.1
starcoder2-7b/dp3/epochs/epoch_001/adapter/adapter_config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": null,
6
+ "base_model_name_or_path": "bigcode/starcoder2-7b",
7
+ "bias": "none",
8
+ "corda_config": null,
9
+ "ensure_weight_tying": true,
10
+ "eva_config": null,
11
+ "exclude_modules": null,
12
+ "fan_in_fan_out": false,
13
+ "inference_mode": true,
14
+ "init_lora_weights": true,
15
+ "layer_replication": null,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "loftq_config": {},
19
+ "lora_alpha": 32,
20
+ "lora_bias": false,
21
+ "lora_dropout": 0.05,
22
+ "megatron_config": null,
23
+ "megatron_core": "megatron.core",
24
+ "modules_to_save": [
25
+ "lm_head",
26
+ "embed_tokens"
27
+ ],
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 16,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "k_proj",
36
+ "q_proj",
37
+ "v_proj",
38
+ "o_proj"
39
+ ],
40
+ "target_parameters": null,
41
+ "task_type": "CAUSAL_LM",
42
+ "trainable_token_indices": null,
43
+ "use_dora": false,
44
+ "use_qalora": false,
45
+ "use_rslora": false
46
+ }
starcoder2-7b/dp3/epochs/epoch_001/adapter/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1eb0abf981ab5eb12b4b55a3c95748d96446558dfe488188b7929dae1d80c5b4
3
+ size 2804312360
starcoder2-7b/dp3/epochs/epoch_001/audit_results.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "delta": 1e-05,
3
+ "num_canaries": 500,
4
+ "num_members": 250,
5
+ "paper_guess_fraction": 0.2,
6
+ "paper_guess_steps": 20,
7
+ "loss": {
8
+ "auc": 0.527168,
9
+ "empirical_epsilon": {
10
+ "0.05": 0.09125839080661535,
11
+ "0.01": 0.0
12
+ },
13
+ "empirical_epsilon_details": {
14
+ "0.05": {
15
+ "epsilon": 0.09125839080661535,
16
+ "num_guesses": 100,
17
+ "correct_guesses": 61,
18
+ "candidate_num_guesses": [
19
+ 5,
20
+ 10,
21
+ 15,
22
+ 20,
23
+ 25,
24
+ 30,
25
+ 35,
26
+ 40,
27
+ 45,
28
+ 50,
29
+ 55,
30
+ 60,
31
+ 65,
32
+ 70,
33
+ 75,
34
+ 80,
35
+ 85,
36
+ 90,
37
+ 95,
38
+ 100
39
+ ],
40
+ "direction": "higher"
41
+ },
42
+ "0.01": {
43
+ "epsilon": 0.0,
44
+ "num_guesses": 0,
45
+ "correct_guesses": 0,
46
+ "candidate_num_guesses": [
47
+ 5,
48
+ 10,
49
+ 15,
50
+ 20,
51
+ 25,
52
+ 30,
53
+ 35,
54
+ 40,
55
+ 45,
56
+ 50,
57
+ 55,
58
+ 60,
59
+ 65,
60
+ 70,
61
+ 75,
62
+ 80,
63
+ 85,
64
+ 90,
65
+ 95,
66
+ 100
67
+ ],
68
+ "direction": "lower"
69
+ }
70
+ }
71
+ },
72
+ "embedding": {
73
+ "auc": 0.514792,
74
+ "empirical_epsilon": {
75
+ "0.05": 0.0,
76
+ "0.01": 0.0
77
+ },
78
+ "empirical_epsilon_details": {
79
+ "0.05": {
80
+ "epsilon": 0.0,
81
+ "num_guesses": 0,
82
+ "correct_guesses": 0,
83
+ "candidate_num_guesses": [
84
+ 5,
85
+ 10,
86
+ 15,
87
+ 20,
88
+ 25,
89
+ 30,
90
+ 35,
91
+ 40,
92
+ 45,
93
+ 50,
94
+ 55,
95
+ 60,
96
+ 65,
97
+ 70,
98
+ 75,
99
+ 80,
100
+ 85,
101
+ 90,
102
+ 95,
103
+ 100
104
+ ],
105
+ "direction": "lower"
106
+ },
107
+ "0.01": {
108
+ "epsilon": 0.0,
109
+ "num_guesses": 0,
110
+ "correct_guesses": 0,
111
+ "candidate_num_guesses": [
112
+ 5,
113
+ 10,
114
+ 15,
115
+ 20,
116
+ 25,
117
+ 30,
118
+ 35,
119
+ 40,
120
+ 45,
121
+ 50,
122
+ 55,
123
+ 60,
124
+ 65,
125
+ 70,
126
+ 75,
127
+ 80,
128
+ 85,
129
+ 90,
130
+ 95,
131
+ 100
132
+ ],
133
+ "direction": "lower"
134
+ }
135
+ }
136
+ }
137
+ }
starcoder2-7b/dp3/epochs/epoch_001/audit_scores.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3dedd61e7d2e9b43bbdd58fe90dba89fa6bab75a49f7c8b244e1a9d3a2cb374
3
+ size 12784
starcoder2-7b/dp3/epochs/epoch_002/adapter/README.md ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: bigcode/starcoder2-7b
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:bigcode/starcoder2-7b
7
+ - lora
8
+ - transformers
9
+ ---
10
+
11
+ # Model Card for Model ID
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+
16
+
17
+ ## Model Details
18
+
19
+ ### Model Description
20
+
21
+ <!-- Provide a longer summary of what this model is. -->
22
+
23
+
24
+
25
+ - **Developed by:** [More Information Needed]
26
+ - **Funded by [optional]:** [More Information Needed]
27
+ - **Shared by [optional]:** [More Information Needed]
28
+ - **Model type:** [More Information Needed]
29
+ - **Language(s) (NLP):** [More Information Needed]
30
+ - **License:** [More Information Needed]
31
+ - **Finetuned from model [optional]:** [More Information Needed]
32
+
33
+ ### Model Sources [optional]
34
+
35
+ <!-- Provide the basic links for the model. -->
36
+
37
+ - **Repository:** [More Information Needed]
38
+ - **Paper [optional]:** [More Information Needed]
39
+ - **Demo [optional]:** [More Information Needed]
40
+
41
+ ## Uses
42
+
43
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
+
45
+ ### Direct Use
46
+
47
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
48
+
49
+ [More Information Needed]
50
+
51
+ ### Downstream Use [optional]
52
+
53
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
54
+
55
+ [More Information Needed]
56
+
57
+ ### Out-of-Scope Use
58
+
59
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
+
61
+ [More Information Needed]
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ [More Information Needed]
68
+
69
+ ### Recommendations
70
+
71
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
72
+
73
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
74
+
75
+ ## How to Get Started with the Model
76
+
77
+ Use the code below to get started with the model.
78
+
79
+ [More Information Needed]
80
+
81
+ ## Training Details
82
+
83
+ ### Training Data
84
+
85
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
86
+
87
+ [More Information Needed]
88
+
89
+ ### Training Procedure
90
+
91
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
92
+
93
+ #### Preprocessing [optional]
94
+
95
+ [More Information Needed]
96
+
97
+
98
+ #### Training Hyperparameters
99
+
100
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
101
+
102
+ #### Speeds, Sizes, Times [optional]
103
+
104
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
105
+
106
+ [More Information Needed]
107
+
108
+ ## Evaluation
109
+
110
+ <!-- This section describes the evaluation protocols and provides the results. -->
111
+
112
+ ### Testing Data, Factors & Metrics
113
+
114
+ #### Testing Data
115
+
116
+ <!-- This should link to a Dataset Card if possible. -->
117
+
118
+ [More Information Needed]
119
+
120
+ #### Factors
121
+
122
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
123
+
124
+ [More Information Needed]
125
+
126
+ #### Metrics
127
+
128
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
129
+
130
+ [More Information Needed]
131
+
132
+ ### Results
133
+
134
+ [More Information Needed]
135
+
136
+ #### Summary
137
+
138
+
139
+
140
+ ## Model Examination [optional]
141
+
142
+ <!-- Relevant interpretability work for the model goes here -->
143
+
144
+ [More Information Needed]
145
+
146
+ ## Environmental Impact
147
+
148
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
149
+
150
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
151
+
152
+ - **Hardware Type:** [More Information Needed]
153
+ - **Hours used:** [More Information Needed]
154
+ - **Cloud Provider:** [More Information Needed]
155
+ - **Compute Region:** [More Information Needed]
156
+ - **Carbon Emitted:** [More Information Needed]
157
+
158
+ ## Technical Specifications [optional]
159
+
160
+ ### Model Architecture and Objective
161
+
162
+ [More Information Needed]
163
+
164
+ ### Compute Infrastructure
165
+
166
+ [More Information Needed]
167
+
168
+ #### Hardware
169
+
170
+ [More Information Needed]
171
+
172
+ #### Software
173
+
174
+ [More Information Needed]
175
+
176
+ ## Citation [optional]
177
+
178
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
179
+
180
+ **BibTeX:**
181
+
182
+ [More Information Needed]
183
+
184
+ **APA:**
185
+
186
+ [More Information Needed]
187
+
188
+ ## Glossary [optional]
189
+
190
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
191
+
192
+ [More Information Needed]
193
+
194
+ ## More Information [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Authors [optional]
199
+
200
+ [More Information Needed]
201
+
202
+ ## Model Card Contact
203
+
204
+ [More Information Needed]
205
+ ### Framework versions
206
+
207
+ - PEFT 0.18.1
starcoder2-7b/dp3/epochs/epoch_002/adapter/adapter_config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": null,
6
+ "base_model_name_or_path": "bigcode/starcoder2-7b",
7
+ "bias": "none",
8
+ "corda_config": null,
9
+ "ensure_weight_tying": true,
10
+ "eva_config": null,
11
+ "exclude_modules": null,
12
+ "fan_in_fan_out": false,
13
+ "inference_mode": true,
14
+ "init_lora_weights": true,
15
+ "layer_replication": null,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "loftq_config": {},
19
+ "lora_alpha": 32,
20
+ "lora_bias": false,
21
+ "lora_dropout": 0.05,
22
+ "megatron_config": null,
23
+ "megatron_core": "megatron.core",
24
+ "modules_to_save": [
25
+ "lm_head",
26
+ "embed_tokens"
27
+ ],
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 16,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "k_proj",
36
+ "q_proj",
37
+ "v_proj",
38
+ "o_proj"
39
+ ],
40
+ "target_parameters": null,
41
+ "task_type": "CAUSAL_LM",
42
+ "trainable_token_indices": null,
43
+ "use_dora": false,
44
+ "use_qalora": false,
45
+ "use_rslora": false
46
+ }
starcoder2-7b/dp3/epochs/epoch_002/adapter/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:853fac2b8cce7de6abb5b5d69b51e3a5f349fcbb8c2008b2dfeca74d3f065d9a
3
+ size 2804312360
starcoder2-7b/dp3/epochs/epoch_002/audit_results.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "delta": 1e-05,
3
+ "num_canaries": 500,
4
+ "num_members": 250,
5
+ "paper_guess_fraction": 0.2,
6
+ "paper_guess_steps": 20,
7
+ "loss": {
8
+ "auc": 0.525808,
9
+ "empirical_epsilon": {
10
+ "0.05": 0.05073561053723097,
11
+ "0.01": 0.0
12
+ },
13
+ "empirical_epsilon_details": {
14
+ "0.05": {
15
+ "epsilon": 0.05073561053723097,
16
+ "num_guesses": 100,
17
+ "correct_guesses": 60,
18
+ "candidate_num_guesses": [
19
+ 5,
20
+ 10,
21
+ 15,
22
+ 20,
23
+ 25,
24
+ 30,
25
+ 35,
26
+ 40,
27
+ 45,
28
+ 50,
29
+ 55,
30
+ 60,
31
+ 65,
32
+ 70,
33
+ 75,
34
+ 80,
35
+ 85,
36
+ 90,
37
+ 95,
38
+ 100
39
+ ],
40
+ "direction": "higher"
41
+ },
42
+ "0.01": {
43
+ "epsilon": 0.0,
44
+ "num_guesses": 0,
45
+ "correct_guesses": 0,
46
+ "candidate_num_guesses": [
47
+ 5,
48
+ 10,
49
+ 15,
50
+ 20,
51
+ 25,
52
+ 30,
53
+ 35,
54
+ 40,
55
+ 45,
56
+ 50,
57
+ 55,
58
+ 60,
59
+ 65,
60
+ 70,
61
+ 75,
62
+ 80,
63
+ 85,
64
+ 90,
65
+ 95,
66
+ 100
67
+ ],
68
+ "direction": "lower"
69
+ }
70
+ }
71
+ },
72
+ "embedding": {
73
+ "auc": 0.520928,
74
+ "empirical_epsilon": {
75
+ "0.05": 0.0,
76
+ "0.01": 0.0
77
+ },
78
+ "empirical_epsilon_details": {
79
+ "0.05": {
80
+ "epsilon": 0.0,
81
+ "num_guesses": 0,
82
+ "correct_guesses": 0,
83
+ "candidate_num_guesses": [
84
+ 5,
85
+ 10,
86
+ 15,
87
+ 20,
88
+ 25,
89
+ 30,
90
+ 35,
91
+ 40,
92
+ 45,
93
+ 50,
94
+ 55,
95
+ 60,
96
+ 65,
97
+ 70,
98
+ 75,
99
+ 80,
100
+ 85,
101
+ 90,
102
+ 95,
103
+ 100
104
+ ],
105
+ "direction": "lower"
106
+ },
107
+ "0.01": {
108
+ "epsilon": 0.0,
109
+ "num_guesses": 0,
110
+ "correct_guesses": 0,
111
+ "candidate_num_guesses": [
112
+ 5,
113
+ 10,
114
+ 15,
115
+ 20,
116
+ 25,
117
+ 30,
118
+ 35,
119
+ 40,
120
+ 45,
121
+ 50,
122
+ 55,
123
+ 60,
124
+ 65,
125
+ 70,
126
+ 75,
127
+ 80,
128
+ 85,
129
+ 90,
130
+ 95,
131
+ 100
132
+ ],
133
+ "direction": "lower"
134
+ }
135
+ }
136
+ }
137
+ }
starcoder2-7b/dp3/epochs/epoch_002/audit_scores.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fe821d6aec2e4b9779349695c9909f92bfbca3b25bd240d418bb85f362550a5
3
+ size 12784
starcoder2-7b/dp3/metrics.jsonl ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"timestamp": 1774096500.8353336, "event": "train_step", "step": 10, "epoch": 1, "metrics": {"train/step_loss": 1.6323321043555417, "train/step_real_loss": 1.330917976796627, "train/lr": 0.00018181818181818183, "train/step_canary_loss": 8.062500158945719, "perf/step_duration_sec": 17.560116425156593, "perf/samples_per_sec": 7.573981674145647, "perf/tokens_per_sec": 6185.665138551686, "perf/logical_batch_size": 133.0, "perf/logical_token_count": 108621.0, "perf/physical_batches": 18.0, "privacy/epsilon": 0.7131647471248268, "system/cuda_memory_allocated_gb": 16.63217782974243, "system/cuda_max_memory_allocated_gb": 74.69534254074097}}
2
+ {"timestamp": 1774096676.393492, "event": "train_step", "step": 20, "epoch": 1, "metrics": {"train/step_loss": 1.826857232699429, "train/step_real_loss": 1.3395854756236076, "train/lr": 0.00019897180218885507, "train/step_canary_loss": 8.756944444444445, "perf/step_duration_sec": 18.843991773203015, "perf/samples_per_sec": 7.217154498729827, "perf/tokens_per_sec": 5434.1989336686165, "perf/logical_batch_size": 136.0, "perf/logical_token_count": 102402.0, "perf/physical_batches": 18.0, "privacy/epsilon": 0.9542480237478311, "system/cuda_memory_allocated_gb": 16.916476249694824, "system/cuda_max_memory_allocated_gb": 74.69534397125244}}
3
+ {"timestamp": 1774096852.7017646, "event": "train_step", "step": 30, "epoch": 1, "metrics": {"train/step_loss": 1.5258880311792546, "train/step_real_loss": 1.3621462360024452, "train/lr": 0.00019544467510209388, "train/step_canary_loss": 6.765625476837158, "perf/step_duration_sec": 18.006671600043774, "perf/samples_per_sec": 7.275081309289916, "perf/tokens_per_sec": 5874.1561099910805, "perf/logical_batch_size": 131.0, "perf/logical_token_count": 105774.0, "perf/physical_batches": 18.0, "privacy/epsilon": 1.145336977893831, "system/cuda_memory_allocated_gb": 16.442769050598145, "system/cuda_max_memory_allocated_gb": 74.69534397125244}}
4
+ {"timestamp": 1774097029.0227332, "event": "train_step", "step": 40, "epoch": 1, "metrics": {"train/step_loss": 1.6870967794347693, "train/step_real_loss": 1.2007466033101082, "train/lr": 0.00018949541262593762, "train/step_canary_loss": 10.580357142857142, "perf/step_duration_sec": 17.532358843833208, "perf/samples_per_sec": 7.643010344106256, "perf/tokens_per_sec": 6436.156195815627, "perf/logical_batch_size": 134.0, "perf/logical_token_count": 112841.0, "perf/physical_batches": 18.0, "privacy/epsilon": 1.311193108872294, "system/cuda_memory_allocated_gb": 16.726882934570312, "system/cuda_max_memory_allocated_gb": 74.69534683227539}}
5
+ {"timestamp": 1774097204.3534977, "event": "train_step", "step": 50, "epoch": 1, "metrics": {"train/step_loss": 1.5818433902881763, "train/step_real_loss": 1.2425691708922386, "train/lr": 0.00018127499143005268, "train/step_canary_loss": 7.785714830671038, "perf/step_duration_sec": 17.701514894142747, "perf/samples_per_sec": 7.569973575783576, "perf/tokens_per_sec": 6216.078152520672, "perf/logical_batch_size": 134.0, "perf/logical_token_count": 110034.0, "perf/physical_batches": 18.0, "privacy/epsilon": 1.4583641061524852, "system/cuda_memory_allocated_gb": 16.726882934570312, "system/cuda_max_memory_allocated_gb": 74.69534683227539}}
6
+ {"timestamp": 1774097231.3920577, "event": "eval_step", "step": 50, "epoch": 1, "metrics": {"eval/loss": 0.8423196817266531, "eval/duration_sec": 27.033981669694185}}
7
+ {"timestamp": 1774097408.5970793, "event": "train_step", "step": 60, "epoch": 1, "metrics": {"train/step_loss": 1.6164713336112804, "train/step_real_loss": 1.3143803551793098, "train/lr": 0.0001709920242324663, "train/step_canary_loss": 9.350000381469727, "perf/step_duration_sec": 17.616566620767117, "perf/samples_per_sec": 7.549711749349289, "perf/tokens_per_sec": 5881.736335493052, "perf/logical_batch_size": 133.0, "perf/logical_token_count": 103616.0, "perf/physical_batches": 17.0, "privacy/epsilon": 1.5929146459722179, "system/cuda_memory_allocated_gb": 16.63217782974243, "system/cuda_max_memory_allocated_gb": 74.69534969329834}}
8
+ {"timestamp": 1774097584.0989442, "event": "train_step", "step": 70, "epoch": 1, "metrics": {"train/step_loss": 1.5537053346633911, "train/step_real_loss": 1.274133626371622, "train/lr": 0.00015890746575622231, "train/step_canary_loss": 10.5, "perf/step_duration_sec": 17.34652972780168, "perf/samples_per_sec": 7.609591201889827, "perf/tokens_per_sec": 6165.613622913036, "perf/logical_batch_size": 132.0, "perf/logical_token_count": 106952.0, "perf/physical_batches": 17.0, "privacy/epsilon": 1.7182200620091768, "system/cuda_memory_allocated_gb": 16.537474155426025, "system/cuda_max_memory_allocated_gb": 74.69534969329834}}
9
+ {"timestamp": 1774097758.6071105, "event": "train_step", "step": 80, "epoch": 1, "metrics": {"train/step_loss": 1.6813924115403254, "train/step_real_loss": 1.3261733427643776, "train/lr": 0.00014532799038330385, "train/step_canary_loss": 10.77500057220459, "perf/step_duration_sec": 16.975525552406907, "perf/samples_per_sec": 7.834808977748694, "perf/tokens_per_sec": 6001.228043602786, "perf/logical_batch_size": 133.0, "perf/logical_token_count": 101874.0, "perf/physical_batches": 17.0, "privacy/epsilon": 1.8360751937320303, "system/cuda_memory_allocated_gb": 16.63217782974243, "system/cuda_max_memory_allocated_gb": 74.69534969329834}}
10
+ {"timestamp": 1774097934.2494926, "event": "train_step", "step": 90, "epoch": 1, "metrics": {"train/step_loss": 1.8558794460157408, "train/step_real_loss": 1.2407655008137226, "train/lr": 0.00013059820956358998, "train/step_canary_loss": 10.604166666666666, "perf/step_duration_sec": 17.994612561538815, "perf/samples_per_sec": 7.557817626520207, "perf/tokens_per_sec": 6203.967971981345, "perf/logical_batch_size": 136.0, "perf/logical_token_count": 111638.0, "perf/physical_batches": 18.0, "privacy/epsilon": 1.9476934830264792, "system/cuda_memory_allocated_gb": 16.916476249694824, "system/cuda_max_memory_allocated_gb": 74.69534969329834}}
11
+ {"timestamp": 1774098110.4647262, "event": "train_step", "step": 100, "epoch": 1, "metrics": {"train/step_loss": 1.86213939777319, "train/step_real_loss": 1.2256366163492203, "train/lr": 0.00011509192648058249, "train/step_canary_loss": 10.009375, "perf/step_duration_sec": 17.502219082787633, "perf/samples_per_sec": 7.884714466619527, "perf/tokens_per_sec": 6282.689039593839, "perf/logical_batch_size": 138.0, "perf/logical_token_count": 109961.0, "perf/physical_batches": 18.0, "privacy/epsilon": 2.054057754637441, "system/cuda_memory_allocated_gb": 16.34807252883911, "system/cuda_max_memory_allocated_gb": 74.69534969329834}}
12
+ {"timestamp": 1774098137.4743512, "event": "eval_step", "step": 100, "epoch": 1, "metrics": {"eval/loss": 0.8425039929905157, "eval/duration_sec": 27.006801065057516}}
13
+ {"timestamp": 1774098228.5487797, "event": "train_epoch", "step": 104, "epoch": 1, "metrics": {"train/epoch_loss": 1.7094427731604633, "train/epoch_real_loss": 1.285309421313279, "train/epoch_canary_loss": 9.719154772066274, "perf/epoch_duration_sec": 1878.0538540109992, "perf/epoch_samples_per_sec": 29.64824458099586, "perf/epoch_tokens_per_sec": 23429.532069074678, "perf/epoch_samples": 55681.0, "perf/epoch_tokens": 44001923.0, "system/cuda_epoch_peak_memory_gb": 74.69534969329834, "eval/loss": 0.8425441932404178, "eval/duration_sec": 26.999734753742814, "privacy/epsilon": 2.0952814257505974}}
14
+ {"timestamp": 1774098238.4572744, "event": "audit_epoch", "step": 104, "epoch": 1, "metrics": {"audit/delta": 1e-05, "audit/num_canaries": 500.0, "audit/num_members": 250.0, "audit/paper_guess_fraction": 0.2, "audit/paper_guess_steps": 20.0, "audit/loss/auc": 0.527168, "audit/loss/empirical_epsilon/0.05": 0.09125839080661535, "audit/loss/empirical_epsilon/0.01": 0.0, "audit/loss/empirical_epsilon_details/0.05/epsilon": 0.09125839080661535, "audit/loss/empirical_epsilon_details/0.05/num_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.05/correct_guesses": 61.0, "audit/loss/empirical_epsilon_details/0.01/epsilon": 0.0, "audit/loss/empirical_epsilon_details/0.01/num_guesses": 0.0, "audit/loss/empirical_epsilon_details/0.01/correct_guesses": 0.0, "audit/embedding/auc": 0.514792, "audit/embedding/empirical_epsilon/0.05": 0.0, "audit/embedding/empirical_epsilon/0.01": 0.0, "audit/embedding/empirical_epsilon_details/0.05/epsilon": 0.0, "audit/embedding/empirical_epsilon_details/0.05/num_guesses": 0.0, "audit/embedding/empirical_epsilon_details/0.05/correct_guesses": 0.0, "audit/embedding/empirical_epsilon_details/0.01/epsilon": 0.0, "audit/embedding/empirical_epsilon_details/0.01/num_guesses": 0.0, "audit/embedding/empirical_epsilon_details/0.01/correct_guesses": 0.0, "perf/audit_duration_sec": 5.7471471559256315}}
15
+ {"timestamp": 1774098344.630848, "event": "train_step", "step": 110, "epoch": 2, "metrics": {"train/step_loss": 1.570878519940732, "train/step_real_loss": 1.2690251618623734, "train/lr": 9.920264990753837e-05, "train/step_canary_loss": 8.010416825612387, "perf/step_duration_sec": 17.30536743812263, "perf/samples_per_sec": 7.685476802243991, "perf/tokens_per_sec": 6186.173184867881, "perf/logical_batch_size": 133.0, "perf/logical_token_count": 107054.0, "perf/physical_batches": 18.0, "privacy/epsilon": 2.155895236118489, "system/cuda_memory_allocated_gb": 16.63217782974243, "system/cuda_max_memory_allocated_gb": 74.69534254074097}}
16
+ {"timestamp": 1774098520.7483177, "event": "train_step", "step": 120, "epoch": 2, "metrics": {"train/step_loss": 1.733701615437974, "train/step_real_loss": 1.2730829790234566, "train/lr": 8.333360798744496e-05, "train/step_canary_loss": 8.284722222222221, "perf/step_duration_sec": 18.148367216810584, "perf/samples_per_sec": 7.493787092539381, "perf/tokens_per_sec": 5505.233545608104, "perf/logical_batch_size": 136.0, "perf/logical_token_count": 99911.0, "perf/physical_batches": 18.0, "privacy/epsilon": 2.2538737991591966, "system/cuda_memory_allocated_gb": 16.916476249694824, "system/cuda_max_memory_allocated_gb": 74.69534254074097}}
17
+ {"timestamp": 1774098695.815589, "event": "train_step", "step": 130, "epoch": 2, "metrics": {"train/step_loss": 1.4695805564071194, "train/step_real_loss": 1.2108174487948418, "train/lr": 6.788751536089739e-05, "train/step_canary_loss": 9.75, "perf/step_duration_sec": 16.7934517133981, "perf/samples_per_sec": 7.860206600331495, "perf/tokens_per_sec": 6584.88807942771, "perf/logical_batch_size": 132.0, "perf/logical_token_count": 110583.0, "perf/physical_batches": 17.0, "privacy/epsilon": 2.34829265801524, "system/cuda_memory_allocated_gb": 16.537474155426025, "system/cuda_max_memory_allocated_gb": 74.69534254074097}}
18
+ {"timestamp": 1774098871.6479974, "event": "train_step", "step": 140, "epoch": 2, "metrics": {"train/step_loss": 1.409767813354958, "train/step_real_loss": 1.280211590230465, "train/lr": 5.325635332531864e-05, "train/step_canary_loss": 6.9375, "perf/step_duration_sec": 17.296851823106408, "perf/samples_per_sec": 7.515818562215838, "perf/tokens_per_sec": 5869.218343212228, "perf/logical_batch_size": 130.0, "perf/logical_token_count": 101519.0, "perf/physical_batches": 18.0, "privacy/epsilon": 2.4397183333948855, "system/cuda_memory_allocated_gb": 16.348063945770264, "system/cuda_max_memory_allocated_gb": 74.69534254074097}}
19
+ {"timestamp": 1774099047.329602, "event": "train_step", "step": 150, "epoch": 2, "metrics": {"train/step_loss": 1.470641034250041, "train/step_real_loss": 1.2892888560891151, "train/lr": 3.981142237826332e-05, "train/step_canary_loss": 9.208333969116211, "perf/step_duration_sec": 17.535365080460906, "perf/samples_per_sec": 7.4706172012334715, "perf/tokens_per_sec": 5992.347437184814, "perf/logical_batch_size": 131.0, "perf/logical_token_count": 105078.0, "perf/physical_batches": 17.0, "privacy/epsilon": 2.5283910517887054, "system/cuda_memory_allocated_gb": 16.442769050598145, "system/cuda_max_memory_allocated_gb": 74.69534254074097}}
20
+ {"timestamp": 1774099074.3730319, "event": "eval_step", "step": 150, "epoch": 2, "metrics": {"eval/loss": 0.842570746081999, "eval/duration_sec": 27.040043698623776}}
21
+ {"timestamp": 1774099249.8653634, "event": "train_step", "step": 160, "epoch": 2, "metrics": {"train/step_loss": 1.536541504646415, "train/step_real_loss": 1.2477270364761353, "train/lr": 2.789391958515183e-05, "train/step_canary_loss": 7.697916825612386, "perf/step_duration_sec": 17.38564824871719, "perf/samples_per_sec": 7.649987972683935, "perf/tokens_per_sec": 6056.662282222899, "perf/logical_batch_size": 133.0, "perf/logical_token_count": 105299.0, "perf/physical_batches": 18.0, "privacy/epsilon": 2.6145698431381854, "system/cuda_memory_allocated_gb": 16.63217782974243, "system/cuda_max_memory_allocated_gb": 74.69534540176392}}
22
+ {"timestamp": 1774099424.502899, "event": "train_step", "step": 170, "epoch": 2, "metrics": {"train/step_loss": 1.4997789441174223, "train/step_real_loss": 1.2844417318701744, "train/lr": 1.7806279893114875e-05, "train/step_canary_loss": 10.6875, "perf/step_duration_sec": 17.294984631240368, "perf/samples_per_sec": 7.574450211616343, "perf/tokens_per_sec": 5730.794337970551, "perf/logical_batch_size": 131.0, "perf/logical_token_count": 99114.0, "perf/physical_batches": 17.0, "privacy/epsilon": 2.6985357679843074, "system/cuda_memory_allocated_gb": 16.442769050598145, "system/cuda_max_memory_allocated_gb": 74.69534540176392}}
23
+ {"timestamp": 1774099602.7806635, "event": "train_step", "step": 180, "epoch": 2, "metrics": {"train/step_loss": 1.5966231734664351, "train/step_real_loss": 1.2403349727392197, "train/lr": 9.804501125681243e-06, "train/step_canary_loss": 8.111607415335518, "perf/step_duration_sec": 17.99831911176443, "perf/samples_per_sec": 7.445139691539982, "perf/tokens_per_sec": 5983.892125215344, "perf/logical_batch_size": 134.0, "perf/logical_token_count": 107700.0, "perf/physical_batches": 18.0, "privacy/epsilon": 2.780402267783889, "system/cuda_memory_allocated_gb": 16.726882934570312, "system/cuda_max_memory_allocated_gb": 74.69534540176392}}
24
+ {"timestamp": 1774099786.8860722, "event": "train_step", "step": 190, "epoch": 2, "metrics": {"train/step_loss": 1.301622830904447, "train/step_real_loss": 1.246277093887329, "train/lr": 4.091647429802869e-06, "train/step_canary_loss": 4.84375, "perf/step_duration_sec": 17.514903953298926, "perf/samples_per_sec": 7.3651560033649455, "perf/tokens_per_sec": 6243.596898480441, "perf/logical_batch_size": 129.0, "perf/logical_token_count": 109356.0, "perf/physical_batches": 18.0, "privacy/epsilon": 2.860377969759561, "system/cuda_memory_allocated_gb": 16.253358840942383, "system/cuda_max_memory_allocated_gb": 74.69534540176392}}
25
+ {"timestamp": 1774099963.4199688, "event": "train_step", "step": 200, "epoch": 2, "metrics": {"train/step_loss": 1.5310632604541201, "train/step_real_loss": 1.2371121123433113, "train/lr": 8.126960406835249e-07, "train/step_canary_loss": 10.9375, "perf/step_duration_sec": 16.793452875688672, "perf/samples_per_sec": 7.860206056319249, "perf/tokens_per_sec": 6343.96039865215, "perf/logical_batch_size": 132.0, "perf/logical_token_count": 106537.0, "perf/physical_batches": 17.0, "privacy/epsilon": 2.938565800812133, "system/cuda_memory_allocated_gb": 16.537474155426025, "system/cuda_max_memory_allocated_gb": 74.69534540176392}}
26
+ {"timestamp": 1774099990.6645026, "event": "eval_step", "step": 200, "epoch": 2, "metrics": {"eval/loss": 0.8425621900750303, "eval/duration_sec": 27.24077794700861}}
27
+ {"timestamp": 1774100151.6417675, "event": "train_epoch", "step": 208, "epoch": 2, "metrics": {"train/epoch_loss": 1.6631105757675884, "train/epoch_real_loss": 1.2492633356516678, "train/epoch_canary_loss": 9.74998009501868, "perf/epoch_duration_sec": 1886.0955602359027, "perf/epoch_samples_per_sec": 29.488961838732855, "perf/epoch_tokens_per_sec": 23327.959583605018, "perf/epoch_samples": 55619.0, "perf/epoch_tokens": 43998761.0, "system/cuda_epoch_peak_memory_gb": 74.69534540176392, "eval/loss": 0.8425623961221212, "eval/duration_sec": 27.067450830712914, "privacy/epsilon": 2.9999680995370417}}
28
+ {"timestamp": 1774100161.8989484, "event": "audit_epoch", "step": 208, "epoch": 2, "metrics": {"audit/delta": 1e-05, "audit/num_canaries": 500.0, "audit/num_members": 250.0, "audit/paper_guess_fraction": 0.2, "audit/paper_guess_steps": 20.0, "audit/loss/auc": 0.525808, "audit/loss/empirical_epsilon/0.05": 0.05073561053723097, "audit/loss/empirical_epsilon/0.01": 0.0, "audit/loss/empirical_epsilon_details/0.05/epsilon": 0.05073561053723097, "audit/loss/empirical_epsilon_details/0.05/num_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.05/correct_guesses": 60.0, "audit/loss/empirical_epsilon_details/0.01/epsilon": 0.0, "audit/loss/empirical_epsilon_details/0.01/num_guesses": 0.0, "audit/loss/empirical_epsilon_details/0.01/correct_guesses": 0.0, "audit/embedding/auc": 0.520928, "audit/embedding/empirical_epsilon/0.05": 0.0, "audit/embedding/empirical_epsilon/0.01": 0.0, "audit/embedding/empirical_epsilon_details/0.05/epsilon": 0.0, "audit/embedding/empirical_epsilon_details/0.05/num_guesses": 0.0, "audit/embedding/empirical_epsilon_details/0.05/correct_guesses": 0.0, "audit/embedding/empirical_epsilon_details/0.01/epsilon": 0.0, "audit/embedding/empirical_epsilon_details/0.01/num_guesses": 0.0, "audit/embedding/empirical_epsilon_details/0.01/correct_guesses": 0.0, "perf/audit_duration_sec": 5.841585950925946}}
29
+ {"timestamp": 1774100172.4001799, "event": "audit_final", "step": 208, "epoch": 2, "metrics": {"audit/delta": 1e-05, "audit/num_canaries": 500.0, "audit/num_members": 250.0, "audit/paper_guess_fraction": 0.2, "audit/paper_guess_steps": 20.0, "audit/loss/auc": 0.525808, "audit/loss/empirical_epsilon/0.05": 0.05073561053723097, "audit/loss/empirical_epsilon/0.01": 0.0, "audit/loss/empirical_epsilon_details/0.05/epsilon": 0.05073561053723097, "audit/loss/empirical_epsilon_details/0.05/num_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.05/correct_guesses": 60.0, "audit/loss/empirical_epsilon_details/0.01/epsilon": 0.0, "audit/loss/empirical_epsilon_details/0.01/num_guesses": 0.0, "audit/loss/empirical_epsilon_details/0.01/correct_guesses": 0.0, "audit/embedding/auc": 0.520928, "audit/embedding/empirical_epsilon/0.05": 0.0, "audit/embedding/empirical_epsilon/0.01": 0.0, "audit/embedding/empirical_epsilon_details/0.05/epsilon": 0.0, "audit/embedding/empirical_epsilon_details/0.05/num_guesses": 0.0, "audit/embedding/empirical_epsilon_details/0.05/correct_guesses": 0.0, "audit/embedding/empirical_epsilon_details/0.01/epsilon": 0.0, "audit/embedding/empirical_epsilon_details/0.01/num_guesses": 0.0, "audit/embedding/empirical_epsilon_details/0.01/correct_guesses": 0.0}}
30
+ {"timestamp": 1774100172.924808, "event": "energy_final", "step": 208, "epoch": null, "metrics": {"energy/codecarbon/duration": 3938.0607158355415, "energy/codecarbon/emissions": 0.5702959606126596, "energy/codecarbon/emissions_rate": 0.00014481644691749236, "energy/codecarbon/cpu_power": 179.92904019353898, "energy/codecarbon/gpu_power": 2298.550082608512, "energy/codecarbon/ram_power": 70.0, "energy/codecarbon/cpu_energy": 0.18975838183590477, "energy/codecarbon/gpu_energy": 2.514705143984827, "energy/codecarbon/ram_energy": 0.07382037755078118, "energy/codecarbon/energy_consumed": 2.778283903371513, "energy/codecarbon/water_consumed": 0.0, "energy/codecarbon/cpu_count": 224.0, "energy/codecarbon/gpu_count": 4.0, "energy/codecarbon/longitude": -121.9552, "energy/codecarbon/latitude": 37.3541, "energy/codecarbon/ram_total_size": 2015.5625190734863, "energy/codecarbon/cpu_utilization_percent": 2.4580389144905275, "energy/codecarbon/gpu_utilization_percent": 95.87682458386683, "energy/codecarbon/ram_utilization_percent": 2.36410650281618, "energy/codecarbon/ram_used_gb": 47.93522234837092, "energy/codecarbon/pue": 1.0, "energy/codecarbon/wue": 0.0}}
starcoder2-7b/dp3/pretrain_lm_head.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9a9b40bb4ecc60cfc3a7e2f7cd31be2121d8f32f92d1a4ecdac554172fb4e5b
3
+ size 457594472
starcoder2-7b/dp3/resolved_config.yaml ADDED
@@ -0,0 +1,101 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ name: bigcode/starcoder2-7b
3
+ tokenizer_name: bigcode/starcoder2-7b
4
+ max_length: 1024
5
+ dtype: bfloat16
6
+ trust_remote_code: true
7
+ use_fast_tokenizer: true
8
+ cache_dir: null
9
+ local_files_only: false
10
+ low_cpu_mem_usage: true
11
+ tie_word_embeddings: true
12
+ gradient_checkpointing: false
13
+ use_chat_template: false
14
+ dataset:
15
+ name: melihcatal/codedp-cpt
16
+ split: train
17
+ mode: cpt
18
+ text_column: text
19
+ validation_ratio: 0.05
20
+ max_samples: -1
21
+ lora:
22
+ enabled: true
23
+ r: 16
24
+ alpha: 32
25
+ dropout: 0.05
26
+ target_modules:
27
+ - q_proj
28
+ - k_proj
29
+ - v_proj
30
+ - o_proj
31
+ modules_to_save:
32
+ - lm_head
33
+ bias: none
34
+ training:
35
+ seed: 42
36
+ epochs: 2
37
+ warmup_steps: null
38
+ warmup_ratio: 0.05
39
+ mixed_precision: false
40
+ mixed_precision_dtype: bfloat16
41
+ batch_size: 8
42
+ eval_batch_size: 8
43
+ eval_every_steps: 50
44
+ eval_every_epochs: 1
45
+ learning_rate: 0.0002
46
+ optimizer: adamw
47
+ lr_scheduler: cosine
48
+ adam_beta1: 0.9
49
+ adam_beta2: 0.999
50
+ adam_epsilon: 1.0e-08
51
+ sgd_momentum: 0.9
52
+ weight_decay: 0.01
53
+ max_grad_norm: 1.0
54
+ log_every: 10
55
+ gradient_accumulation_steps: 16
56
+ num_workers: 4
57
+ output_dir: runs/cpt/starcoder2-7b/dp3
58
+ distributed:
59
+ strategy: dpddp
60
+ backend: nccl
61
+ devices: null
62
+ dp:
63
+ module_validator: auto
64
+ target_delta: 1.0e-05
65
+ noise_multiplier: null
66
+ max_grad_norm: 1.0
67
+ grad_sample_mode: hooks
68
+ clipping: flat
69
+ secure_mode: false
70
+ enabled: true
71
+ target_epsilon: 3.0
72
+ audit:
73
+ enabled: true
74
+ run_every_epoch: true
75
+ epoch_device: cuda
76
+ q_canary: auto
77
+ num_canaries: 500
78
+ prefix_length: 49
79
+ num_digits: 12
80
+ batch_size: 32
81
+ delta: 1.0e-05
82
+ p_values:
83
+ - 0.05
84
+ - 0.01
85
+ paper_guess_fraction: 0.2
86
+ paper_guess_steps: 20
87
+ enable_holdout_empirical_epsilon: false
88
+ holdout_seed: 42
89
+ tie_seed: 42
90
+ tracking:
91
+ enabled: true
92
+ tensorboard: true
93
+ wandb: false
94
+ wandb_project: codedp-finetune-h200-audit
95
+ wandb_run_name: starcoder2-7b-cpt-dp3
96
+ wandb_mode: online
97
+ codecarbon: true
98
+ codecarbon_output_file: codecarbon.csv
99
+ codecarbon_measure_power_secs: 15
100
+ codecarbon_country_iso_code: null
101
+ codecarbon_project_name: codedp-starcoder2-7b-cpt-dp3
starcoder2-7b/dp3/scalars.csv ADDED
@@ -0,0 +1,386 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ timestamp,event,step,epoch,key,value
2
+ 1774096500.8353336,train_step,10,1,train/step_loss,1.6323321043555417
3
+ 1774096500.8353336,train_step,10,1,train/step_real_loss,1.330917976796627
4
+ 1774096500.8353336,train_step,10,1,train/lr,0.00018181818181818183
5
+ 1774096500.8353336,train_step,10,1,train/step_canary_loss,8.062500158945719
6
+ 1774096500.8353336,train_step,10,1,perf/step_duration_sec,17.560116425156593
7
+ 1774096500.8353336,train_step,10,1,perf/samples_per_sec,7.573981674145647
8
+ 1774096500.8353336,train_step,10,1,perf/tokens_per_sec,6185.665138551686
9
+ 1774096500.8353336,train_step,10,1,perf/logical_batch_size,133.0
10
+ 1774096500.8353336,train_step,10,1,perf/logical_token_count,108621.0
11
+ 1774096500.8353336,train_step,10,1,perf/physical_batches,18.0
12
+ 1774096500.8353336,train_step,10,1,privacy/epsilon,0.7131647471248268
13
+ 1774096500.8353336,train_step,10,1,system/cuda_memory_allocated_gb,16.63217782974243
14
+ 1774096500.8353336,train_step,10,1,system/cuda_max_memory_allocated_gb,74.69534254074097
15
+ 1774096676.393492,train_step,20,1,train/step_loss,1.826857232699429
16
+ 1774096676.393492,train_step,20,1,train/step_real_loss,1.3395854756236076
17
+ 1774096676.393492,train_step,20,1,train/lr,0.00019897180218885507
18
+ 1774096676.393492,train_step,20,1,train/step_canary_loss,8.756944444444445
19
+ 1774096676.393492,train_step,20,1,perf/step_duration_sec,18.843991773203015
20
+ 1774096676.393492,train_step,20,1,perf/samples_per_sec,7.217154498729827
21
+ 1774096676.393492,train_step,20,1,perf/tokens_per_sec,5434.1989336686165
22
+ 1774096676.393492,train_step,20,1,perf/logical_batch_size,136.0
23
+ 1774096676.393492,train_step,20,1,perf/logical_token_count,102402.0
24
+ 1774096676.393492,train_step,20,1,perf/physical_batches,18.0
25
+ 1774096676.393492,train_step,20,1,privacy/epsilon,0.9542480237478311
26
+ 1774096676.393492,train_step,20,1,system/cuda_memory_allocated_gb,16.916476249694824
27
+ 1774096676.393492,train_step,20,1,system/cuda_max_memory_allocated_gb,74.69534397125244
28
+ 1774096852.7017646,train_step,30,1,train/step_loss,1.5258880311792546
29
+ 1774096852.7017646,train_step,30,1,train/step_real_loss,1.3621462360024452
30
+ 1774096852.7017646,train_step,30,1,train/lr,0.00019544467510209388
31
+ 1774096852.7017646,train_step,30,1,train/step_canary_loss,6.765625476837158
32
+ 1774096852.7017646,train_step,30,1,perf/step_duration_sec,18.006671600043774
33
+ 1774096852.7017646,train_step,30,1,perf/samples_per_sec,7.275081309289916
34
+ 1774096852.7017646,train_step,30,1,perf/tokens_per_sec,5874.1561099910805
35
+ 1774096852.7017646,train_step,30,1,perf/logical_batch_size,131.0
36
+ 1774096852.7017646,train_step,30,1,perf/logical_token_count,105774.0
37
+ 1774096852.7017646,train_step,30,1,perf/physical_batches,18.0
38
+ 1774096852.7017646,train_step,30,1,privacy/epsilon,1.145336977893831
39
+ 1774096852.7017646,train_step,30,1,system/cuda_memory_allocated_gb,16.442769050598145
40
+ 1774096852.7017646,train_step,30,1,system/cuda_max_memory_allocated_gb,74.69534397125244
41
+ 1774097029.0227332,train_step,40,1,train/step_loss,1.6870967794347693
42
+ 1774097029.0227332,train_step,40,1,train/step_real_loss,1.2007466033101082
43
+ 1774097029.0227332,train_step,40,1,train/lr,0.00018949541262593762
44
+ 1774097029.0227332,train_step,40,1,train/step_canary_loss,10.580357142857142
45
+ 1774097029.0227332,train_step,40,1,perf/step_duration_sec,17.532358843833208
46
+ 1774097029.0227332,train_step,40,1,perf/samples_per_sec,7.643010344106256
47
+ 1774097029.0227332,train_step,40,1,perf/tokens_per_sec,6436.156195815627
48
+ 1774097029.0227332,train_step,40,1,perf/logical_batch_size,134.0
49
+ 1774097029.0227332,train_step,40,1,perf/logical_token_count,112841.0
50
+ 1774097029.0227332,train_step,40,1,perf/physical_batches,18.0
51
+ 1774097029.0227332,train_step,40,1,privacy/epsilon,1.311193108872294
52
+ 1774097029.0227332,train_step,40,1,system/cuda_memory_allocated_gb,16.726882934570312
53
+ 1774097029.0227332,train_step,40,1,system/cuda_max_memory_allocated_gb,74.69534683227539
54
+ 1774097204.3534977,train_step,50,1,train/step_loss,1.5818433902881763
55
+ 1774097204.3534977,train_step,50,1,train/step_real_loss,1.2425691708922386
56
+ 1774097204.3534977,train_step,50,1,train/lr,0.00018127499143005268
57
+ 1774097204.3534977,train_step,50,1,train/step_canary_loss,7.785714830671038
58
+ 1774097204.3534977,train_step,50,1,perf/step_duration_sec,17.701514894142747
59
+ 1774097204.3534977,train_step,50,1,perf/samples_per_sec,7.569973575783576
60
+ 1774097204.3534977,train_step,50,1,perf/tokens_per_sec,6216.078152520672
61
+ 1774097204.3534977,train_step,50,1,perf/logical_batch_size,134.0
62
+ 1774097204.3534977,train_step,50,1,perf/logical_token_count,110034.0
63
+ 1774097204.3534977,train_step,50,1,perf/physical_batches,18.0
64
+ 1774097204.3534977,train_step,50,1,privacy/epsilon,1.4583641061524852
65
+ 1774097204.3534977,train_step,50,1,system/cuda_memory_allocated_gb,16.726882934570312
66
+ 1774097204.3534977,train_step,50,1,system/cuda_max_memory_allocated_gb,74.69534683227539
67
+ 1774097231.3920577,eval_step,50,1,eval/loss,0.8423196817266531
68
+ 1774097231.3920577,eval_step,50,1,eval/duration_sec,27.033981669694185
69
+ 1774097408.5970793,train_step,60,1,train/step_loss,1.6164713336112804
70
+ 1774097408.5970793,train_step,60,1,train/step_real_loss,1.3143803551793098
71
+ 1774097408.5970793,train_step,60,1,train/lr,0.0001709920242324663
72
+ 1774097408.5970793,train_step,60,1,train/step_canary_loss,9.350000381469727
73
+ 1774097408.5970793,train_step,60,1,perf/step_duration_sec,17.616566620767117
74
+ 1774097408.5970793,train_step,60,1,perf/samples_per_sec,7.549711749349289
75
+ 1774097408.5970793,train_step,60,1,perf/tokens_per_sec,5881.736335493052
76
+ 1774097408.5970793,train_step,60,1,perf/logical_batch_size,133.0
77
+ 1774097408.5970793,train_step,60,1,perf/logical_token_count,103616.0
78
+ 1774097408.5970793,train_step,60,1,perf/physical_batches,17.0
79
+ 1774097408.5970793,train_step,60,1,privacy/epsilon,1.5929146459722179
80
+ 1774097408.5970793,train_step,60,1,system/cuda_memory_allocated_gb,16.63217782974243
81
+ 1774097408.5970793,train_step,60,1,system/cuda_max_memory_allocated_gb,74.69534969329834
82
+ 1774097584.0989442,train_step,70,1,train/step_loss,1.5537053346633911
83
+ 1774097584.0989442,train_step,70,1,train/step_real_loss,1.274133626371622
84
+ 1774097584.0989442,train_step,70,1,train/lr,0.00015890746575622231
85
+ 1774097584.0989442,train_step,70,1,train/step_canary_loss,10.5
86
+ 1774097584.0989442,train_step,70,1,perf/step_duration_sec,17.34652972780168
87
+ 1774097584.0989442,train_step,70,1,perf/samples_per_sec,7.609591201889827
88
+ 1774097584.0989442,train_step,70,1,perf/tokens_per_sec,6165.613622913036
89
+ 1774097584.0989442,train_step,70,1,perf/logical_batch_size,132.0
90
+ 1774097584.0989442,train_step,70,1,perf/logical_token_count,106952.0
91
+ 1774097584.0989442,train_step,70,1,perf/physical_batches,17.0
92
+ 1774097584.0989442,train_step,70,1,privacy/epsilon,1.7182200620091768
93
+ 1774097584.0989442,train_step,70,1,system/cuda_memory_allocated_gb,16.537474155426025
94
+ 1774097584.0989442,train_step,70,1,system/cuda_max_memory_allocated_gb,74.69534969329834
95
+ 1774097758.6071105,train_step,80,1,train/step_loss,1.6813924115403254
96
+ 1774097758.6071105,train_step,80,1,train/step_real_loss,1.3261733427643776
97
+ 1774097758.6071105,train_step,80,1,train/lr,0.00014532799038330385
98
+ 1774097758.6071105,train_step,80,1,train/step_canary_loss,10.77500057220459
99
+ 1774097758.6071105,train_step,80,1,perf/step_duration_sec,16.975525552406907
100
+ 1774097758.6071105,train_step,80,1,perf/samples_per_sec,7.834808977748694
101
+ 1774097758.6071105,train_step,80,1,perf/tokens_per_sec,6001.228043602786
102
+ 1774097758.6071105,train_step,80,1,perf/logical_batch_size,133.0
103
+ 1774097758.6071105,train_step,80,1,perf/logical_token_count,101874.0
104
+ 1774097758.6071105,train_step,80,1,perf/physical_batches,17.0
105
+ 1774097758.6071105,train_step,80,1,privacy/epsilon,1.8360751937320303
106
+ 1774097758.6071105,train_step,80,1,system/cuda_memory_allocated_gb,16.63217782974243
107
+ 1774097758.6071105,train_step,80,1,system/cuda_max_memory_allocated_gb,74.69534969329834
108
+ 1774097934.2494926,train_step,90,1,train/step_loss,1.8558794460157408
109
+ 1774097934.2494926,train_step,90,1,train/step_real_loss,1.2407655008137226
110
+ 1774097934.2494926,train_step,90,1,train/lr,0.00013059820956358998
111
+ 1774097934.2494926,train_step,90,1,train/step_canary_loss,10.604166666666666
112
+ 1774097934.2494926,train_step,90,1,perf/step_duration_sec,17.994612561538815
113
+ 1774097934.2494926,train_step,90,1,perf/samples_per_sec,7.557817626520207
114
+ 1774097934.2494926,train_step,90,1,perf/tokens_per_sec,6203.967971981345
115
+ 1774097934.2494926,train_step,90,1,perf/logical_batch_size,136.0
116
+ 1774097934.2494926,train_step,90,1,perf/logical_token_count,111638.0
117
+ 1774097934.2494926,train_step,90,1,perf/physical_batches,18.0
118
+ 1774097934.2494926,train_step,90,1,privacy/epsilon,1.9476934830264792
119
+ 1774097934.2494926,train_step,90,1,system/cuda_memory_allocated_gb,16.916476249694824
120
+ 1774097934.2494926,train_step,90,1,system/cuda_max_memory_allocated_gb,74.69534969329834
121
+ 1774098110.4647262,train_step,100,1,train/step_loss,1.86213939777319
122
+ 1774098110.4647262,train_step,100,1,train/step_real_loss,1.2256366163492203
123
+ 1774098110.4647262,train_step,100,1,train/lr,0.00011509192648058249
124
+ 1774098110.4647262,train_step,100,1,train/step_canary_loss,10.009375
125
+ 1774098110.4647262,train_step,100,1,perf/step_duration_sec,17.502219082787633
126
+ 1774098110.4647262,train_step,100,1,perf/samples_per_sec,7.884714466619527
127
+ 1774098110.4647262,train_step,100,1,perf/tokens_per_sec,6282.689039593839
128
+ 1774098110.4647262,train_step,100,1,perf/logical_batch_size,138.0
129
+ 1774098110.4647262,train_step,100,1,perf/logical_token_count,109961.0
130
+ 1774098110.4647262,train_step,100,1,perf/physical_batches,18.0
131
+ 1774098110.4647262,train_step,100,1,privacy/epsilon,2.054057754637441
132
+ 1774098110.4647262,train_step,100,1,system/cuda_memory_allocated_gb,16.34807252883911
133
+ 1774098110.4647262,train_step,100,1,system/cuda_max_memory_allocated_gb,74.69534969329834
134
+ 1774098137.4743512,eval_step,100,1,eval/loss,0.8425039929905157
135
+ 1774098137.4743512,eval_step,100,1,eval/duration_sec,27.006801065057516
136
+ 1774098228.5487797,train_epoch,104,1,train/epoch_loss,1.7094427731604633
137
+ 1774098228.5487797,train_epoch,104,1,train/epoch_real_loss,1.285309421313279
138
+ 1774098228.5487797,train_epoch,104,1,train/epoch_canary_loss,9.719154772066274
139
+ 1774098228.5487797,train_epoch,104,1,perf/epoch_duration_sec,1878.0538540109992
140
+ 1774098228.5487797,train_epoch,104,1,perf/epoch_samples_per_sec,29.64824458099586
141
+ 1774098228.5487797,train_epoch,104,1,perf/epoch_tokens_per_sec,23429.532069074678
142
+ 1774098228.5487797,train_epoch,104,1,perf/epoch_samples,55681.0
143
+ 1774098228.5487797,train_epoch,104,1,perf/epoch_tokens,44001923.0
144
+ 1774098228.5487797,train_epoch,104,1,system/cuda_epoch_peak_memory_gb,74.69534969329834
145
+ 1774098228.5487797,train_epoch,104,1,eval/loss,0.8425441932404178
146
+ 1774098228.5487797,train_epoch,104,1,eval/duration_sec,26.999734753742814
147
+ 1774098228.5487797,train_epoch,104,1,privacy/epsilon,2.0952814257505974
148
+ 1774098238.4572744,audit_epoch,104,1,audit/delta,1e-05
149
+ 1774098238.4572744,audit_epoch,104,1,audit/num_canaries,500.0
150
+ 1774098238.4572744,audit_epoch,104,1,audit/num_members,250.0
151
+ 1774098238.4572744,audit_epoch,104,1,audit/paper_guess_fraction,0.2
152
+ 1774098238.4572744,audit_epoch,104,1,audit/paper_guess_steps,20.0
153
+ 1774098238.4572744,audit_epoch,104,1,audit/loss/auc,0.527168
154
+ 1774098238.4572744,audit_epoch,104,1,audit/loss/empirical_epsilon/0.05,0.09125839080661535
155
+ 1774098238.4572744,audit_epoch,104,1,audit/loss/empirical_epsilon/0.01,0.0
156
+ 1774098238.4572744,audit_epoch,104,1,audit/loss/empirical_epsilon_details/0.05/epsilon,0.09125839080661535
157
+ 1774098238.4572744,audit_epoch,104,1,audit/loss/empirical_epsilon_details/0.05/num_guesses,100.0
158
+ 1774098238.4572744,audit_epoch,104,1,audit/loss/empirical_epsilon_details/0.05/correct_guesses,61.0
159
+ 1774098238.4572744,audit_epoch,104,1,audit/loss/empirical_epsilon_details/0.01/epsilon,0.0
160
+ 1774098238.4572744,audit_epoch,104,1,audit/loss/empirical_epsilon_details/0.01/num_guesses,0.0
161
+ 1774098238.4572744,audit_epoch,104,1,audit/loss/empirical_epsilon_details/0.01/correct_guesses,0.0
162
+ 1774098238.4572744,audit_epoch,104,1,audit/embedding/auc,0.514792
163
+ 1774098238.4572744,audit_epoch,104,1,audit/embedding/empirical_epsilon/0.05,0.0
164
+ 1774098238.4572744,audit_epoch,104,1,audit/embedding/empirical_epsilon/0.01,0.0
165
+ 1774098238.4572744,audit_epoch,104,1,audit/embedding/empirical_epsilon_details/0.05/epsilon,0.0
166
+ 1774098238.4572744,audit_epoch,104,1,audit/embedding/empirical_epsilon_details/0.05/num_guesses,0.0
167
+ 1774098238.4572744,audit_epoch,104,1,audit/embedding/empirical_epsilon_details/0.05/correct_guesses,0.0
168
+ 1774098238.4572744,audit_epoch,104,1,audit/embedding/empirical_epsilon_details/0.01/epsilon,0.0
169
+ 1774098238.4572744,audit_epoch,104,1,audit/embedding/empirical_epsilon_details/0.01/num_guesses,0.0
170
+ 1774098238.4572744,audit_epoch,104,1,audit/embedding/empirical_epsilon_details/0.01/correct_guesses,0.0
171
+ 1774098238.4572744,audit_epoch,104,1,perf/audit_duration_sec,5.7471471559256315
172
+ 1774098344.630848,train_step,110,2,train/step_loss,1.570878519940732
173
+ 1774098344.630848,train_step,110,2,train/step_real_loss,1.2690251618623734
174
+ 1774098344.630848,train_step,110,2,train/lr,9.920264990753837e-05
175
+ 1774098344.630848,train_step,110,2,train/step_canary_loss,8.010416825612387
176
+ 1774098344.630848,train_step,110,2,perf/step_duration_sec,17.30536743812263
177
+ 1774098344.630848,train_step,110,2,perf/samples_per_sec,7.685476802243991
178
+ 1774098344.630848,train_step,110,2,perf/tokens_per_sec,6186.173184867881
179
+ 1774098344.630848,train_step,110,2,perf/logical_batch_size,133.0
180
+ 1774098344.630848,train_step,110,2,perf/logical_token_count,107054.0
181
+ 1774098344.630848,train_step,110,2,perf/physical_batches,18.0
182
+ 1774098344.630848,train_step,110,2,privacy/epsilon,2.155895236118489
183
+ 1774098344.630848,train_step,110,2,system/cuda_memory_allocated_gb,16.63217782974243
184
+ 1774098344.630848,train_step,110,2,system/cuda_max_memory_allocated_gb,74.69534254074097
185
+ 1774098520.7483177,train_step,120,2,train/step_loss,1.733701615437974
186
+ 1774098520.7483177,train_step,120,2,train/step_real_loss,1.2730829790234566
187
+ 1774098520.7483177,train_step,120,2,train/lr,8.333360798744496e-05
188
+ 1774098520.7483177,train_step,120,2,train/step_canary_loss,8.284722222222221
189
+ 1774098520.7483177,train_step,120,2,perf/step_duration_sec,18.148367216810584
190
+ 1774098520.7483177,train_step,120,2,perf/samples_per_sec,7.493787092539381
191
+ 1774098520.7483177,train_step,120,2,perf/tokens_per_sec,5505.233545608104
192
+ 1774098520.7483177,train_step,120,2,perf/logical_batch_size,136.0
193
+ 1774098520.7483177,train_step,120,2,perf/logical_token_count,99911.0
194
+ 1774098520.7483177,train_step,120,2,perf/physical_batches,18.0
195
+ 1774098520.7483177,train_step,120,2,privacy/epsilon,2.2538737991591966
196
+ 1774098520.7483177,train_step,120,2,system/cuda_memory_allocated_gb,16.916476249694824
197
+ 1774098520.7483177,train_step,120,2,system/cuda_max_memory_allocated_gb,74.69534254074097
198
+ 1774098695.815589,train_step,130,2,train/step_loss,1.4695805564071194
199
+ 1774098695.815589,train_step,130,2,train/step_real_loss,1.2108174487948418
200
+ 1774098695.815589,train_step,130,2,train/lr,6.788751536089739e-05
201
+ 1774098695.815589,train_step,130,2,train/step_canary_loss,9.75
202
+ 1774098695.815589,train_step,130,2,perf/step_duration_sec,16.7934517133981
203
+ 1774098695.815589,train_step,130,2,perf/samples_per_sec,7.860206600331495
204
+ 1774098695.815589,train_step,130,2,perf/tokens_per_sec,6584.88807942771
205
+ 1774098695.815589,train_step,130,2,perf/logical_batch_size,132.0
206
+ 1774098695.815589,train_step,130,2,perf/logical_token_count,110583.0
207
+ 1774098695.815589,train_step,130,2,perf/physical_batches,17.0
208
+ 1774098695.815589,train_step,130,2,privacy/epsilon,2.34829265801524
209
+ 1774098695.815589,train_step,130,2,system/cuda_memory_allocated_gb,16.537474155426025
210
+ 1774098695.815589,train_step,130,2,system/cuda_max_memory_allocated_gb,74.69534254074097
211
+ 1774098871.6479974,train_step,140,2,train/step_loss,1.409767813354958
212
+ 1774098871.6479974,train_step,140,2,train/step_real_loss,1.280211590230465
213
+ 1774098871.6479974,train_step,140,2,train/lr,5.325635332531864e-05
214
+ 1774098871.6479974,train_step,140,2,train/step_canary_loss,6.9375
215
+ 1774098871.6479974,train_step,140,2,perf/step_duration_sec,17.296851823106408
216
+ 1774098871.6479974,train_step,140,2,perf/samples_per_sec,7.515818562215838
217
+ 1774098871.6479974,train_step,140,2,perf/tokens_per_sec,5869.218343212228
218
+ 1774098871.6479974,train_step,140,2,perf/logical_batch_size,130.0
219
+ 1774098871.6479974,train_step,140,2,perf/logical_token_count,101519.0
220
+ 1774098871.6479974,train_step,140,2,perf/physical_batches,18.0
221
+ 1774098871.6479974,train_step,140,2,privacy/epsilon,2.4397183333948855
222
+ 1774098871.6479974,train_step,140,2,system/cuda_memory_allocated_gb,16.348063945770264
223
+ 1774098871.6479974,train_step,140,2,system/cuda_max_memory_allocated_gb,74.69534254074097
224
+ 1774099047.329602,train_step,150,2,train/step_loss,1.470641034250041
225
+ 1774099047.329602,train_step,150,2,train/step_real_loss,1.2892888560891151
226
+ 1774099047.329602,train_step,150,2,train/lr,3.981142237826332e-05
227
+ 1774099047.329602,train_step,150,2,train/step_canary_loss,9.208333969116211
228
+ 1774099047.329602,train_step,150,2,perf/step_duration_sec,17.535365080460906
229
+ 1774099047.329602,train_step,150,2,perf/samples_per_sec,7.4706172012334715
230
+ 1774099047.329602,train_step,150,2,perf/tokens_per_sec,5992.347437184814
231
+ 1774099047.329602,train_step,150,2,perf/logical_batch_size,131.0
232
+ 1774099047.329602,train_step,150,2,perf/logical_token_count,105078.0
233
+ 1774099047.329602,train_step,150,2,perf/physical_batches,17.0
234
+ 1774099047.329602,train_step,150,2,privacy/epsilon,2.5283910517887054
235
+ 1774099047.329602,train_step,150,2,system/cuda_memory_allocated_gb,16.442769050598145
236
+ 1774099047.329602,train_step,150,2,system/cuda_max_memory_allocated_gb,74.69534254074097
237
+ 1774099074.3730319,eval_step,150,2,eval/loss,0.842570746081999
238
+ 1774099074.3730319,eval_step,150,2,eval/duration_sec,27.040043698623776
239
+ 1774099249.8653634,train_step,160,2,train/step_loss,1.536541504646415
240
+ 1774099249.8653634,train_step,160,2,train/step_real_loss,1.2477270364761353
241
+ 1774099249.8653634,train_step,160,2,train/lr,2.789391958515183e-05
242
+ 1774099249.8653634,train_step,160,2,train/step_canary_loss,7.697916825612386
243
+ 1774099249.8653634,train_step,160,2,perf/step_duration_sec,17.38564824871719
244
+ 1774099249.8653634,train_step,160,2,perf/samples_per_sec,7.649987972683935
245
+ 1774099249.8653634,train_step,160,2,perf/tokens_per_sec,6056.662282222899
246
+ 1774099249.8653634,train_step,160,2,perf/logical_batch_size,133.0
247
+ 1774099249.8653634,train_step,160,2,perf/logical_token_count,105299.0
248
+ 1774099249.8653634,train_step,160,2,perf/physical_batches,18.0
249
+ 1774099249.8653634,train_step,160,2,privacy/epsilon,2.6145698431381854
250
+ 1774099249.8653634,train_step,160,2,system/cuda_memory_allocated_gb,16.63217782974243
251
+ 1774099249.8653634,train_step,160,2,system/cuda_max_memory_allocated_gb,74.69534540176392
252
+ 1774099424.502899,train_step,170,2,train/step_loss,1.4997789441174223
253
+ 1774099424.502899,train_step,170,2,train/step_real_loss,1.2844417318701744
254
+ 1774099424.502899,train_step,170,2,train/lr,1.7806279893114875e-05
255
+ 1774099424.502899,train_step,170,2,train/step_canary_loss,10.6875
256
+ 1774099424.502899,train_step,170,2,perf/step_duration_sec,17.294984631240368
257
+ 1774099424.502899,train_step,170,2,perf/samples_per_sec,7.574450211616343
258
+ 1774099424.502899,train_step,170,2,perf/tokens_per_sec,5730.794337970551
259
+ 1774099424.502899,train_step,170,2,perf/logical_batch_size,131.0
260
+ 1774099424.502899,train_step,170,2,perf/logical_token_count,99114.0
261
+ 1774099424.502899,train_step,170,2,perf/physical_batches,17.0
262
+ 1774099424.502899,train_step,170,2,privacy/epsilon,2.6985357679843074
263
+ 1774099424.502899,train_step,170,2,system/cuda_memory_allocated_gb,16.442769050598145
264
+ 1774099424.502899,train_step,170,2,system/cuda_max_memory_allocated_gb,74.69534540176392
265
+ 1774099602.7806635,train_step,180,2,train/step_loss,1.5966231734664351
266
+ 1774099602.7806635,train_step,180,2,train/step_real_loss,1.2403349727392197
267
+ 1774099602.7806635,train_step,180,2,train/lr,9.804501125681243e-06
268
+ 1774099602.7806635,train_step,180,2,train/step_canary_loss,8.111607415335518
269
+ 1774099602.7806635,train_step,180,2,perf/step_duration_sec,17.99831911176443
270
+ 1774099602.7806635,train_step,180,2,perf/samples_per_sec,7.445139691539982
271
+ 1774099602.7806635,train_step,180,2,perf/tokens_per_sec,5983.892125215344
272
+ 1774099602.7806635,train_step,180,2,perf/logical_batch_size,134.0
273
+ 1774099602.7806635,train_step,180,2,perf/logical_token_count,107700.0
274
+ 1774099602.7806635,train_step,180,2,perf/physical_batches,18.0
275
+ 1774099602.7806635,train_step,180,2,privacy/epsilon,2.780402267783889
276
+ 1774099602.7806635,train_step,180,2,system/cuda_memory_allocated_gb,16.726882934570312
277
+ 1774099602.7806635,train_step,180,2,system/cuda_max_memory_allocated_gb,74.69534540176392
278
+ 1774099786.8860722,train_step,190,2,train/step_loss,1.301622830904447
279
+ 1774099786.8860722,train_step,190,2,train/step_real_loss,1.246277093887329
280
+ 1774099786.8860722,train_step,190,2,train/lr,4.091647429802869e-06
281
+ 1774099786.8860722,train_step,190,2,train/step_canary_loss,4.84375
282
+ 1774099786.8860722,train_step,190,2,perf/step_duration_sec,17.514903953298926
283
+ 1774099786.8860722,train_step,190,2,perf/samples_per_sec,7.3651560033649455
284
+ 1774099786.8860722,train_step,190,2,perf/tokens_per_sec,6243.596898480441
285
+ 1774099786.8860722,train_step,190,2,perf/logical_batch_size,129.0
286
+ 1774099786.8860722,train_step,190,2,perf/logical_token_count,109356.0
287
+ 1774099786.8860722,train_step,190,2,perf/physical_batches,18.0
288
+ 1774099786.8860722,train_step,190,2,privacy/epsilon,2.860377969759561
289
+ 1774099786.8860722,train_step,190,2,system/cuda_memory_allocated_gb,16.253358840942383
290
+ 1774099786.8860722,train_step,190,2,system/cuda_max_memory_allocated_gb,74.69534540176392
291
+ 1774099963.4199688,train_step,200,2,train/step_loss,1.5310632604541201
292
+ 1774099963.4199688,train_step,200,2,train/step_real_loss,1.2371121123433113
293
+ 1774099963.4199688,train_step,200,2,train/lr,8.126960406835249e-07
294
+ 1774099963.4199688,train_step,200,2,train/step_canary_loss,10.9375
295
+ 1774099963.4199688,train_step,200,2,perf/step_duration_sec,16.793452875688672
296
+ 1774099963.4199688,train_step,200,2,perf/samples_per_sec,7.860206056319249
297
+ 1774099963.4199688,train_step,200,2,perf/tokens_per_sec,6343.96039865215
298
+ 1774099963.4199688,train_step,200,2,perf/logical_batch_size,132.0
299
+ 1774099963.4199688,train_step,200,2,perf/logical_token_count,106537.0
300
+ 1774099963.4199688,train_step,200,2,perf/physical_batches,17.0
301
+ 1774099963.4199688,train_step,200,2,privacy/epsilon,2.938565800812133
302
+ 1774099963.4199688,train_step,200,2,system/cuda_memory_allocated_gb,16.537474155426025
303
+ 1774099963.4199688,train_step,200,2,system/cuda_max_memory_allocated_gb,74.69534540176392
304
+ 1774099990.6645026,eval_step,200,2,eval/loss,0.8425621900750303
305
+ 1774099990.6645026,eval_step,200,2,eval/duration_sec,27.24077794700861
306
+ 1774100151.6417675,train_epoch,208,2,train/epoch_loss,1.6631105757675884
307
+ 1774100151.6417675,train_epoch,208,2,train/epoch_real_loss,1.2492633356516678
308
+ 1774100151.6417675,train_epoch,208,2,train/epoch_canary_loss,9.74998009501868
309
+ 1774100151.6417675,train_epoch,208,2,perf/epoch_duration_sec,1886.0955602359027
310
+ 1774100151.6417675,train_epoch,208,2,perf/epoch_samples_per_sec,29.488961838732855
311
+ 1774100151.6417675,train_epoch,208,2,perf/epoch_tokens_per_sec,23327.959583605018
312
+ 1774100151.6417675,train_epoch,208,2,perf/epoch_samples,55619.0
313
+ 1774100151.6417675,train_epoch,208,2,perf/epoch_tokens,43998761.0
314
+ 1774100151.6417675,train_epoch,208,2,system/cuda_epoch_peak_memory_gb,74.69534540176392
315
+ 1774100151.6417675,train_epoch,208,2,eval/loss,0.8425623961221212
316
+ 1774100151.6417675,train_epoch,208,2,eval/duration_sec,27.067450830712914
317
+ 1774100151.6417675,train_epoch,208,2,privacy/epsilon,2.9999680995370417
318
+ 1774100161.8989484,audit_epoch,208,2,audit/delta,1e-05
319
+ 1774100161.8989484,audit_epoch,208,2,audit/num_canaries,500.0
320
+ 1774100161.8989484,audit_epoch,208,2,audit/num_members,250.0
321
+ 1774100161.8989484,audit_epoch,208,2,audit/paper_guess_fraction,0.2
322
+ 1774100161.8989484,audit_epoch,208,2,audit/paper_guess_steps,20.0
323
+ 1774100161.8989484,audit_epoch,208,2,audit/loss/auc,0.525808
324
+ 1774100161.8989484,audit_epoch,208,2,audit/loss/empirical_epsilon/0.05,0.05073561053723097
325
+ 1774100161.8989484,audit_epoch,208,2,audit/loss/empirical_epsilon/0.01,0.0
326
+ 1774100161.8989484,audit_epoch,208,2,audit/loss/empirical_epsilon_details/0.05/epsilon,0.05073561053723097
327
+ 1774100161.8989484,audit_epoch,208,2,audit/loss/empirical_epsilon_details/0.05/num_guesses,100.0
328
+ 1774100161.8989484,audit_epoch,208,2,audit/loss/empirical_epsilon_details/0.05/correct_guesses,60.0
329
+ 1774100161.8989484,audit_epoch,208,2,audit/loss/empirical_epsilon_details/0.01/epsilon,0.0
330
+ 1774100161.8989484,audit_epoch,208,2,audit/loss/empirical_epsilon_details/0.01/num_guesses,0.0
331
+ 1774100161.8989484,audit_epoch,208,2,audit/loss/empirical_epsilon_details/0.01/correct_guesses,0.0
332
+ 1774100161.8989484,audit_epoch,208,2,audit/embedding/auc,0.520928
333
+ 1774100161.8989484,audit_epoch,208,2,audit/embedding/empirical_epsilon/0.05,0.0
334
+ 1774100161.8989484,audit_epoch,208,2,audit/embedding/empirical_epsilon/0.01,0.0
335
+ 1774100161.8989484,audit_epoch,208,2,audit/embedding/empirical_epsilon_details/0.05/epsilon,0.0
336
+ 1774100161.8989484,audit_epoch,208,2,audit/embedding/empirical_epsilon_details/0.05/num_guesses,0.0
337
+ 1774100161.8989484,audit_epoch,208,2,audit/embedding/empirical_epsilon_details/0.05/correct_guesses,0.0
338
+ 1774100161.8989484,audit_epoch,208,2,audit/embedding/empirical_epsilon_details/0.01/epsilon,0.0
339
+ 1774100161.8989484,audit_epoch,208,2,audit/embedding/empirical_epsilon_details/0.01/num_guesses,0.0
340
+ 1774100161.8989484,audit_epoch,208,2,audit/embedding/empirical_epsilon_details/0.01/correct_guesses,0.0
341
+ 1774100161.8989484,audit_epoch,208,2,perf/audit_duration_sec,5.841585950925946
342
+ 1774100172.4001799,audit_final,208,2,audit/delta,1e-05
343
+ 1774100172.4001799,audit_final,208,2,audit/num_canaries,500.0
344
+ 1774100172.4001799,audit_final,208,2,audit/num_members,250.0
345
+ 1774100172.4001799,audit_final,208,2,audit/paper_guess_fraction,0.2
346
+ 1774100172.4001799,audit_final,208,2,audit/paper_guess_steps,20.0
347
+ 1774100172.4001799,audit_final,208,2,audit/loss/auc,0.525808
348
+ 1774100172.4001799,audit_final,208,2,audit/loss/empirical_epsilon/0.05,0.05073561053723097
349
+ 1774100172.4001799,audit_final,208,2,audit/loss/empirical_epsilon/0.01,0.0
350
+ 1774100172.4001799,audit_final,208,2,audit/loss/empirical_epsilon_details/0.05/epsilon,0.05073561053723097
351
+ 1774100172.4001799,audit_final,208,2,audit/loss/empirical_epsilon_details/0.05/num_guesses,100.0
352
+ 1774100172.4001799,audit_final,208,2,audit/loss/empirical_epsilon_details/0.05/correct_guesses,60.0
353
+ 1774100172.4001799,audit_final,208,2,audit/loss/empirical_epsilon_details/0.01/epsilon,0.0
354
+ 1774100172.4001799,audit_final,208,2,audit/loss/empirical_epsilon_details/0.01/num_guesses,0.0
355
+ 1774100172.4001799,audit_final,208,2,audit/loss/empirical_epsilon_details/0.01/correct_guesses,0.0
356
+ 1774100172.4001799,audit_final,208,2,audit/embedding/auc,0.520928
357
+ 1774100172.4001799,audit_final,208,2,audit/embedding/empirical_epsilon/0.05,0.0
358
+ 1774100172.4001799,audit_final,208,2,audit/embedding/empirical_epsilon/0.01,0.0
359
+ 1774100172.4001799,audit_final,208,2,audit/embedding/empirical_epsilon_details/0.05/epsilon,0.0
360
+ 1774100172.4001799,audit_final,208,2,audit/embedding/empirical_epsilon_details/0.05/num_guesses,0.0
361
+ 1774100172.4001799,audit_final,208,2,audit/embedding/empirical_epsilon_details/0.05/correct_guesses,0.0
362
+ 1774100172.4001799,audit_final,208,2,audit/embedding/empirical_epsilon_details/0.01/epsilon,0.0
363
+ 1774100172.4001799,audit_final,208,2,audit/embedding/empirical_epsilon_details/0.01/num_guesses,0.0
364
+ 1774100172.4001799,audit_final,208,2,audit/embedding/empirical_epsilon_details/0.01/correct_guesses,0.0
365
+ 1774100172.924808,energy_final,208,,energy/codecarbon/duration,3938.0607158355415
366
+ 1774100172.924808,energy_final,208,,energy/codecarbon/emissions,0.5702959606126596
367
+ 1774100172.924808,energy_final,208,,energy/codecarbon/emissions_rate,0.00014481644691749236
368
+ 1774100172.924808,energy_final,208,,energy/codecarbon/cpu_power,179.92904019353898
369
+ 1774100172.924808,energy_final,208,,energy/codecarbon/gpu_power,2298.550082608512
370
+ 1774100172.924808,energy_final,208,,energy/codecarbon/ram_power,70.0
371
+ 1774100172.924808,energy_final,208,,energy/codecarbon/cpu_energy,0.18975838183590477
372
+ 1774100172.924808,energy_final,208,,energy/codecarbon/gpu_energy,2.514705143984827
373
+ 1774100172.924808,energy_final,208,,energy/codecarbon/ram_energy,0.07382037755078118
374
+ 1774100172.924808,energy_final,208,,energy/codecarbon/energy_consumed,2.778283903371513
375
+ 1774100172.924808,energy_final,208,,energy/codecarbon/water_consumed,0.0
376
+ 1774100172.924808,energy_final,208,,energy/codecarbon/cpu_count,224.0
377
+ 1774100172.924808,energy_final,208,,energy/codecarbon/gpu_count,4.0
378
+ 1774100172.924808,energy_final,208,,energy/codecarbon/longitude,-121.9552
379
+ 1774100172.924808,energy_final,208,,energy/codecarbon/latitude,37.3541
380
+ 1774100172.924808,energy_final,208,,energy/codecarbon/ram_total_size,2015.5625190734863
381
+ 1774100172.924808,energy_final,208,,energy/codecarbon/cpu_utilization_percent,2.4580389144905275
382
+ 1774100172.924808,energy_final,208,,energy/codecarbon/gpu_utilization_percent,95.87682458386683
383
+ 1774100172.924808,energy_final,208,,energy/codecarbon/ram_utilization_percent,2.36410650281618
384
+ 1774100172.924808,energy_final,208,,energy/codecarbon/ram_used_gb,47.93522234837092
385
+ 1774100172.924808,energy_final,208,,energy/codecarbon/pue,1.0
386
+ 1774100172.924808,energy_final,208,,energy/codecarbon/wue,0.0
starcoder2-7b/dp3/summary.json ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "audit/delta": 1e-05,
3
+ "audit/embedding/auc": 0.520928,
4
+ "audit/embedding/empirical_epsilon/0.01": 0.0,
5
+ "audit/embedding/empirical_epsilon/0.05": 0.0,
6
+ "audit/embedding/empirical_epsilon_details/0.01/correct_guesses": 0.0,
7
+ "audit/embedding/empirical_epsilon_details/0.01/epsilon": 0.0,
8
+ "audit/embedding/empirical_epsilon_details/0.01/num_guesses": 0.0,
9
+ "audit/embedding/empirical_epsilon_details/0.05/correct_guesses": 0.0,
10
+ "audit/embedding/empirical_epsilon_details/0.05/epsilon": 0.0,
11
+ "audit/embedding/empirical_epsilon_details/0.05/num_guesses": 0.0,
12
+ "audit/loss/auc": 0.525808,
13
+ "audit/loss/empirical_epsilon/0.01": 0.0,
14
+ "audit/loss/empirical_epsilon/0.05": 0.05073561053723097,
15
+ "audit/loss/empirical_epsilon_details/0.01/correct_guesses": 0.0,
16
+ "audit/loss/empirical_epsilon_details/0.01/epsilon": 0.0,
17
+ "audit/loss/empirical_epsilon_details/0.01/num_guesses": 0.0,
18
+ "audit/loss/empirical_epsilon_details/0.05/correct_guesses": 60.0,
19
+ "audit/loss/empirical_epsilon_details/0.05/epsilon": 0.05073561053723097,
20
+ "audit/loss/empirical_epsilon_details/0.05/num_guesses": 100.0,
21
+ "audit/num_canaries": 500.0,
22
+ "audit/num_members": 250.0,
23
+ "audit/paper_guess_fraction": 0.2,
24
+ "audit/paper_guess_steps": 20.0,
25
+ "energy/codecarbon/cpu_count": 224.0,
26
+ "energy/codecarbon/cpu_energy": 0.18975838183590477,
27
+ "energy/codecarbon/cpu_power": 179.92904019353898,
28
+ "energy/codecarbon/cpu_utilization_percent": 2.4580389144905275,
29
+ "energy/codecarbon/duration": 3938.0607158355415,
30
+ "energy/codecarbon/emissions": 0.5702959606126596,
31
+ "energy/codecarbon/emissions_rate": 0.00014481644691749236,
32
+ "energy/codecarbon/energy_consumed": 2.778283903371513,
33
+ "energy/codecarbon/gpu_count": 4.0,
34
+ "energy/codecarbon/gpu_energy": 2.514705143984827,
35
+ "energy/codecarbon/gpu_power": 2298.550082608512,
36
+ "energy/codecarbon/gpu_utilization_percent": 95.87682458386683,
37
+ "energy/codecarbon/latitude": 37.3541,
38
+ "energy/codecarbon/longitude": -121.9552,
39
+ "energy/codecarbon/pue": 1.0,
40
+ "energy/codecarbon/ram_energy": 0.07382037755078118,
41
+ "energy/codecarbon/ram_power": 70.0,
42
+ "energy/codecarbon/ram_total_size": 2015.5625190734863,
43
+ "energy/codecarbon/ram_used_gb": 47.93522234837092,
44
+ "energy/codecarbon/ram_utilization_percent": 2.36410650281618,
45
+ "energy/codecarbon/water_consumed": 0.0,
46
+ "energy/codecarbon/wue": 0.0,
47
+ "eval/duration_sec": 27.067450830712914,
48
+ "eval/loss": 0.8425623961221212,
49
+ "perf/audit_duration_sec": 5.841585950925946,
50
+ "perf/epoch_duration_sec": 1886.0955602359027,
51
+ "perf/epoch_samples": 55619.0,
52
+ "perf/epoch_samples_per_sec": 29.488961838732855,
53
+ "perf/epoch_tokens": 43998761.0,
54
+ "perf/epoch_tokens_per_sec": 23327.959583605018,
55
+ "perf/logical_batch_size": 132.0,
56
+ "perf/logical_token_count": 106537.0,
57
+ "perf/physical_batches": 17.0,
58
+ "perf/samples_per_sec": 7.860206056319249,
59
+ "perf/step_duration_sec": 16.793452875688672,
60
+ "perf/tokens_per_sec": 6343.96039865215,
61
+ "privacy/epsilon": 2.9999680995370417,
62
+ "system/cuda_epoch_peak_memory_gb": 74.69534540176392,
63
+ "system/cuda_max_memory_allocated_gb": 74.69534540176392,
64
+ "system/cuda_memory_allocated_gb": 16.537474155426025,
65
+ "train/epoch_canary_loss": 9.74998009501868,
66
+ "train/epoch_loss": 1.6631105757675884,
67
+ "train/epoch_real_loss": 1.2492633356516678,
68
+ "train/lr": 8.126960406835249e-07,
69
+ "train/step_canary_loss": 10.9375,
70
+ "train/step_loss": 1.5310632604541201,
71
+ "train/step_real_loss": 1.2371121123433113
72
+ }
starcoder2-7b/dp3/tensorboard/events.out.tfevents.1774096221.364c4f8de9dd.12837.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b07c9c3ed310b74cc126d10e56bdacca0e2e7e9e9015a0a823e61040e77f436
3
+ size 26671
starcoder2-7b/dp3/tokenizer/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff