Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes. See raw diff
- starcoder2-7b/base/adapter/README.md +207 -0
- starcoder2-7b/base/adapter/adapter_config.json +46 -0
- starcoder2-7b/base/adapter/adapter_model.safetensors +3 -0
- starcoder2-7b/base/audit_results.json +137 -0
- starcoder2-7b/base/audit_scores.npz +3 -0
- starcoder2-7b/base/canary_meta.json +0 -0
- starcoder2-7b/base/codecarbon.csv +2 -0
- starcoder2-7b/base/epochs/epoch_001/adapter/README.md +207 -0
- starcoder2-7b/base/epochs/epoch_001/adapter/adapter_config.json +46 -0
- starcoder2-7b/base/epochs/epoch_001/adapter/adapter_model.safetensors +3 -0
- starcoder2-7b/base/epochs/epoch_001/audit_results.json +137 -0
- starcoder2-7b/base/epochs/epoch_001/audit_scores.npz +3 -0
- starcoder2-7b/base/epochs/epoch_002/adapter/README.md +207 -0
- starcoder2-7b/base/epochs/epoch_002/adapter/adapter_config.json +46 -0
- starcoder2-7b/base/epochs/epoch_002/adapter/adapter_model.safetensors +3 -0
- starcoder2-7b/base/epochs/epoch_002/audit_results.json +137 -0
- starcoder2-7b/base/epochs/epoch_002/audit_scores.npz +3 -0
- starcoder2-7b/base/metrics.jsonl +55 -0
- starcoder2-7b/base/pretrain_lm_head.pt +3 -0
- starcoder2-7b/base/resolved_config.yaml +100 -0
- starcoder2-7b/base/scalars.csv +613 -0
- starcoder2-7b/base/summary.json +71 -0
- starcoder2-7b/base/tensorboard/events.out.tfevents.1774090870.364c4f8de9dd.6145.0 +3 -0
- starcoder2-7b/base/tokenizer/tokenizer.json +0 -0
- starcoder2-7b/base/tokenizer/tokenizer_config.json +516 -0
- starcoder2-7b/base/train.log +49 -0
- starcoder2-7b/dp3/adapter/README.md +207 -0
- starcoder2-7b/dp3/adapter/adapter_config.json +46 -0
- starcoder2-7b/dp3/adapter/adapter_model.safetensors +3 -0
- starcoder2-7b/dp3/audit_results.json +137 -0
- starcoder2-7b/dp3/audit_scores.npz +3 -0
- starcoder2-7b/dp3/canary_meta.json +0 -0
- starcoder2-7b/dp3/codecarbon.csv +2 -0
- starcoder2-7b/dp3/epochs/epoch_001/adapter/README.md +207 -0
- starcoder2-7b/dp3/epochs/epoch_001/adapter/adapter_config.json +46 -0
- starcoder2-7b/dp3/epochs/epoch_001/adapter/adapter_model.safetensors +3 -0
- starcoder2-7b/dp3/epochs/epoch_001/audit_results.json +137 -0
- starcoder2-7b/dp3/epochs/epoch_001/audit_scores.npz +3 -0
- starcoder2-7b/dp3/epochs/epoch_002/adapter/README.md +207 -0
- starcoder2-7b/dp3/epochs/epoch_002/adapter/adapter_config.json +46 -0
- starcoder2-7b/dp3/epochs/epoch_002/adapter/adapter_model.safetensors +3 -0
- starcoder2-7b/dp3/epochs/epoch_002/audit_results.json +137 -0
- starcoder2-7b/dp3/epochs/epoch_002/audit_scores.npz +3 -0
- starcoder2-7b/dp3/metrics.jsonl +30 -0
- starcoder2-7b/dp3/pretrain_lm_head.pt +3 -0
- starcoder2-7b/dp3/resolved_config.yaml +101 -0
- starcoder2-7b/dp3/scalars.csv +386 -0
- starcoder2-7b/dp3/summary.json +72 -0
- starcoder2-7b/dp3/tensorboard/events.out.tfevents.1774096221.364c4f8de9dd.12837.0 +3 -0
- starcoder2-7b/dp3/tokenizer/tokenizer.json +0 -0
starcoder2-7b/base/adapter/README.md
ADDED
|
@@ -0,0 +1,207 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
base_model: bigcode/starcoder2-7b
|
| 3 |
+
library_name: peft
|
| 4 |
+
pipeline_tag: text-generation
|
| 5 |
+
tags:
|
| 6 |
+
- base_model:adapter:bigcode/starcoder2-7b
|
| 7 |
+
- lora
|
| 8 |
+
- transformers
|
| 9 |
+
---
|
| 10 |
+
|
| 11 |
+
# Model Card for Model ID
|
| 12 |
+
|
| 13 |
+
<!-- Provide a quick summary of what the model is/does. -->
|
| 14 |
+
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
## Model Details
|
| 18 |
+
|
| 19 |
+
### Model Description
|
| 20 |
+
|
| 21 |
+
<!-- Provide a longer summary of what this model is. -->
|
| 22 |
+
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
- **Developed by:** [More Information Needed]
|
| 26 |
+
- **Funded by [optional]:** [More Information Needed]
|
| 27 |
+
- **Shared by [optional]:** [More Information Needed]
|
| 28 |
+
- **Model type:** [More Information Needed]
|
| 29 |
+
- **Language(s) (NLP):** [More Information Needed]
|
| 30 |
+
- **License:** [More Information Needed]
|
| 31 |
+
- **Finetuned from model [optional]:** [More Information Needed]
|
| 32 |
+
|
| 33 |
+
### Model Sources [optional]
|
| 34 |
+
|
| 35 |
+
<!-- Provide the basic links for the model. -->
|
| 36 |
+
|
| 37 |
+
- **Repository:** [More Information Needed]
|
| 38 |
+
- **Paper [optional]:** [More Information Needed]
|
| 39 |
+
- **Demo [optional]:** [More Information Needed]
|
| 40 |
+
|
| 41 |
+
## Uses
|
| 42 |
+
|
| 43 |
+
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
| 44 |
+
|
| 45 |
+
### Direct Use
|
| 46 |
+
|
| 47 |
+
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
| 48 |
+
|
| 49 |
+
[More Information Needed]
|
| 50 |
+
|
| 51 |
+
### Downstream Use [optional]
|
| 52 |
+
|
| 53 |
+
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
|
| 54 |
+
|
| 55 |
+
[More Information Needed]
|
| 56 |
+
|
| 57 |
+
### Out-of-Scope Use
|
| 58 |
+
|
| 59 |
+
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
|
| 60 |
+
|
| 61 |
+
[More Information Needed]
|
| 62 |
+
|
| 63 |
+
## Bias, Risks, and Limitations
|
| 64 |
+
|
| 65 |
+
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
|
| 66 |
+
|
| 67 |
+
[More Information Needed]
|
| 68 |
+
|
| 69 |
+
### Recommendations
|
| 70 |
+
|
| 71 |
+
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
|
| 72 |
+
|
| 73 |
+
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
|
| 74 |
+
|
| 75 |
+
## How to Get Started with the Model
|
| 76 |
+
|
| 77 |
+
Use the code below to get started with the model.
|
| 78 |
+
|
| 79 |
+
[More Information Needed]
|
| 80 |
+
|
| 81 |
+
## Training Details
|
| 82 |
+
|
| 83 |
+
### Training Data
|
| 84 |
+
|
| 85 |
+
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
| 86 |
+
|
| 87 |
+
[More Information Needed]
|
| 88 |
+
|
| 89 |
+
### Training Procedure
|
| 90 |
+
|
| 91 |
+
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
|
| 92 |
+
|
| 93 |
+
#### Preprocessing [optional]
|
| 94 |
+
|
| 95 |
+
[More Information Needed]
|
| 96 |
+
|
| 97 |
+
|
| 98 |
+
#### Training Hyperparameters
|
| 99 |
+
|
| 100 |
+
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
| 101 |
+
|
| 102 |
+
#### Speeds, Sizes, Times [optional]
|
| 103 |
+
|
| 104 |
+
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|
| 105 |
+
|
| 106 |
+
[More Information Needed]
|
| 107 |
+
|
| 108 |
+
## Evaluation
|
| 109 |
+
|
| 110 |
+
<!-- This section describes the evaluation protocols and provides the results. -->
|
| 111 |
+
|
| 112 |
+
### Testing Data, Factors & Metrics
|
| 113 |
+
|
| 114 |
+
#### Testing Data
|
| 115 |
+
|
| 116 |
+
<!-- This should link to a Dataset Card if possible. -->
|
| 117 |
+
|
| 118 |
+
[More Information Needed]
|
| 119 |
+
|
| 120 |
+
#### Factors
|
| 121 |
+
|
| 122 |
+
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
|
| 123 |
+
|
| 124 |
+
[More Information Needed]
|
| 125 |
+
|
| 126 |
+
#### Metrics
|
| 127 |
+
|
| 128 |
+
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
|
| 129 |
+
|
| 130 |
+
[More Information Needed]
|
| 131 |
+
|
| 132 |
+
### Results
|
| 133 |
+
|
| 134 |
+
[More Information Needed]
|
| 135 |
+
|
| 136 |
+
#### Summary
|
| 137 |
+
|
| 138 |
+
|
| 139 |
+
|
| 140 |
+
## Model Examination [optional]
|
| 141 |
+
|
| 142 |
+
<!-- Relevant interpretability work for the model goes here -->
|
| 143 |
+
|
| 144 |
+
[More Information Needed]
|
| 145 |
+
|
| 146 |
+
## Environmental Impact
|
| 147 |
+
|
| 148 |
+
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
|
| 149 |
+
|
| 150 |
+
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
| 151 |
+
|
| 152 |
+
- **Hardware Type:** [More Information Needed]
|
| 153 |
+
- **Hours used:** [More Information Needed]
|
| 154 |
+
- **Cloud Provider:** [More Information Needed]
|
| 155 |
+
- **Compute Region:** [More Information Needed]
|
| 156 |
+
- **Carbon Emitted:** [More Information Needed]
|
| 157 |
+
|
| 158 |
+
## Technical Specifications [optional]
|
| 159 |
+
|
| 160 |
+
### Model Architecture and Objective
|
| 161 |
+
|
| 162 |
+
[More Information Needed]
|
| 163 |
+
|
| 164 |
+
### Compute Infrastructure
|
| 165 |
+
|
| 166 |
+
[More Information Needed]
|
| 167 |
+
|
| 168 |
+
#### Hardware
|
| 169 |
+
|
| 170 |
+
[More Information Needed]
|
| 171 |
+
|
| 172 |
+
#### Software
|
| 173 |
+
|
| 174 |
+
[More Information Needed]
|
| 175 |
+
|
| 176 |
+
## Citation [optional]
|
| 177 |
+
|
| 178 |
+
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
| 179 |
+
|
| 180 |
+
**BibTeX:**
|
| 181 |
+
|
| 182 |
+
[More Information Needed]
|
| 183 |
+
|
| 184 |
+
**APA:**
|
| 185 |
+
|
| 186 |
+
[More Information Needed]
|
| 187 |
+
|
| 188 |
+
## Glossary [optional]
|
| 189 |
+
|
| 190 |
+
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
|
| 191 |
+
|
| 192 |
+
[More Information Needed]
|
| 193 |
+
|
| 194 |
+
## More Information [optional]
|
| 195 |
+
|
| 196 |
+
[More Information Needed]
|
| 197 |
+
|
| 198 |
+
## Model Card Authors [optional]
|
| 199 |
+
|
| 200 |
+
[More Information Needed]
|
| 201 |
+
|
| 202 |
+
## Model Card Contact
|
| 203 |
+
|
| 204 |
+
[More Information Needed]
|
| 205 |
+
### Framework versions
|
| 206 |
+
|
| 207 |
+
- PEFT 0.18.1
|
starcoder2-7b/base/adapter/adapter_config.json
ADDED
|
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"alora_invocation_tokens": null,
|
| 3 |
+
"alpha_pattern": {},
|
| 4 |
+
"arrow_config": null,
|
| 5 |
+
"auto_mapping": null,
|
| 6 |
+
"base_model_name_or_path": "bigcode/starcoder2-7b",
|
| 7 |
+
"bias": "none",
|
| 8 |
+
"corda_config": null,
|
| 9 |
+
"ensure_weight_tying": true,
|
| 10 |
+
"eva_config": null,
|
| 11 |
+
"exclude_modules": null,
|
| 12 |
+
"fan_in_fan_out": false,
|
| 13 |
+
"inference_mode": true,
|
| 14 |
+
"init_lora_weights": true,
|
| 15 |
+
"layer_replication": null,
|
| 16 |
+
"layers_pattern": null,
|
| 17 |
+
"layers_to_transform": null,
|
| 18 |
+
"loftq_config": {},
|
| 19 |
+
"lora_alpha": 32,
|
| 20 |
+
"lora_bias": false,
|
| 21 |
+
"lora_dropout": 0.05,
|
| 22 |
+
"megatron_config": null,
|
| 23 |
+
"megatron_core": "megatron.core",
|
| 24 |
+
"modules_to_save": [
|
| 25 |
+
"lm_head",
|
| 26 |
+
"embed_tokens"
|
| 27 |
+
],
|
| 28 |
+
"peft_type": "LORA",
|
| 29 |
+
"peft_version": "0.18.1",
|
| 30 |
+
"qalora_group_size": 16,
|
| 31 |
+
"r": 16,
|
| 32 |
+
"rank_pattern": {},
|
| 33 |
+
"revision": null,
|
| 34 |
+
"target_modules": [
|
| 35 |
+
"v_proj",
|
| 36 |
+
"k_proj",
|
| 37 |
+
"q_proj",
|
| 38 |
+
"o_proj"
|
| 39 |
+
],
|
| 40 |
+
"target_parameters": null,
|
| 41 |
+
"task_type": "CAUSAL_LM",
|
| 42 |
+
"trainable_token_indices": null,
|
| 43 |
+
"use_dora": false,
|
| 44 |
+
"use_qalora": false,
|
| 45 |
+
"use_rslora": false
|
| 46 |
+
}
|
starcoder2-7b/base/adapter/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e8722c41c3297c10da9786673bd5f7f3556116c882e76e37035eedd5ae393938
|
| 3 |
+
size 2804312360
|
starcoder2-7b/base/audit_results.json
ADDED
|
@@ -0,0 +1,137 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"delta": 1e-05,
|
| 3 |
+
"num_canaries": 500,
|
| 4 |
+
"num_members": 250,
|
| 5 |
+
"paper_guess_fraction": 0.2,
|
| 6 |
+
"paper_guess_steps": 20,
|
| 7 |
+
"loss": {
|
| 8 |
+
"auc": 1.0,
|
| 9 |
+
"empirical_epsilon": {
|
| 10 |
+
"0.05": 3.4791953936219215,
|
| 11 |
+
"0.01": 3.023197554051876
|
| 12 |
+
},
|
| 13 |
+
"empirical_epsilon_details": {
|
| 14 |
+
"0.05": {
|
| 15 |
+
"epsilon": 3.4791953936219215,
|
| 16 |
+
"num_guesses": 100,
|
| 17 |
+
"correct_guesses": 100,
|
| 18 |
+
"candidate_num_guesses": [
|
| 19 |
+
5,
|
| 20 |
+
10,
|
| 21 |
+
15,
|
| 22 |
+
20,
|
| 23 |
+
25,
|
| 24 |
+
30,
|
| 25 |
+
35,
|
| 26 |
+
40,
|
| 27 |
+
45,
|
| 28 |
+
50,
|
| 29 |
+
55,
|
| 30 |
+
60,
|
| 31 |
+
65,
|
| 32 |
+
70,
|
| 33 |
+
75,
|
| 34 |
+
80,
|
| 35 |
+
85,
|
| 36 |
+
90,
|
| 37 |
+
95,
|
| 38 |
+
100
|
| 39 |
+
],
|
| 40 |
+
"direction": "lower"
|
| 41 |
+
},
|
| 42 |
+
"0.01": {
|
| 43 |
+
"epsilon": 3.023197554051876,
|
| 44 |
+
"num_guesses": 100,
|
| 45 |
+
"correct_guesses": 100,
|
| 46 |
+
"candidate_num_guesses": [
|
| 47 |
+
5,
|
| 48 |
+
10,
|
| 49 |
+
15,
|
| 50 |
+
20,
|
| 51 |
+
25,
|
| 52 |
+
30,
|
| 53 |
+
35,
|
| 54 |
+
40,
|
| 55 |
+
45,
|
| 56 |
+
50,
|
| 57 |
+
55,
|
| 58 |
+
60,
|
| 59 |
+
65,
|
| 60 |
+
70,
|
| 61 |
+
75,
|
| 62 |
+
80,
|
| 63 |
+
85,
|
| 64 |
+
90,
|
| 65 |
+
95,
|
| 66 |
+
100
|
| 67 |
+
],
|
| 68 |
+
"direction": "lower"
|
| 69 |
+
}
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"embedding": {
|
| 73 |
+
"auc": 0.916224,
|
| 74 |
+
"empirical_epsilon": {
|
| 75 |
+
"0.05": 3.4791953936219215,
|
| 76 |
+
"0.01": 3.023197554051876
|
| 77 |
+
},
|
| 78 |
+
"empirical_epsilon_details": {
|
| 79 |
+
"0.05": {
|
| 80 |
+
"epsilon": 3.4791953936219215,
|
| 81 |
+
"num_guesses": 100,
|
| 82 |
+
"correct_guesses": 100,
|
| 83 |
+
"candidate_num_guesses": [
|
| 84 |
+
5,
|
| 85 |
+
10,
|
| 86 |
+
15,
|
| 87 |
+
20,
|
| 88 |
+
25,
|
| 89 |
+
30,
|
| 90 |
+
35,
|
| 91 |
+
40,
|
| 92 |
+
45,
|
| 93 |
+
50,
|
| 94 |
+
55,
|
| 95 |
+
60,
|
| 96 |
+
65,
|
| 97 |
+
70,
|
| 98 |
+
75,
|
| 99 |
+
80,
|
| 100 |
+
85,
|
| 101 |
+
90,
|
| 102 |
+
95,
|
| 103 |
+
100
|
| 104 |
+
],
|
| 105 |
+
"direction": "lower"
|
| 106 |
+
},
|
| 107 |
+
"0.01": {
|
| 108 |
+
"epsilon": 3.023197554051876,
|
| 109 |
+
"num_guesses": 100,
|
| 110 |
+
"correct_guesses": 100,
|
| 111 |
+
"candidate_num_guesses": [
|
| 112 |
+
5,
|
| 113 |
+
10,
|
| 114 |
+
15,
|
| 115 |
+
20,
|
| 116 |
+
25,
|
| 117 |
+
30,
|
| 118 |
+
35,
|
| 119 |
+
40,
|
| 120 |
+
45,
|
| 121 |
+
50,
|
| 122 |
+
55,
|
| 123 |
+
60,
|
| 124 |
+
65,
|
| 125 |
+
70,
|
| 126 |
+
75,
|
| 127 |
+
80,
|
| 128 |
+
85,
|
| 129 |
+
90,
|
| 130 |
+
95,
|
| 131 |
+
100
|
| 132 |
+
],
|
| 133 |
+
"direction": "lower"
|
| 134 |
+
}
|
| 135 |
+
}
|
| 136 |
+
}
|
| 137 |
+
}
|
starcoder2-7b/base/audit_scores.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:df7a9ad49c78c2b7e30773bfe7d47ce312de900740e940d72ac0b3117dff09bf
|
| 3 |
+
size 12784
|
starcoder2-7b/base/canary_meta.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
starcoder2-7b/base/codecarbon.csv
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
timestamp,project_name,run_id,experiment_id,duration,emissions,emissions_rate,cpu_power,gpu_power,ram_power,cpu_energy,gpu_energy,ram_energy,energy_consumed,water_consumed,country_name,country_iso_code,region,cloud_provider,cloud_region,os,python_version,codecarbon_version,cpu_count,cpu_model,gpu_count,gpu_model,longitude,latitude,ram_total_size,tracking_mode,cpu_utilization_percent,gpu_utilization_percent,ram_utilization_percent,ram_used_gb,on_cloud,pue,wue
|
| 2 |
+
2026-03-21T12:05:04,codedp-starcoder2-7b-cpt-base,499ee9b6-ab5e-4714-be7f-0b592eb40909,5b0fa12a-3dd7-45bb-9766-cc326314d9f1,3820.65162669681,0.5568149811716148,0.0001457382236267942,179.3439937196561,2316.658494927477,70.0,0.18368298514637418,2.457235179675422,0.07169108754482326,2.7126092523666183,0.0,United States,USA,california,,,Linux-5.15.0-157-generic-x86_64-with-glibc2.39,3.12.13,3.2.5,224,Intel(R) Xeon(R) Platinum 8480C,4,4 x NVIDIA H200,-121.9552,37.3541,2015.5625190734863,machine,2.2497114375655825,96.45330535152151,2.3057974816369358,46.78091819539025,N,1.0,0.0
|
starcoder2-7b/base/epochs/epoch_001/adapter/README.md
ADDED
|
@@ -0,0 +1,207 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
base_model: bigcode/starcoder2-7b
|
| 3 |
+
library_name: peft
|
| 4 |
+
pipeline_tag: text-generation
|
| 5 |
+
tags:
|
| 6 |
+
- base_model:adapter:bigcode/starcoder2-7b
|
| 7 |
+
- lora
|
| 8 |
+
- transformers
|
| 9 |
+
---
|
| 10 |
+
|
| 11 |
+
# Model Card for Model ID
|
| 12 |
+
|
| 13 |
+
<!-- Provide a quick summary of what the model is/does. -->
|
| 14 |
+
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
## Model Details
|
| 18 |
+
|
| 19 |
+
### Model Description
|
| 20 |
+
|
| 21 |
+
<!-- Provide a longer summary of what this model is. -->
|
| 22 |
+
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
- **Developed by:** [More Information Needed]
|
| 26 |
+
- **Funded by [optional]:** [More Information Needed]
|
| 27 |
+
- **Shared by [optional]:** [More Information Needed]
|
| 28 |
+
- **Model type:** [More Information Needed]
|
| 29 |
+
- **Language(s) (NLP):** [More Information Needed]
|
| 30 |
+
- **License:** [More Information Needed]
|
| 31 |
+
- **Finetuned from model [optional]:** [More Information Needed]
|
| 32 |
+
|
| 33 |
+
### Model Sources [optional]
|
| 34 |
+
|
| 35 |
+
<!-- Provide the basic links for the model. -->
|
| 36 |
+
|
| 37 |
+
- **Repository:** [More Information Needed]
|
| 38 |
+
- **Paper [optional]:** [More Information Needed]
|
| 39 |
+
- **Demo [optional]:** [More Information Needed]
|
| 40 |
+
|
| 41 |
+
## Uses
|
| 42 |
+
|
| 43 |
+
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
| 44 |
+
|
| 45 |
+
### Direct Use
|
| 46 |
+
|
| 47 |
+
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
| 48 |
+
|
| 49 |
+
[More Information Needed]
|
| 50 |
+
|
| 51 |
+
### Downstream Use [optional]
|
| 52 |
+
|
| 53 |
+
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
|
| 54 |
+
|
| 55 |
+
[More Information Needed]
|
| 56 |
+
|
| 57 |
+
### Out-of-Scope Use
|
| 58 |
+
|
| 59 |
+
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
|
| 60 |
+
|
| 61 |
+
[More Information Needed]
|
| 62 |
+
|
| 63 |
+
## Bias, Risks, and Limitations
|
| 64 |
+
|
| 65 |
+
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
|
| 66 |
+
|
| 67 |
+
[More Information Needed]
|
| 68 |
+
|
| 69 |
+
### Recommendations
|
| 70 |
+
|
| 71 |
+
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
|
| 72 |
+
|
| 73 |
+
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
|
| 74 |
+
|
| 75 |
+
## How to Get Started with the Model
|
| 76 |
+
|
| 77 |
+
Use the code below to get started with the model.
|
| 78 |
+
|
| 79 |
+
[More Information Needed]
|
| 80 |
+
|
| 81 |
+
## Training Details
|
| 82 |
+
|
| 83 |
+
### Training Data
|
| 84 |
+
|
| 85 |
+
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
| 86 |
+
|
| 87 |
+
[More Information Needed]
|
| 88 |
+
|
| 89 |
+
### Training Procedure
|
| 90 |
+
|
| 91 |
+
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
|
| 92 |
+
|
| 93 |
+
#### Preprocessing [optional]
|
| 94 |
+
|
| 95 |
+
[More Information Needed]
|
| 96 |
+
|
| 97 |
+
|
| 98 |
+
#### Training Hyperparameters
|
| 99 |
+
|
| 100 |
+
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
| 101 |
+
|
| 102 |
+
#### Speeds, Sizes, Times [optional]
|
| 103 |
+
|
| 104 |
+
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|
| 105 |
+
|
| 106 |
+
[More Information Needed]
|
| 107 |
+
|
| 108 |
+
## Evaluation
|
| 109 |
+
|
| 110 |
+
<!-- This section describes the evaluation protocols and provides the results. -->
|
| 111 |
+
|
| 112 |
+
### Testing Data, Factors & Metrics
|
| 113 |
+
|
| 114 |
+
#### Testing Data
|
| 115 |
+
|
| 116 |
+
<!-- This should link to a Dataset Card if possible. -->
|
| 117 |
+
|
| 118 |
+
[More Information Needed]
|
| 119 |
+
|
| 120 |
+
#### Factors
|
| 121 |
+
|
| 122 |
+
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
|
| 123 |
+
|
| 124 |
+
[More Information Needed]
|
| 125 |
+
|
| 126 |
+
#### Metrics
|
| 127 |
+
|
| 128 |
+
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
|
| 129 |
+
|
| 130 |
+
[More Information Needed]
|
| 131 |
+
|
| 132 |
+
### Results
|
| 133 |
+
|
| 134 |
+
[More Information Needed]
|
| 135 |
+
|
| 136 |
+
#### Summary
|
| 137 |
+
|
| 138 |
+
|
| 139 |
+
|
| 140 |
+
## Model Examination [optional]
|
| 141 |
+
|
| 142 |
+
<!-- Relevant interpretability work for the model goes here -->
|
| 143 |
+
|
| 144 |
+
[More Information Needed]
|
| 145 |
+
|
| 146 |
+
## Environmental Impact
|
| 147 |
+
|
| 148 |
+
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
|
| 149 |
+
|
| 150 |
+
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
| 151 |
+
|
| 152 |
+
- **Hardware Type:** [More Information Needed]
|
| 153 |
+
- **Hours used:** [More Information Needed]
|
| 154 |
+
- **Cloud Provider:** [More Information Needed]
|
| 155 |
+
- **Compute Region:** [More Information Needed]
|
| 156 |
+
- **Carbon Emitted:** [More Information Needed]
|
| 157 |
+
|
| 158 |
+
## Technical Specifications [optional]
|
| 159 |
+
|
| 160 |
+
### Model Architecture and Objective
|
| 161 |
+
|
| 162 |
+
[More Information Needed]
|
| 163 |
+
|
| 164 |
+
### Compute Infrastructure
|
| 165 |
+
|
| 166 |
+
[More Information Needed]
|
| 167 |
+
|
| 168 |
+
#### Hardware
|
| 169 |
+
|
| 170 |
+
[More Information Needed]
|
| 171 |
+
|
| 172 |
+
#### Software
|
| 173 |
+
|
| 174 |
+
[More Information Needed]
|
| 175 |
+
|
| 176 |
+
## Citation [optional]
|
| 177 |
+
|
| 178 |
+
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
| 179 |
+
|
| 180 |
+
**BibTeX:**
|
| 181 |
+
|
| 182 |
+
[More Information Needed]
|
| 183 |
+
|
| 184 |
+
**APA:**
|
| 185 |
+
|
| 186 |
+
[More Information Needed]
|
| 187 |
+
|
| 188 |
+
## Glossary [optional]
|
| 189 |
+
|
| 190 |
+
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
|
| 191 |
+
|
| 192 |
+
[More Information Needed]
|
| 193 |
+
|
| 194 |
+
## More Information [optional]
|
| 195 |
+
|
| 196 |
+
[More Information Needed]
|
| 197 |
+
|
| 198 |
+
## Model Card Authors [optional]
|
| 199 |
+
|
| 200 |
+
[More Information Needed]
|
| 201 |
+
|
| 202 |
+
## Model Card Contact
|
| 203 |
+
|
| 204 |
+
[More Information Needed]
|
| 205 |
+
### Framework versions
|
| 206 |
+
|
| 207 |
+
- PEFT 0.18.1
|
starcoder2-7b/base/epochs/epoch_001/adapter/adapter_config.json
ADDED
|
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"alora_invocation_tokens": null,
|
| 3 |
+
"alpha_pattern": {},
|
| 4 |
+
"arrow_config": null,
|
| 5 |
+
"auto_mapping": null,
|
| 6 |
+
"base_model_name_or_path": "bigcode/starcoder2-7b",
|
| 7 |
+
"bias": "none",
|
| 8 |
+
"corda_config": null,
|
| 9 |
+
"ensure_weight_tying": true,
|
| 10 |
+
"eva_config": null,
|
| 11 |
+
"exclude_modules": null,
|
| 12 |
+
"fan_in_fan_out": false,
|
| 13 |
+
"inference_mode": true,
|
| 14 |
+
"init_lora_weights": true,
|
| 15 |
+
"layer_replication": null,
|
| 16 |
+
"layers_pattern": null,
|
| 17 |
+
"layers_to_transform": null,
|
| 18 |
+
"loftq_config": {},
|
| 19 |
+
"lora_alpha": 32,
|
| 20 |
+
"lora_bias": false,
|
| 21 |
+
"lora_dropout": 0.05,
|
| 22 |
+
"megatron_config": null,
|
| 23 |
+
"megatron_core": "megatron.core",
|
| 24 |
+
"modules_to_save": [
|
| 25 |
+
"lm_head",
|
| 26 |
+
"embed_tokens"
|
| 27 |
+
],
|
| 28 |
+
"peft_type": "LORA",
|
| 29 |
+
"peft_version": "0.18.1",
|
| 30 |
+
"qalora_group_size": 16,
|
| 31 |
+
"r": 16,
|
| 32 |
+
"rank_pattern": {},
|
| 33 |
+
"revision": null,
|
| 34 |
+
"target_modules": [
|
| 35 |
+
"v_proj",
|
| 36 |
+
"k_proj",
|
| 37 |
+
"q_proj",
|
| 38 |
+
"o_proj"
|
| 39 |
+
],
|
| 40 |
+
"target_parameters": null,
|
| 41 |
+
"task_type": "CAUSAL_LM",
|
| 42 |
+
"trainable_token_indices": null,
|
| 43 |
+
"use_dora": false,
|
| 44 |
+
"use_qalora": false,
|
| 45 |
+
"use_rslora": false
|
| 46 |
+
}
|
starcoder2-7b/base/epochs/epoch_001/adapter/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8ab7c236acdfacb363251d6f1c092cab8b23cfee5381584cd586426a1ecf16d0
|
| 3 |
+
size 2804312360
|
starcoder2-7b/base/epochs/epoch_001/audit_results.json
ADDED
|
@@ -0,0 +1,137 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"delta": 1e-05,
|
| 3 |
+
"num_canaries": 500,
|
| 4 |
+
"num_members": 250,
|
| 5 |
+
"paper_guess_fraction": 0.2,
|
| 6 |
+
"paper_guess_steps": 20,
|
| 7 |
+
"loss": {
|
| 8 |
+
"auc": 0.998072,
|
| 9 |
+
"empirical_epsilon": {
|
| 10 |
+
"0.05": 3.4791953936219215,
|
| 11 |
+
"0.01": 3.023197554051876
|
| 12 |
+
},
|
| 13 |
+
"empirical_epsilon_details": {
|
| 14 |
+
"0.05": {
|
| 15 |
+
"epsilon": 3.4791953936219215,
|
| 16 |
+
"num_guesses": 100,
|
| 17 |
+
"correct_guesses": 100,
|
| 18 |
+
"candidate_num_guesses": [
|
| 19 |
+
5,
|
| 20 |
+
10,
|
| 21 |
+
15,
|
| 22 |
+
20,
|
| 23 |
+
25,
|
| 24 |
+
30,
|
| 25 |
+
35,
|
| 26 |
+
40,
|
| 27 |
+
45,
|
| 28 |
+
50,
|
| 29 |
+
55,
|
| 30 |
+
60,
|
| 31 |
+
65,
|
| 32 |
+
70,
|
| 33 |
+
75,
|
| 34 |
+
80,
|
| 35 |
+
85,
|
| 36 |
+
90,
|
| 37 |
+
95,
|
| 38 |
+
100
|
| 39 |
+
],
|
| 40 |
+
"direction": "lower"
|
| 41 |
+
},
|
| 42 |
+
"0.01": {
|
| 43 |
+
"epsilon": 3.023197554051876,
|
| 44 |
+
"num_guesses": 100,
|
| 45 |
+
"correct_guesses": 100,
|
| 46 |
+
"candidate_num_guesses": [
|
| 47 |
+
5,
|
| 48 |
+
10,
|
| 49 |
+
15,
|
| 50 |
+
20,
|
| 51 |
+
25,
|
| 52 |
+
30,
|
| 53 |
+
35,
|
| 54 |
+
40,
|
| 55 |
+
45,
|
| 56 |
+
50,
|
| 57 |
+
55,
|
| 58 |
+
60,
|
| 59 |
+
65,
|
| 60 |
+
70,
|
| 61 |
+
75,
|
| 62 |
+
80,
|
| 63 |
+
85,
|
| 64 |
+
90,
|
| 65 |
+
95,
|
| 66 |
+
100
|
| 67 |
+
],
|
| 68 |
+
"direction": "lower"
|
| 69 |
+
}
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"embedding": {
|
| 73 |
+
"auc": 0.96,
|
| 74 |
+
"empirical_epsilon": {
|
| 75 |
+
"0.05": 3.4791953936219215,
|
| 76 |
+
"0.01": 3.023197554051876
|
| 77 |
+
},
|
| 78 |
+
"empirical_epsilon_details": {
|
| 79 |
+
"0.05": {
|
| 80 |
+
"epsilon": 3.4791953936219215,
|
| 81 |
+
"num_guesses": 100,
|
| 82 |
+
"correct_guesses": 100,
|
| 83 |
+
"candidate_num_guesses": [
|
| 84 |
+
5,
|
| 85 |
+
10,
|
| 86 |
+
15,
|
| 87 |
+
20,
|
| 88 |
+
25,
|
| 89 |
+
30,
|
| 90 |
+
35,
|
| 91 |
+
40,
|
| 92 |
+
45,
|
| 93 |
+
50,
|
| 94 |
+
55,
|
| 95 |
+
60,
|
| 96 |
+
65,
|
| 97 |
+
70,
|
| 98 |
+
75,
|
| 99 |
+
80,
|
| 100 |
+
85,
|
| 101 |
+
90,
|
| 102 |
+
95,
|
| 103 |
+
100
|
| 104 |
+
],
|
| 105 |
+
"direction": "lower"
|
| 106 |
+
},
|
| 107 |
+
"0.01": {
|
| 108 |
+
"epsilon": 3.023197554051876,
|
| 109 |
+
"num_guesses": 100,
|
| 110 |
+
"correct_guesses": 100,
|
| 111 |
+
"candidate_num_guesses": [
|
| 112 |
+
5,
|
| 113 |
+
10,
|
| 114 |
+
15,
|
| 115 |
+
20,
|
| 116 |
+
25,
|
| 117 |
+
30,
|
| 118 |
+
35,
|
| 119 |
+
40,
|
| 120 |
+
45,
|
| 121 |
+
50,
|
| 122 |
+
55,
|
| 123 |
+
60,
|
| 124 |
+
65,
|
| 125 |
+
70,
|
| 126 |
+
75,
|
| 127 |
+
80,
|
| 128 |
+
85,
|
| 129 |
+
90,
|
| 130 |
+
95,
|
| 131 |
+
100
|
| 132 |
+
],
|
| 133 |
+
"direction": "lower"
|
| 134 |
+
}
|
| 135 |
+
}
|
| 136 |
+
}
|
| 137 |
+
}
|
starcoder2-7b/base/epochs/epoch_001/audit_scores.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ad12d1a5732cbbe84adb0a8ad0d9063e66354e84483f65f8d0ad9cbad178b42f
|
| 3 |
+
size 12784
|
starcoder2-7b/base/epochs/epoch_002/adapter/README.md
ADDED
|
@@ -0,0 +1,207 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
base_model: bigcode/starcoder2-7b
|
| 3 |
+
library_name: peft
|
| 4 |
+
pipeline_tag: text-generation
|
| 5 |
+
tags:
|
| 6 |
+
- base_model:adapter:bigcode/starcoder2-7b
|
| 7 |
+
- lora
|
| 8 |
+
- transformers
|
| 9 |
+
---
|
| 10 |
+
|
| 11 |
+
# Model Card for Model ID
|
| 12 |
+
|
| 13 |
+
<!-- Provide a quick summary of what the model is/does. -->
|
| 14 |
+
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
## Model Details
|
| 18 |
+
|
| 19 |
+
### Model Description
|
| 20 |
+
|
| 21 |
+
<!-- Provide a longer summary of what this model is. -->
|
| 22 |
+
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
- **Developed by:** [More Information Needed]
|
| 26 |
+
- **Funded by [optional]:** [More Information Needed]
|
| 27 |
+
- **Shared by [optional]:** [More Information Needed]
|
| 28 |
+
- **Model type:** [More Information Needed]
|
| 29 |
+
- **Language(s) (NLP):** [More Information Needed]
|
| 30 |
+
- **License:** [More Information Needed]
|
| 31 |
+
- **Finetuned from model [optional]:** [More Information Needed]
|
| 32 |
+
|
| 33 |
+
### Model Sources [optional]
|
| 34 |
+
|
| 35 |
+
<!-- Provide the basic links for the model. -->
|
| 36 |
+
|
| 37 |
+
- **Repository:** [More Information Needed]
|
| 38 |
+
- **Paper [optional]:** [More Information Needed]
|
| 39 |
+
- **Demo [optional]:** [More Information Needed]
|
| 40 |
+
|
| 41 |
+
## Uses
|
| 42 |
+
|
| 43 |
+
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
| 44 |
+
|
| 45 |
+
### Direct Use
|
| 46 |
+
|
| 47 |
+
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
| 48 |
+
|
| 49 |
+
[More Information Needed]
|
| 50 |
+
|
| 51 |
+
### Downstream Use [optional]
|
| 52 |
+
|
| 53 |
+
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
|
| 54 |
+
|
| 55 |
+
[More Information Needed]
|
| 56 |
+
|
| 57 |
+
### Out-of-Scope Use
|
| 58 |
+
|
| 59 |
+
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
|
| 60 |
+
|
| 61 |
+
[More Information Needed]
|
| 62 |
+
|
| 63 |
+
## Bias, Risks, and Limitations
|
| 64 |
+
|
| 65 |
+
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
|
| 66 |
+
|
| 67 |
+
[More Information Needed]
|
| 68 |
+
|
| 69 |
+
### Recommendations
|
| 70 |
+
|
| 71 |
+
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
|
| 72 |
+
|
| 73 |
+
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
|
| 74 |
+
|
| 75 |
+
## How to Get Started with the Model
|
| 76 |
+
|
| 77 |
+
Use the code below to get started with the model.
|
| 78 |
+
|
| 79 |
+
[More Information Needed]
|
| 80 |
+
|
| 81 |
+
## Training Details
|
| 82 |
+
|
| 83 |
+
### Training Data
|
| 84 |
+
|
| 85 |
+
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
| 86 |
+
|
| 87 |
+
[More Information Needed]
|
| 88 |
+
|
| 89 |
+
### Training Procedure
|
| 90 |
+
|
| 91 |
+
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
|
| 92 |
+
|
| 93 |
+
#### Preprocessing [optional]
|
| 94 |
+
|
| 95 |
+
[More Information Needed]
|
| 96 |
+
|
| 97 |
+
|
| 98 |
+
#### Training Hyperparameters
|
| 99 |
+
|
| 100 |
+
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
| 101 |
+
|
| 102 |
+
#### Speeds, Sizes, Times [optional]
|
| 103 |
+
|
| 104 |
+
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|
| 105 |
+
|
| 106 |
+
[More Information Needed]
|
| 107 |
+
|
| 108 |
+
## Evaluation
|
| 109 |
+
|
| 110 |
+
<!-- This section describes the evaluation protocols and provides the results. -->
|
| 111 |
+
|
| 112 |
+
### Testing Data, Factors & Metrics
|
| 113 |
+
|
| 114 |
+
#### Testing Data
|
| 115 |
+
|
| 116 |
+
<!-- This should link to a Dataset Card if possible. -->
|
| 117 |
+
|
| 118 |
+
[More Information Needed]
|
| 119 |
+
|
| 120 |
+
#### Factors
|
| 121 |
+
|
| 122 |
+
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
|
| 123 |
+
|
| 124 |
+
[More Information Needed]
|
| 125 |
+
|
| 126 |
+
#### Metrics
|
| 127 |
+
|
| 128 |
+
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
|
| 129 |
+
|
| 130 |
+
[More Information Needed]
|
| 131 |
+
|
| 132 |
+
### Results
|
| 133 |
+
|
| 134 |
+
[More Information Needed]
|
| 135 |
+
|
| 136 |
+
#### Summary
|
| 137 |
+
|
| 138 |
+
|
| 139 |
+
|
| 140 |
+
## Model Examination [optional]
|
| 141 |
+
|
| 142 |
+
<!-- Relevant interpretability work for the model goes here -->
|
| 143 |
+
|
| 144 |
+
[More Information Needed]
|
| 145 |
+
|
| 146 |
+
## Environmental Impact
|
| 147 |
+
|
| 148 |
+
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
|
| 149 |
+
|
| 150 |
+
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
| 151 |
+
|
| 152 |
+
- **Hardware Type:** [More Information Needed]
|
| 153 |
+
- **Hours used:** [More Information Needed]
|
| 154 |
+
- **Cloud Provider:** [More Information Needed]
|
| 155 |
+
- **Compute Region:** [More Information Needed]
|
| 156 |
+
- **Carbon Emitted:** [More Information Needed]
|
| 157 |
+
|
| 158 |
+
## Technical Specifications [optional]
|
| 159 |
+
|
| 160 |
+
### Model Architecture and Objective
|
| 161 |
+
|
| 162 |
+
[More Information Needed]
|
| 163 |
+
|
| 164 |
+
### Compute Infrastructure
|
| 165 |
+
|
| 166 |
+
[More Information Needed]
|
| 167 |
+
|
| 168 |
+
#### Hardware
|
| 169 |
+
|
| 170 |
+
[More Information Needed]
|
| 171 |
+
|
| 172 |
+
#### Software
|
| 173 |
+
|
| 174 |
+
[More Information Needed]
|
| 175 |
+
|
| 176 |
+
## Citation [optional]
|
| 177 |
+
|
| 178 |
+
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
| 179 |
+
|
| 180 |
+
**BibTeX:**
|
| 181 |
+
|
| 182 |
+
[More Information Needed]
|
| 183 |
+
|
| 184 |
+
**APA:**
|
| 185 |
+
|
| 186 |
+
[More Information Needed]
|
| 187 |
+
|
| 188 |
+
## Glossary [optional]
|
| 189 |
+
|
| 190 |
+
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
|
| 191 |
+
|
| 192 |
+
[More Information Needed]
|
| 193 |
+
|
| 194 |
+
## More Information [optional]
|
| 195 |
+
|
| 196 |
+
[More Information Needed]
|
| 197 |
+
|
| 198 |
+
## Model Card Authors [optional]
|
| 199 |
+
|
| 200 |
+
[More Information Needed]
|
| 201 |
+
|
| 202 |
+
## Model Card Contact
|
| 203 |
+
|
| 204 |
+
[More Information Needed]
|
| 205 |
+
### Framework versions
|
| 206 |
+
|
| 207 |
+
- PEFT 0.18.1
|
starcoder2-7b/base/epochs/epoch_002/adapter/adapter_config.json
ADDED
|
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"alora_invocation_tokens": null,
|
| 3 |
+
"alpha_pattern": {},
|
| 4 |
+
"arrow_config": null,
|
| 5 |
+
"auto_mapping": null,
|
| 6 |
+
"base_model_name_or_path": "bigcode/starcoder2-7b",
|
| 7 |
+
"bias": "none",
|
| 8 |
+
"corda_config": null,
|
| 9 |
+
"ensure_weight_tying": true,
|
| 10 |
+
"eva_config": null,
|
| 11 |
+
"exclude_modules": null,
|
| 12 |
+
"fan_in_fan_out": false,
|
| 13 |
+
"inference_mode": true,
|
| 14 |
+
"init_lora_weights": true,
|
| 15 |
+
"layer_replication": null,
|
| 16 |
+
"layers_pattern": null,
|
| 17 |
+
"layers_to_transform": null,
|
| 18 |
+
"loftq_config": {},
|
| 19 |
+
"lora_alpha": 32,
|
| 20 |
+
"lora_bias": false,
|
| 21 |
+
"lora_dropout": 0.05,
|
| 22 |
+
"megatron_config": null,
|
| 23 |
+
"megatron_core": "megatron.core",
|
| 24 |
+
"modules_to_save": [
|
| 25 |
+
"lm_head",
|
| 26 |
+
"embed_tokens"
|
| 27 |
+
],
|
| 28 |
+
"peft_type": "LORA",
|
| 29 |
+
"peft_version": "0.18.1",
|
| 30 |
+
"qalora_group_size": 16,
|
| 31 |
+
"r": 16,
|
| 32 |
+
"rank_pattern": {},
|
| 33 |
+
"revision": null,
|
| 34 |
+
"target_modules": [
|
| 35 |
+
"v_proj",
|
| 36 |
+
"k_proj",
|
| 37 |
+
"q_proj",
|
| 38 |
+
"o_proj"
|
| 39 |
+
],
|
| 40 |
+
"target_parameters": null,
|
| 41 |
+
"task_type": "CAUSAL_LM",
|
| 42 |
+
"trainable_token_indices": null,
|
| 43 |
+
"use_dora": false,
|
| 44 |
+
"use_qalora": false,
|
| 45 |
+
"use_rslora": false
|
| 46 |
+
}
|
starcoder2-7b/base/epochs/epoch_002/adapter/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e8722c41c3297c10da9786673bd5f7f3556116c882e76e37035eedd5ae393938
|
| 3 |
+
size 2804312360
|
starcoder2-7b/base/epochs/epoch_002/audit_results.json
ADDED
|
@@ -0,0 +1,137 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"delta": 1e-05,
|
| 3 |
+
"num_canaries": 500,
|
| 4 |
+
"num_members": 250,
|
| 5 |
+
"paper_guess_fraction": 0.2,
|
| 6 |
+
"paper_guess_steps": 20,
|
| 7 |
+
"loss": {
|
| 8 |
+
"auc": 1.0,
|
| 9 |
+
"empirical_epsilon": {
|
| 10 |
+
"0.05": 3.4791953936219215,
|
| 11 |
+
"0.01": 3.023197554051876
|
| 12 |
+
},
|
| 13 |
+
"empirical_epsilon_details": {
|
| 14 |
+
"0.05": {
|
| 15 |
+
"epsilon": 3.4791953936219215,
|
| 16 |
+
"num_guesses": 100,
|
| 17 |
+
"correct_guesses": 100,
|
| 18 |
+
"candidate_num_guesses": [
|
| 19 |
+
5,
|
| 20 |
+
10,
|
| 21 |
+
15,
|
| 22 |
+
20,
|
| 23 |
+
25,
|
| 24 |
+
30,
|
| 25 |
+
35,
|
| 26 |
+
40,
|
| 27 |
+
45,
|
| 28 |
+
50,
|
| 29 |
+
55,
|
| 30 |
+
60,
|
| 31 |
+
65,
|
| 32 |
+
70,
|
| 33 |
+
75,
|
| 34 |
+
80,
|
| 35 |
+
85,
|
| 36 |
+
90,
|
| 37 |
+
95,
|
| 38 |
+
100
|
| 39 |
+
],
|
| 40 |
+
"direction": "lower"
|
| 41 |
+
},
|
| 42 |
+
"0.01": {
|
| 43 |
+
"epsilon": 3.023197554051876,
|
| 44 |
+
"num_guesses": 100,
|
| 45 |
+
"correct_guesses": 100,
|
| 46 |
+
"candidate_num_guesses": [
|
| 47 |
+
5,
|
| 48 |
+
10,
|
| 49 |
+
15,
|
| 50 |
+
20,
|
| 51 |
+
25,
|
| 52 |
+
30,
|
| 53 |
+
35,
|
| 54 |
+
40,
|
| 55 |
+
45,
|
| 56 |
+
50,
|
| 57 |
+
55,
|
| 58 |
+
60,
|
| 59 |
+
65,
|
| 60 |
+
70,
|
| 61 |
+
75,
|
| 62 |
+
80,
|
| 63 |
+
85,
|
| 64 |
+
90,
|
| 65 |
+
95,
|
| 66 |
+
100
|
| 67 |
+
],
|
| 68 |
+
"direction": "lower"
|
| 69 |
+
}
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"embedding": {
|
| 73 |
+
"auc": 0.916224,
|
| 74 |
+
"empirical_epsilon": {
|
| 75 |
+
"0.05": 3.4791953936219215,
|
| 76 |
+
"0.01": 3.023197554051876
|
| 77 |
+
},
|
| 78 |
+
"empirical_epsilon_details": {
|
| 79 |
+
"0.05": {
|
| 80 |
+
"epsilon": 3.4791953936219215,
|
| 81 |
+
"num_guesses": 100,
|
| 82 |
+
"correct_guesses": 100,
|
| 83 |
+
"candidate_num_guesses": [
|
| 84 |
+
5,
|
| 85 |
+
10,
|
| 86 |
+
15,
|
| 87 |
+
20,
|
| 88 |
+
25,
|
| 89 |
+
30,
|
| 90 |
+
35,
|
| 91 |
+
40,
|
| 92 |
+
45,
|
| 93 |
+
50,
|
| 94 |
+
55,
|
| 95 |
+
60,
|
| 96 |
+
65,
|
| 97 |
+
70,
|
| 98 |
+
75,
|
| 99 |
+
80,
|
| 100 |
+
85,
|
| 101 |
+
90,
|
| 102 |
+
95,
|
| 103 |
+
100
|
| 104 |
+
],
|
| 105 |
+
"direction": "lower"
|
| 106 |
+
},
|
| 107 |
+
"0.01": {
|
| 108 |
+
"epsilon": 3.023197554051876,
|
| 109 |
+
"num_guesses": 100,
|
| 110 |
+
"correct_guesses": 100,
|
| 111 |
+
"candidate_num_guesses": [
|
| 112 |
+
5,
|
| 113 |
+
10,
|
| 114 |
+
15,
|
| 115 |
+
20,
|
| 116 |
+
25,
|
| 117 |
+
30,
|
| 118 |
+
35,
|
| 119 |
+
40,
|
| 120 |
+
45,
|
| 121 |
+
50,
|
| 122 |
+
55,
|
| 123 |
+
60,
|
| 124 |
+
65,
|
| 125 |
+
70,
|
| 126 |
+
75,
|
| 127 |
+
80,
|
| 128 |
+
85,
|
| 129 |
+
90,
|
| 130 |
+
95,
|
| 131 |
+
100
|
| 132 |
+
],
|
| 133 |
+
"direction": "lower"
|
| 134 |
+
}
|
| 135 |
+
}
|
| 136 |
+
}
|
| 137 |
+
}
|
starcoder2-7b/base/epochs/epoch_002/audit_scores.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:df7a9ad49c78c2b7e30773bfe7d47ce312de900740e940d72ac0b3117dff09bf
|
| 3 |
+
size 12784
|
starcoder2-7b/base/metrics.jsonl
ADDED
|
@@ -0,0 +1,55 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"timestamp": 1774091071.9870658, "event": "train_step", "step": 10, "epoch": 1, "metrics": {"train/step_loss": 1.5507386380975896, "train/step_real_loss": 1.2866992354393005, "train/lr": 4.761904761904762e-05, "train/step_canary_loss": 10.0, "perf/step_duration_sec": 8.244300998747349, "perf/samples_per_sec": 8.00553012438873, "perf/tokens_per_sec": 6440.812872803661, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 53100.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.16860580444336}}
|
| 2 |
+
{"timestamp": 1774091154.380765, "event": "train_step", "step": 20, "epoch": 1, "metrics": {"train/step_loss": 1.4504926845208923, "train/step_real_loss": 1.2006134390830994, "train/lr": 9.523809523809524e-05, "train/step_canary_loss": 6.78125, "perf/step_duration_sec": 8.244972918182611, "perf/samples_per_sec": 8.126163744242886, "perf/tokens_per_sec": 6422.095078472549, "perf/logical_batch_size": 67.0, "perf/logical_token_count": 52950.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.911073684692383, "system/cuda_max_memory_allocated_gb": 82.16860580444336}}
|
| 3 |
+
{"timestamp": 1774091235.6533117, "event": "train_step", "step": 30, "epoch": 1, "metrics": {"train/step_loss": 0.9861254319548607, "train/step_real_loss": 0.9861254319548607, "train/lr": 9.98706541985615e-05, "perf/step_duration_sec": 7.973596462979913, "perf/samples_per_sec": 8.026490968929942, "perf/tokens_per_sec": 7148.9697609673985, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 57003.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.16860580444336}}
|
| 4 |
+
{"timestamp": 1774091318.662035, "event": "train_step", "step": 40, "epoch": 1, "metrics": {"train/step_loss": 1.0557814985513687, "train/step_real_loss": 1.0557814985513687, "train/lr": 9.942439201095397e-05, "perf/step_duration_sec": 8.119743634015322, "perf/samples_per_sec": 7.882022251526572, "perf/tokens_per_sec": 6340.347961767047, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 51482.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
|
| 5 |
+
{"timestamp": 1774091402.0365882, "event": "train_step", "step": 50, "epoch": 1, "metrics": {"train/step_loss": 1.017115435233483, "train/step_real_loss": 0.9500000476837158, "train/lr": 9.866246608261724e-05, "train/step_canary_loss": 5.3125, "perf/step_duration_sec": 8.119036318734288, "perf/samples_per_sec": 8.005876245437603, "perf/tokens_per_sec": 6307.645142790008, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 51212.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
|
| 6 |
+
{"timestamp": 1774091429.1958544, "event": "eval_step", "step": 50, "epoch": 1, "metrics": {"eval/loss": 0.8033588177513802, "eval/duration_sec": 27.156014366075397}}
|
| 7 |
+
{"timestamp": 1774091511.9092546, "event": "train_step", "step": 60, "epoch": 1, "metrics": {"train/step_loss": 1.2237938749256418, "train/step_real_loss": 0.9667061120271683, "train/lr": 9.7589742682592e-05, "train/step_canary_loss": 6.708333333333333, "perf/step_duration_sec": 8.495171755552292, "perf/samples_per_sec": 7.886832889071372, "perf/tokens_per_sec": 6228.008276044599, "perf/logical_batch_size": 67.0, "perf/logical_token_count": 52908.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
|
| 8 |
+
{"timestamp": 1774091593.5344725, "event": "train_step", "step": 70, "epoch": 1, "metrics": {"train/step_loss": 1.0743033794256358, "train/step_real_loss": 0.9387456253170967, "train/lr": 9.621307308142384e-05, "train/step_canary_loss": 9.75, "perf/step_duration_sec": 8.257235972210765, "perf/samples_per_sec": 7.8718835478062665, "perf/tokens_per_sec": 6568.057420488078, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 54234.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
|
| 9 |
+
{"timestamp": 1774091676.5485113, "event": "train_step", "step": 80, "epoch": 1, "metrics": {"train/step_loss": 0.8626528829336166, "train/step_real_loss": 0.8626528829336166, "train/lr": 9.454124979346391e-05, "perf/step_duration_sec": 8.386640733107924, "perf/samples_per_sec": 7.631184169765056, "perf/tokens_per_sec": 6658.446662625317, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 55842.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
|
| 10 |
+
{"timestamp": 1774091757.879373, "event": "train_step", "step": 90, "epoch": 1, "metrics": {"train/step_loss": 0.9340124875307083, "train/step_real_loss": 0.9340124875307083, "train/lr": 9.258495042083221e-05, "perf/step_duration_sec": 8.493970146402717, "perf/samples_per_sec": 7.534756880103312, "perf/tokens_per_sec": 5785.162786428072, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 49139.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
|
| 11 |
+
{"timestamp": 1774091839.6918592, "event": "train_step", "step": 100, "epoch": 1, "metrics": {"train/step_loss": 0.9577739440477812, "train/step_real_loss": 0.880453996360302, "train/lr": 9.035666945770107e-05, "train/step_canary_loss": 5.90625, "perf/step_duration_sec": 8.234961070120335, "perf/samples_per_sec": 7.893176354633353, "perf/tokens_per_sec": 6771.616711381143, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 55764.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
|
| 12 |
+
{"timestamp": 1774091866.8768253, "event": "eval_step", "step": 100, "epoch": 1, "metrics": {"eval/loss": 0.7754902612717671, "eval/duration_sec": 27.182465851306915}}
|
| 13 |
+
{"timestamp": 1774091950.1670244, "event": "train_step", "step": 110, "epoch": 1, "metrics": {"train/step_loss": 0.9567192720644402, "train/step_real_loss": 0.860762245953083, "train/lr": 8.787063849045118e-05, "train/step_canary_loss": 4.02734375, "perf/step_duration_sec": 8.036363879218698, "perf/samples_per_sec": 8.21266943507498, "perf/tokens_per_sec": 6478.054102878841, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 52060.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
|
| 14 |
+
{"timestamp": 1774092031.8751383, "event": "train_step", "step": 120, "epoch": 1, "metrics": {"train/step_loss": 0.9170611575245857, "train/step_real_loss": 0.9170611575245857, "train/lr": 8.5142735303366e-05, "perf/step_duration_sec": 7.79720464348793, "perf/samples_per_sec": 8.208069805305357, "perf/tokens_per_sec": 6600.4423832818775, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 51465.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
|
| 15 |
+
{"timestamp": 1774092114.4179726, "event": "train_step", "step": 130, "epoch": 1, "metrics": {"train/step_loss": 0.9158061426697355, "train/step_real_loss": 0.8711828961968422, "train/lr": 8.219038247038819e-05, "train/step_canary_loss": 2.34375, "perf/step_duration_sec": 8.22744246572256, "perf/samples_per_sec": 8.021933945447975, "perf/tokens_per_sec": 6342.311139506375, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 52181.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
|
| 16 |
+
{"timestamp": 1774092196.3902268, "event": "train_step", "step": 140, "epoch": 1, "metrics": {"train/step_loss": 0.9856266433542425, "train/step_real_loss": 0.8489470109343529, "train/lr": 7.903243608061246e-05, "train/step_canary_loss": 5.359375, "perf/step_duration_sec": 8.333684524521232, "perf/samples_per_sec": 7.9196662419605675, "perf/tokens_per_sec": 6461.007714123103, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 53844.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
|
| 17 |
+
{"timestamp": 1774092278.982639, "event": "train_step", "step": 150, "epoch": 1, "metrics": {"train/step_loss": 0.94221530854702, "train/step_real_loss": 0.94221530854702, "train/lr": 7.568906530820282e-05, "perf/step_duration_sec": 8.388992108404636, "perf/samples_per_sec": 7.629045202686584, "perf/tokens_per_sec": 6081.421861022835, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 51017.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
|
| 18 |
+
{"timestamp": 1774092306.1516783, "event": "eval_step", "step": 150, "epoch": 1, "metrics": {"eval/loss": 0.7622188641422096, "eval/duration_sec": 27.166858203709126}}
|
| 19 |
+
{"timestamp": 1774092388.1349308, "event": "train_step", "step": 160, "epoch": 1, "metrics": {"train/step_loss": 0.9667639210820198, "train/step_real_loss": 0.9667639210820198, "train/lr": 7.21816235958972e-05, "perf/step_duration_sec": 8.235355000942945, "perf/samples_per_sec": 7.771371117902267, "perf/tokens_per_sec": 6008.24129552819, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 49480.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
|
| 20 |
+
{"timestamp": 1774092469.9282126, "event": "train_step", "step": 170, "epoch": 1, "metrics": {"train/step_loss": 0.9262635037302971, "train/step_real_loss": 0.9262635037302971, "train/lr": 6.853251227482479e-05, "perf/step_duration_sec": 7.9527543764561415, "perf/samples_per_sec": 8.047526299752175, "perf/tokens_per_sec": 6280.842791759705, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 49950.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
|
| 21 |
+
{"timestamp": 1774092551.943189, "event": "train_step", "step": 180, "epoch": 1, "metrics": {"train/step_loss": 0.8705217910535408, "train/step_real_loss": 0.8467002063989639, "train/lr": 6.476503749166904e-05, "train/step_canary_loss": 1.6328125, "perf/step_duration_sec": 8.540382800623775, "perf/samples_per_sec": 7.7279908337574135, "perf/tokens_per_sec": 6385.076790236751, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 54531.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
|
| 22 |
+
{"timestamp": 1774092634.1860914, "event": "train_step", "step": 190, "epoch": 1, "metrics": {"train/step_loss": 0.8329041105050307, "train/step_real_loss": 0.830781526863575, "train/lr": 6.090326135695403e-05, "train/step_canary_loss": 0.96875, "perf/step_duration_sec": 7.938739079982042, "perf/samples_per_sec": 8.18769824088324, "perf/tokens_per_sec": 6963.07051322375, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 55278.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
|
| 23 |
+
{"timestamp": 1774092717.69117, "event": "train_step", "step": 200, "epoch": 1, "metrics": {"train/step_loss": 0.9225348509274997, "train/step_real_loss": 0.8622424304485321, "train/lr": 5.697184826514057e-05, "train/step_canary_loss": 4.78125, "perf/step_duration_sec": 8.257816776633263, "perf/samples_per_sec": 7.871329887571168, "perf/tokens_per_sec": 6772.4922352662325, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 55926.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
|
| 24 |
+
{"timestamp": 1774092744.8800561, "event": "eval_step", "step": 200, "epoch": 1, "metrics": {"eval/loss": 0.7524948388166811, "eval/duration_sec": 27.18629032932222}}
|
| 25 |
+
{"timestamp": 1774092830.2611096, "event": "train_epoch", "step": 207, "epoch": 1, "metrics": {"train/epoch_loss": 1.036294524908722, "train/epoch_real_loss": 0.957591299003593, "train/epoch_canary_loss": 5.154934052784185, "perf/epoch_duration_sec": 1814.8832574244589, "perf/epoch_samples_per_sec": 29.762244915220545, "perf/epoch_tokens_per_sec": 24175.732417227537, "perf/epoch_samples": 54015.0, "perf/epoch_tokens": 43876132.0, "system/cuda_epoch_peak_memory_gb": 88.38552379608154, "eval/loss": 0.7516984458284817, "eval/duration_sec": 27.281295781955123}}
|
| 26 |
+
{"timestamp": 1774092840.935869, "event": "audit_epoch", "step": 207, "epoch": 1, "metrics": {"audit/delta": 1e-05, "audit/num_canaries": 500.0, "audit/num_members": 250.0, "audit/paper_guess_fraction": 0.2, "audit/paper_guess_steps": 20.0, "audit/loss/auc": 0.998072, "audit/loss/empirical_epsilon/0.05": 3.4791953936219215, "audit/loss/empirical_epsilon/0.01": 3.023197554051876, "audit/loss/empirical_epsilon_details/0.05/epsilon": 3.4791953936219215, "audit/loss/empirical_epsilon_details/0.05/num_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.05/correct_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.01/epsilon": 3.023197554051876, "audit/loss/empirical_epsilon_details/0.01/num_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.01/correct_guesses": 100.0, "audit/embedding/auc": 0.96, "audit/embedding/empirical_epsilon/0.05": 3.4791953936219215, "audit/embedding/empirical_epsilon/0.01": 3.023197554051876, "audit/embedding/empirical_epsilon_details/0.05/epsilon": 3.4791953936219215, "audit/embedding/empirical_epsilon_details/0.05/num_guesses": 100.0, "audit/embedding/empirical_epsilon_details/0.05/correct_guesses": 100.0, "audit/embedding/empirical_epsilon_details/0.01/epsilon": 3.023197554051876, "audit/embedding/empirical_epsilon_details/0.01/num_guesses": 100.0, "audit/embedding/empirical_epsilon_details/0.01/correct_guesses": 100.0, "perf/audit_duration_sec": 6.3129072319716215}}
|
| 27 |
+
{"timestamp": 1774092866.0795767, "event": "train_step", "step": 210, "epoch": 2, "metrics": {"train/step_loss": 0.8786039505944108, "train/step_real_loss": 0.8291560411453247, "train/lr": 5.29959073680547e-05, "train/step_canary_loss": 2.4609375, "perf/step_duration_sec": 8.105620637536049, "perf/samples_per_sec": 8.142498020985931, "perf/tokens_per_sec": 6755.805933775599, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 54760.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.911073684692383, "system/cuda_max_memory_allocated_gb": 75.98184061050415}}
|
| 28 |
+
{"timestamp": 1774092947.7881637, "event": "train_step", "step": 220, "epoch": 2, "metrics": {"train/step_loss": 0.8338463682394761, "train/step_real_loss": 0.8286867365241051, "train/lr": 4.9000832207739676e-05, "train/step_canary_loss": 1.1640625, "perf/step_duration_sec": 8.100939376279712, "perf/samples_per_sec": 8.023760823383757, "perf/tokens_per_sec": 6522.083124668936, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 52835.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
|
| 29 |
+
{"timestamp": 1774093029.3670137, "event": "train_step", "step": 230, "epoch": 2, "metrics": {"train/step_loss": 0.7926744809517494, "train/step_real_loss": 0.8034120723605156, "train/lr": 4.501213853296425e-05, "train/step_canary_loss": 0.10546875, "perf/step_duration_sec": 8.09845926053822, "perf/samples_per_sec": 8.026218063073905, "perf/tokens_per_sec": 6708.683497950849, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 54330.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
|
| 30 |
+
{"timestamp": 1774093111.4853563, "event": "train_step", "step": 240, "epoch": 2, "metrics": {"train/step_loss": 0.8129148510369387, "train/step_real_loss": 0.7906041964888573, "train/lr": 4.1055301335220955e-05, "train/step_canary_loss": 1.52685546875, "perf/step_duration_sec": 8.505769751966, "perf/samples_per_sec": 7.75943881913156, "perf/tokens_per_sec": 6383.901937558238, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 54300.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
|
| 31 |
+
{"timestamp": 1774093192.88336, "event": "train_step", "step": 250, "epoch": 2, "metrics": {"train/step_loss": 0.8410789425556476, "train/step_real_loss": 0.850802831351757, "train/lr": 3.715559214503298e-05, "train/step_canary_loss": 0.21875, "perf/step_duration_sec": 8.232771871611476, "perf/samples_per_sec": 7.895275250385015, "perf/tokens_per_sec": 6654.745309889904, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 54787.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
|
| 32 |
+
{"timestamp": 1774093220.0732572, "event": "eval_step", "step": 250, "epoch": 2, "metrics": {"eval/loss": 0.748522541389383, "eval/duration_sec": 27.187080297619104}}
|
| 33 |
+
{"timestamp": 1774093302.2765276, "event": "train_step", "step": 260, "epoch": 2, "metrics": {"train/step_loss": 0.8724462010643699, "train/step_real_loss": 0.8488068357110023, "train/lr": 3.33379176277258e-05, "train/step_canary_loss": 1.62890625, "perf/step_duration_sec": 8.470736568793654, "perf/samples_per_sec": 7.7915302245551095, "perf/tokens_per_sec": 6277.49423773815, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 53175.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.911073684692383, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
|
| 34 |
+
{"timestamp": 1774093384.391679, "event": "train_step", "step": 270, "epoch": 2, "metrics": {"train/step_loss": 0.8640658855438232, "train/step_real_loss": 0.7926059737801552, "train/lr": 2.962666050951997e-05, "train/step_canary_loss": 5.4375, "perf/step_duration_sec": 7.947357261553407, "perf/samples_per_sec": 8.178819431516908, "perf/tokens_per_sec": 6407.161314659091, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 50920.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
|
| 35 |
+
{"timestamp": 1774093468.0585487, "event": "train_step", "step": 280, "epoch": 2, "metrics": {"train/step_loss": 0.7910019425245431, "train/step_real_loss": 0.8018507286906242, "train/lr": 2.604552384991855e-05, "train/step_canary_loss": 0.0966796875, "perf/step_duration_sec": 8.62639987282455, "perf/samples_per_sec": 7.53500892124967, "perf/tokens_per_sec": 6275.039506402552, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 54131.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
|
| 36 |
+
{"timestamp": 1774093550.7132335, "event": "train_step", "step": 290, "epoch": 2, "metrics": {"train/step_loss": 0.8430270507480159, "train/step_real_loss": 0.8613760396838188, "train/lr": 2.2617379654990623e-05, "train/step_canary_loss": 0.255859375, "perf/step_duration_sec": 8.335046991705894, "perf/samples_per_sec": 7.9183716739300705, "perf/tokens_per_sec": 6261.392413496007, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 52189.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
|
| 37 |
+
{"timestamp": 1774093632.5524132, "event": "train_step", "step": 300, "epoch": 2, "metrics": {"train/step_loss": 0.7691241015087474, "train/step_real_loss": 0.784599058330059, "train/lr": 1.936412279842705e-05, "train/step_canary_loss": 0.27392578125, "perf/step_duration_sec": 8.091827008873224, "perf/samples_per_sec": 8.156378025336753, "perf/tokens_per_sec": 5984.062677922065, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 48422.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.911073684692383, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
|
| 38 |
+
{"timestamp": 1774093659.719083, "event": "eval_step", "step": 300, "epoch": 2, "metrics": {"eval/loss": 0.7462242173737493, "eval/duration_sec": 27.16416385397315}}
|
| 39 |
+
{"timestamp": 1774093742.5612013, "event": "train_step", "step": 310, "epoch": 2, "metrics": {"train/step_loss": 0.7514625552928809, "train/step_real_loss": 0.7710395082831383, "train/lr": 1.6306531183346385e-05, "train/step_canary_loss": 0.125, "perf/step_duration_sec": 8.240195842459798, "perf/samples_per_sec": 8.009518373327666, "perf/tokens_per_sec": 6369.751520897312, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 52488.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
|
| 40 |
+
{"timestamp": 1774093824.76255, "event": "train_step", "step": 320, "epoch": 2, "metrics": {"train/step_loss": 0.8317307600608239, "train/step_real_loss": 0.8416290208697319, "train/lr": 1.3464133037968912e-05, "train/step_canary_loss": 0.1982421875, "perf/step_duration_sec": 7.953533122316003, "perf/samples_per_sec": 8.172468637569782, "perf/tokens_per_sec": 6226.666720107783, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 49524.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
|
| 41 |
+
{"timestamp": 1774093907.919859, "event": "train_step", "step": 330, "epoch": 2, "metrics": {"train/step_loss": 0.8719267258277306, "train/step_real_loss": 0.8635779246687889, "train/lr": 1.0855082192715294e-05, "train/step_canary_loss": 1.40625, "perf/step_duration_sec": 8.101611092686653, "perf/samples_per_sec": 8.023095561656334, "perf/tokens_per_sec": 5623.572827524039, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 45560.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
|
| 42 |
+
{"timestamp": 1774093989.7239172, "event": "train_step", "step": 340, "epoch": 2, "metrics": {"train/step_loss": 0.8476224361043988, "train/step_real_loss": 0.8228105828166008, "train/lr": 8.49604213531004e-06, "train/step_canary_loss": 1.6416015625, "perf/step_duration_sec": 8.274557610973716, "perf/samples_per_sec": 7.976257233676254, "perf/tokens_per_sec": 6674.435371234426, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 55228.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
|
| 43 |
+
{"timestamp": 1774094071.9116077, "event": "train_step", "step": 350, "epoch": 2, "metrics": {"train/step_loss": 0.8573094416355741, "train/step_real_loss": 0.816773310303688, "train/lr": 6.402079584406673e-06, "train/step_canary_loss": 1.376171875, "perf/step_duration_sec": 8.637357847765088, "perf/samples_per_sec": 7.988554048140279, "perf/tokens_per_sec": 6212.547974249382, "perf/logical_batch_size": 69.0, "perf/logical_token_count": 53660.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
|
| 44 |
+
{"timestamp": 1774094099.091823, "event": "eval_step", "step": 350, "epoch": 2, "metrics": {"eval/loss": 0.7456388631942629, "eval/duration_sec": 27.17722495831549}}
|
| 45 |
+
{"timestamp": 1774094181.5654633, "event": "train_step", "step": 360, "epoch": 2, "metrics": {"train/step_loss": 0.8446246770712046, "train/step_real_loss": 0.8453097268939018, "train/lr": 4.586568261458729e-06, "train/step_canary_loss": 0.80078125, "perf/step_duration_sec": 7.976690696552396, "perf/samples_per_sec": 8.148742689508275, "perf/tokens_per_sec": 6948.119478163339, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 55423.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
|
| 46 |
+
{"timestamp": 1774094263.5117958, "event": "train_step", "step": 370, "epoch": 2, "metrics": {"train/step_loss": 0.8721330687403679, "train/step_real_loss": 0.8721330687403679, "train/lr": 3.06110347542643e-06, "perf/step_duration_sec": 8.37641635723412, "perf/samples_per_sec": 7.640498904370688, "perf/tokens_per_sec": 5830.058812420977, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 48835.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
|
| 47 |
+
{"timestamp": 1774094346.1630964, "event": "train_step", "step": 380, "epoch": 2, "metrics": {"train/step_loss": 0.81929341004445, "train/step_real_loss": 0.8280055150389671, "train/lr": 1.8354280658494649e-06, "train/step_canary_loss": 0.26171875, "perf/step_duration_sec": 7.9715049508959055, "perf/samples_per_sec": 8.154043734576712, "perf/tokens_per_sec": 6787.676898315027, "perf/logical_batch_size": 65.0, "perf/logical_token_count": 54108.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
|
| 48 |
+
{"timestamp": 1774094429.2324018, "event": "train_step", "step": 390, "epoch": 2, "metrics": {"train/step_loss": 0.8554338663816452, "train/step_real_loss": 0.8554338663816452, "train/lr": 9.17370177272775e-07, "perf/step_duration_sec": 8.24701151996851, "perf/samples_per_sec": 7.760386880148844, "perf/tokens_per_sec": 6142.34621563781, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 50656.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
|
| 49 |
+
{"timestamp": 1774094512.2904398, "event": "train_step", "step": 400, "epoch": 2, "metrics": {"train/step_loss": 0.7998756021261215, "train/step_real_loss": 0.7998756021261215, "train/lr": 3.127932624475638e-07, "perf/step_duration_sec": 7.943094424903393, "perf/samples_per_sec": 8.0573132555677, "perf/tokens_per_sec": 6869.86671452841, "perf/logical_batch_size": 64.0, "perf/logical_token_count": 54568.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 82.1704511642456}}
|
| 50 |
+
{"timestamp": 1774094539.4831598, "event": "eval_step", "step": 400, "epoch": 2, "metrics": {"eval/loss": 0.7455017728538349, "eval/duration_sec": 27.190383709967136}}
|
| 51 |
+
{"timestamp": 1774094621.4459202, "event": "train_step", "step": 410, "epoch": 2, "metrics": {"train/step_loss": 0.854436350591255, "train/step_real_loss": 0.8517185375094414, "train/lr": 2.5558633627303928e-08, "train/step_canary_loss": 0.94140625, "perf/step_duration_sec": 8.378037076443434, "perf/samples_per_sec": 7.877740262760654, "perf/tokens_per_sec": 6647.619184760483, "perf/logical_batch_size": 66.0, "perf/logical_token_count": 55694.0, "perf/gradient_accumulation_steps": 8.0, "system/cuda_memory_allocated_gb": 17.816345691680908, "system/cuda_max_memory_allocated_gb": 88.38552379608154}}
|
| 52 |
+
{"timestamp": 1774094682.1799755, "event": "train_epoch", "step": 414, "epoch": 2, "metrics": {"train/epoch_loss": 0.840315752633972, "train/epoch_real_loss": 0.8362452851041504, "train/epoch_canary_loss": 1.0832555509625612, "perf/epoch_duration_sec": 1814.1021996028721, "perf/epoch_samples_per_sec": 29.765136722628178, "perf/epoch_tokens_per_sec": 24186.19579955584, "perf/epoch_samples": 53997.0, "perf/epoch_tokens": 43876231.0, "system/cuda_epoch_peak_memory_gb": 88.38552379608154, "eval/loss": 0.7454980848164394, "eval/duration_sec": 27.13153049722314}}
|
| 53 |
+
{"timestamp": 1774094693.161264, "event": "audit_epoch", "step": 414, "epoch": 2, "metrics": {"audit/delta": 1e-05, "audit/num_canaries": 500.0, "audit/num_members": 250.0, "audit/paper_guess_fraction": 0.2, "audit/paper_guess_steps": 20.0, "audit/loss/auc": 1.0, "audit/loss/empirical_epsilon/0.05": 3.4791953936219215, "audit/loss/empirical_epsilon/0.01": 3.023197554051876, "audit/loss/empirical_epsilon_details/0.05/epsilon": 3.4791953936219215, "audit/loss/empirical_epsilon_details/0.05/num_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.05/correct_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.01/epsilon": 3.023197554051876, "audit/loss/empirical_epsilon_details/0.01/num_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.01/correct_guesses": 100.0, "audit/embedding/auc": 0.916224, "audit/embedding/empirical_epsilon/0.05": 3.4791953936219215, "audit/embedding/empirical_epsilon/0.01": 3.023197554051876, "audit/embedding/empirical_epsilon_details/0.05/epsilon": 3.4791953936219215, "audit/embedding/empirical_epsilon_details/0.05/num_guesses": 100.0, "audit/embedding/empirical_epsilon_details/0.05/correct_guesses": 100.0, "audit/embedding/empirical_epsilon_details/0.01/epsilon": 3.023197554051876, "audit/embedding/empirical_epsilon_details/0.01/num_guesses": 100.0, "audit/embedding/empirical_epsilon_details/0.01/correct_guesses": 100.0, "perf/audit_duration_sec": 6.722778998315334}}
|
| 54 |
+
{"timestamp": 1774094704.3854914, "event": "audit_final", "step": 414, "epoch": 2, "metrics": {"audit/delta": 1e-05, "audit/num_canaries": 500.0, "audit/num_members": 250.0, "audit/paper_guess_fraction": 0.2, "audit/paper_guess_steps": 20.0, "audit/loss/auc": 1.0, "audit/loss/empirical_epsilon/0.05": 3.4791953936219215, "audit/loss/empirical_epsilon/0.01": 3.023197554051876, "audit/loss/empirical_epsilon_details/0.05/epsilon": 3.4791953936219215, "audit/loss/empirical_epsilon_details/0.05/num_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.05/correct_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.01/epsilon": 3.023197554051876, "audit/loss/empirical_epsilon_details/0.01/num_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.01/correct_guesses": 100.0, "audit/embedding/auc": 0.916224, "audit/embedding/empirical_epsilon/0.05": 3.4791953936219215, "audit/embedding/empirical_epsilon/0.01": 3.023197554051876, "audit/embedding/empirical_epsilon_details/0.05/epsilon": 3.4791953936219215, "audit/embedding/empirical_epsilon_details/0.05/num_guesses": 100.0, "audit/embedding/empirical_epsilon_details/0.05/correct_guesses": 100.0, "audit/embedding/empirical_epsilon_details/0.01/epsilon": 3.023197554051876, "audit/embedding/empirical_epsilon_details/0.01/num_guesses": 100.0, "audit/embedding/empirical_epsilon_details/0.01/correct_guesses": 100.0}}
|
| 55 |
+
{"timestamp": 1774094704.914049, "event": "energy_final", "step": 414, "epoch": null, "metrics": {"energy/codecarbon/duration": 3820.65162669681, "energy/codecarbon/emissions": 0.5568149811716148, "energy/codecarbon/emissions_rate": 0.0001457382236267942, "energy/codecarbon/cpu_power": 179.3439937196561, "energy/codecarbon/gpu_power": 2316.658494927477, "energy/codecarbon/ram_power": 70.0, "energy/codecarbon/cpu_energy": 0.18368298514637418, "energy/codecarbon/gpu_energy": 2.457235179675422, "energy/codecarbon/ram_energy": 0.07169108754482326, "energy/codecarbon/energy_consumed": 2.7126092523666183, "energy/codecarbon/water_consumed": 0.0, "energy/codecarbon/cpu_count": 224.0, "energy/codecarbon/gpu_count": 4.0, "energy/codecarbon/longitude": -121.9552, "energy/codecarbon/latitude": 37.3541, "energy/codecarbon/ram_total_size": 2015.5625190734863, "energy/codecarbon/cpu_utilization_percent": 2.2497114375655825, "energy/codecarbon/gpu_utilization_percent": 96.45330535152151, "energy/codecarbon/ram_utilization_percent": 2.3057974816369358, "energy/codecarbon/ram_used_gb": 46.78091819539025, "energy/codecarbon/pue": 1.0, "energy/codecarbon/wue": 0.0}}
|
starcoder2-7b/base/pretrain_lm_head.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c9a9b40bb4ecc60cfc3a7e2f7cd31be2121d8f32f92d1a4ecdac554172fb4e5b
|
| 3 |
+
size 457594472
|
starcoder2-7b/base/resolved_config.yaml
ADDED
|
@@ -0,0 +1,100 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
name: bigcode/starcoder2-7b
|
| 3 |
+
tokenizer_name: bigcode/starcoder2-7b
|
| 4 |
+
max_length: 1024
|
| 5 |
+
dtype: bfloat16
|
| 6 |
+
trust_remote_code: true
|
| 7 |
+
use_fast_tokenizer: true
|
| 8 |
+
cache_dir: null
|
| 9 |
+
local_files_only: false
|
| 10 |
+
low_cpu_mem_usage: true
|
| 11 |
+
tie_word_embeddings: true
|
| 12 |
+
gradient_checkpointing: false
|
| 13 |
+
use_chat_template: false
|
| 14 |
+
dataset:
|
| 15 |
+
name: melihcatal/codedp-cpt
|
| 16 |
+
split: train
|
| 17 |
+
mode: cpt
|
| 18 |
+
text_column: text
|
| 19 |
+
validation_ratio: 0.05
|
| 20 |
+
max_samples: -1
|
| 21 |
+
lora:
|
| 22 |
+
enabled: true
|
| 23 |
+
r: 16
|
| 24 |
+
alpha: 32
|
| 25 |
+
dropout: 0.05
|
| 26 |
+
target_modules:
|
| 27 |
+
- q_proj
|
| 28 |
+
- k_proj
|
| 29 |
+
- v_proj
|
| 30 |
+
- o_proj
|
| 31 |
+
modules_to_save:
|
| 32 |
+
- lm_head
|
| 33 |
+
bias: none
|
| 34 |
+
training:
|
| 35 |
+
seed: 42
|
| 36 |
+
epochs: 2
|
| 37 |
+
warmup_steps: null
|
| 38 |
+
warmup_ratio: 0.05
|
| 39 |
+
mixed_precision: false
|
| 40 |
+
mixed_precision_dtype: bfloat16
|
| 41 |
+
batch_size: 8
|
| 42 |
+
eval_batch_size: 8
|
| 43 |
+
eval_every_steps: 50
|
| 44 |
+
eval_every_epochs: 1
|
| 45 |
+
learning_rate: 0.0001
|
| 46 |
+
optimizer: adamw
|
| 47 |
+
lr_scheduler: cosine
|
| 48 |
+
adam_beta1: 0.9
|
| 49 |
+
adam_beta2: 0.999
|
| 50 |
+
adam_epsilon: 1.0e-08
|
| 51 |
+
sgd_momentum: 0.9
|
| 52 |
+
weight_decay: 0.01
|
| 53 |
+
max_grad_norm: 1.0
|
| 54 |
+
log_every: 10
|
| 55 |
+
gradient_accumulation_steps: 8
|
| 56 |
+
num_workers: 4
|
| 57 |
+
output_dir: runs/cpt/starcoder2-7b/base
|
| 58 |
+
distributed:
|
| 59 |
+
strategy: dpddp
|
| 60 |
+
backend: nccl
|
| 61 |
+
devices: null
|
| 62 |
+
dp:
|
| 63 |
+
module_validator: auto
|
| 64 |
+
target_delta: 1.0e-05
|
| 65 |
+
noise_multiplier: null
|
| 66 |
+
max_grad_norm: 1.0
|
| 67 |
+
grad_sample_mode: ghost
|
| 68 |
+
secure_mode: false
|
| 69 |
+
enabled: false
|
| 70 |
+
target_epsilon: 8.0
|
| 71 |
+
audit:
|
| 72 |
+
enabled: true
|
| 73 |
+
run_every_epoch: true
|
| 74 |
+
epoch_device: cuda
|
| 75 |
+
q_canary: auto
|
| 76 |
+
num_canaries: 500
|
| 77 |
+
prefix_length: 49
|
| 78 |
+
num_digits: 12
|
| 79 |
+
batch_size: 32
|
| 80 |
+
delta: 1.0e-05
|
| 81 |
+
p_values:
|
| 82 |
+
- 0.05
|
| 83 |
+
- 0.01
|
| 84 |
+
paper_guess_fraction: 0.2
|
| 85 |
+
paper_guess_steps: 20
|
| 86 |
+
enable_holdout_empirical_epsilon: false
|
| 87 |
+
holdout_seed: 42
|
| 88 |
+
tie_seed: 42
|
| 89 |
+
tracking:
|
| 90 |
+
enabled: true
|
| 91 |
+
tensorboard: true
|
| 92 |
+
wandb: false
|
| 93 |
+
wandb_project: codedp-finetune-h200-audit
|
| 94 |
+
wandb_run_name: starcoder2-7b-cpt-base
|
| 95 |
+
wandb_mode: online
|
| 96 |
+
codecarbon: true
|
| 97 |
+
codecarbon_output_file: codecarbon.csv
|
| 98 |
+
codecarbon_measure_power_secs: 15
|
| 99 |
+
codecarbon_country_iso_code: null
|
| 100 |
+
codecarbon_project_name: codedp-starcoder2-7b-cpt-base
|
starcoder2-7b/base/scalars.csv
ADDED
|
@@ -0,0 +1,613 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
timestamp,event,step,epoch,key,value
|
| 2 |
+
1774091071.9870658,train_step,10,1,train/step_loss,1.5507386380975896
|
| 3 |
+
1774091071.9870658,train_step,10,1,train/step_real_loss,1.2866992354393005
|
| 4 |
+
1774091071.9870658,train_step,10,1,train/lr,4.761904761904762e-05
|
| 5 |
+
1774091071.9870658,train_step,10,1,train/step_canary_loss,10.0
|
| 6 |
+
1774091071.9870658,train_step,10,1,perf/step_duration_sec,8.244300998747349
|
| 7 |
+
1774091071.9870658,train_step,10,1,perf/samples_per_sec,8.00553012438873
|
| 8 |
+
1774091071.9870658,train_step,10,1,perf/tokens_per_sec,6440.812872803661
|
| 9 |
+
1774091071.9870658,train_step,10,1,perf/logical_batch_size,66.0
|
| 10 |
+
1774091071.9870658,train_step,10,1,perf/logical_token_count,53100.0
|
| 11 |
+
1774091071.9870658,train_step,10,1,perf/gradient_accumulation_steps,8.0
|
| 12 |
+
1774091071.9870658,train_step,10,1,system/cuda_memory_allocated_gb,17.816345691680908
|
| 13 |
+
1774091071.9870658,train_step,10,1,system/cuda_max_memory_allocated_gb,82.16860580444336
|
| 14 |
+
1774091154.380765,train_step,20,1,train/step_loss,1.4504926845208923
|
| 15 |
+
1774091154.380765,train_step,20,1,train/step_real_loss,1.2006134390830994
|
| 16 |
+
1774091154.380765,train_step,20,1,train/lr,9.523809523809524e-05
|
| 17 |
+
1774091154.380765,train_step,20,1,train/step_canary_loss,6.78125
|
| 18 |
+
1774091154.380765,train_step,20,1,perf/step_duration_sec,8.244972918182611
|
| 19 |
+
1774091154.380765,train_step,20,1,perf/samples_per_sec,8.126163744242886
|
| 20 |
+
1774091154.380765,train_step,20,1,perf/tokens_per_sec,6422.095078472549
|
| 21 |
+
1774091154.380765,train_step,20,1,perf/logical_batch_size,67.0
|
| 22 |
+
1774091154.380765,train_step,20,1,perf/logical_token_count,52950.0
|
| 23 |
+
1774091154.380765,train_step,20,1,perf/gradient_accumulation_steps,8.0
|
| 24 |
+
1774091154.380765,train_step,20,1,system/cuda_memory_allocated_gb,17.911073684692383
|
| 25 |
+
1774091154.380765,train_step,20,1,system/cuda_max_memory_allocated_gb,82.16860580444336
|
| 26 |
+
1774091235.6533117,train_step,30,1,train/step_loss,0.9861254319548607
|
| 27 |
+
1774091235.6533117,train_step,30,1,train/step_real_loss,0.9861254319548607
|
| 28 |
+
1774091235.6533117,train_step,30,1,train/lr,9.98706541985615e-05
|
| 29 |
+
1774091235.6533117,train_step,30,1,perf/step_duration_sec,7.973596462979913
|
| 30 |
+
1774091235.6533117,train_step,30,1,perf/samples_per_sec,8.026490968929942
|
| 31 |
+
1774091235.6533117,train_step,30,1,perf/tokens_per_sec,7148.9697609673985
|
| 32 |
+
1774091235.6533117,train_step,30,1,perf/logical_batch_size,64.0
|
| 33 |
+
1774091235.6533117,train_step,30,1,perf/logical_token_count,57003.0
|
| 34 |
+
1774091235.6533117,train_step,30,1,perf/gradient_accumulation_steps,8.0
|
| 35 |
+
1774091235.6533117,train_step,30,1,system/cuda_memory_allocated_gb,17.816345691680908
|
| 36 |
+
1774091235.6533117,train_step,30,1,system/cuda_max_memory_allocated_gb,82.16860580444336
|
| 37 |
+
1774091318.662035,train_step,40,1,train/step_loss,1.0557814985513687
|
| 38 |
+
1774091318.662035,train_step,40,1,train/step_real_loss,1.0557814985513687
|
| 39 |
+
1774091318.662035,train_step,40,1,train/lr,9.942439201095397e-05
|
| 40 |
+
1774091318.662035,train_step,40,1,perf/step_duration_sec,8.119743634015322
|
| 41 |
+
1774091318.662035,train_step,40,1,perf/samples_per_sec,7.882022251526572
|
| 42 |
+
1774091318.662035,train_step,40,1,perf/tokens_per_sec,6340.347961767047
|
| 43 |
+
1774091318.662035,train_step,40,1,perf/logical_batch_size,64.0
|
| 44 |
+
1774091318.662035,train_step,40,1,perf/logical_token_count,51482.0
|
| 45 |
+
1774091318.662035,train_step,40,1,perf/gradient_accumulation_steps,8.0
|
| 46 |
+
1774091318.662035,train_step,40,1,system/cuda_memory_allocated_gb,17.816345691680908
|
| 47 |
+
1774091318.662035,train_step,40,1,system/cuda_max_memory_allocated_gb,88.38552379608154
|
| 48 |
+
1774091402.0365882,train_step,50,1,train/step_loss,1.017115435233483
|
| 49 |
+
1774091402.0365882,train_step,50,1,train/step_real_loss,0.9500000476837158
|
| 50 |
+
1774091402.0365882,train_step,50,1,train/lr,9.866246608261724e-05
|
| 51 |
+
1774091402.0365882,train_step,50,1,train/step_canary_loss,5.3125
|
| 52 |
+
1774091402.0365882,train_step,50,1,perf/step_duration_sec,8.119036318734288
|
| 53 |
+
1774091402.0365882,train_step,50,1,perf/samples_per_sec,8.005876245437603
|
| 54 |
+
1774091402.0365882,train_step,50,1,perf/tokens_per_sec,6307.645142790008
|
| 55 |
+
1774091402.0365882,train_step,50,1,perf/logical_batch_size,65.0
|
| 56 |
+
1774091402.0365882,train_step,50,1,perf/logical_token_count,51212.0
|
| 57 |
+
1774091402.0365882,train_step,50,1,perf/gradient_accumulation_steps,8.0
|
| 58 |
+
1774091402.0365882,train_step,50,1,system/cuda_memory_allocated_gb,17.816345691680908
|
| 59 |
+
1774091402.0365882,train_step,50,1,system/cuda_max_memory_allocated_gb,88.38552379608154
|
| 60 |
+
1774091429.1958544,eval_step,50,1,eval/loss,0.8033588177513802
|
| 61 |
+
1774091429.1958544,eval_step,50,1,eval/duration_sec,27.156014366075397
|
| 62 |
+
1774091511.9092546,train_step,60,1,train/step_loss,1.2237938749256418
|
| 63 |
+
1774091511.9092546,train_step,60,1,train/step_real_loss,0.9667061120271683
|
| 64 |
+
1774091511.9092546,train_step,60,1,train/lr,9.7589742682592e-05
|
| 65 |
+
1774091511.9092546,train_step,60,1,train/step_canary_loss,6.708333333333333
|
| 66 |
+
1774091511.9092546,train_step,60,1,perf/step_duration_sec,8.495171755552292
|
| 67 |
+
1774091511.9092546,train_step,60,1,perf/samples_per_sec,7.886832889071372
|
| 68 |
+
1774091511.9092546,train_step,60,1,perf/tokens_per_sec,6228.008276044599
|
| 69 |
+
1774091511.9092546,train_step,60,1,perf/logical_batch_size,67.0
|
| 70 |
+
1774091511.9092546,train_step,60,1,perf/logical_token_count,52908.0
|
| 71 |
+
1774091511.9092546,train_step,60,1,perf/gradient_accumulation_steps,8.0
|
| 72 |
+
1774091511.9092546,train_step,60,1,system/cuda_memory_allocated_gb,17.816345691680908
|
| 73 |
+
1774091511.9092546,train_step,60,1,system/cuda_max_memory_allocated_gb,88.38552379608154
|
| 74 |
+
1774091593.5344725,train_step,70,1,train/step_loss,1.0743033794256358
|
| 75 |
+
1774091593.5344725,train_step,70,1,train/step_real_loss,0.9387456253170967
|
| 76 |
+
1774091593.5344725,train_step,70,1,train/lr,9.621307308142384e-05
|
| 77 |
+
1774091593.5344725,train_step,70,1,train/step_canary_loss,9.75
|
| 78 |
+
1774091593.5344725,train_step,70,1,perf/step_duration_sec,8.257235972210765
|
| 79 |
+
1774091593.5344725,train_step,70,1,perf/samples_per_sec,7.8718835478062665
|
| 80 |
+
1774091593.5344725,train_step,70,1,perf/tokens_per_sec,6568.057420488078
|
| 81 |
+
1774091593.5344725,train_step,70,1,perf/logical_batch_size,65.0
|
| 82 |
+
1774091593.5344725,train_step,70,1,perf/logical_token_count,54234.0
|
| 83 |
+
1774091593.5344725,train_step,70,1,perf/gradient_accumulation_steps,8.0
|
| 84 |
+
1774091593.5344725,train_step,70,1,system/cuda_memory_allocated_gb,17.816345691680908
|
| 85 |
+
1774091593.5344725,train_step,70,1,system/cuda_max_memory_allocated_gb,88.38552379608154
|
| 86 |
+
1774091676.5485113,train_step,80,1,train/step_loss,0.8626528829336166
|
| 87 |
+
1774091676.5485113,train_step,80,1,train/step_real_loss,0.8626528829336166
|
| 88 |
+
1774091676.5485113,train_step,80,1,train/lr,9.454124979346391e-05
|
| 89 |
+
1774091676.5485113,train_step,80,1,perf/step_duration_sec,8.386640733107924
|
| 90 |
+
1774091676.5485113,train_step,80,1,perf/samples_per_sec,7.631184169765056
|
| 91 |
+
1774091676.5485113,train_step,80,1,perf/tokens_per_sec,6658.446662625317
|
| 92 |
+
1774091676.5485113,train_step,80,1,perf/logical_batch_size,64.0
|
| 93 |
+
1774091676.5485113,train_step,80,1,perf/logical_token_count,55842.0
|
| 94 |
+
1774091676.5485113,train_step,80,1,perf/gradient_accumulation_steps,8.0
|
| 95 |
+
1774091676.5485113,train_step,80,1,system/cuda_memory_allocated_gb,17.816345691680908
|
| 96 |
+
1774091676.5485113,train_step,80,1,system/cuda_max_memory_allocated_gb,88.38552379608154
|
| 97 |
+
1774091757.879373,train_step,90,1,train/step_loss,0.9340124875307083
|
| 98 |
+
1774091757.879373,train_step,90,1,train/step_real_loss,0.9340124875307083
|
| 99 |
+
1774091757.879373,train_step,90,1,train/lr,9.258495042083221e-05
|
| 100 |
+
1774091757.879373,train_step,90,1,perf/step_duration_sec,8.493970146402717
|
| 101 |
+
1774091757.879373,train_step,90,1,perf/samples_per_sec,7.534756880103312
|
| 102 |
+
1774091757.879373,train_step,90,1,perf/tokens_per_sec,5785.162786428072
|
| 103 |
+
1774091757.879373,train_step,90,1,perf/logical_batch_size,64.0
|
| 104 |
+
1774091757.879373,train_step,90,1,perf/logical_token_count,49139.0
|
| 105 |
+
1774091757.879373,train_step,90,1,perf/gradient_accumulation_steps,8.0
|
| 106 |
+
1774091757.879373,train_step,90,1,system/cuda_memory_allocated_gb,17.816345691680908
|
| 107 |
+
1774091757.879373,train_step,90,1,system/cuda_max_memory_allocated_gb,88.38552379608154
|
| 108 |
+
1774091839.6918592,train_step,100,1,train/step_loss,0.9577739440477812
|
| 109 |
+
1774091839.6918592,train_step,100,1,train/step_real_loss,0.880453996360302
|
| 110 |
+
1774091839.6918592,train_step,100,1,train/lr,9.035666945770107e-05
|
| 111 |
+
1774091839.6918592,train_step,100,1,train/step_canary_loss,5.90625
|
| 112 |
+
1774091839.6918592,train_step,100,1,perf/step_duration_sec,8.234961070120335
|
| 113 |
+
1774091839.6918592,train_step,100,1,perf/samples_per_sec,7.893176354633353
|
| 114 |
+
1774091839.6918592,train_step,100,1,perf/tokens_per_sec,6771.616711381143
|
| 115 |
+
1774091839.6918592,train_step,100,1,perf/logical_batch_size,65.0
|
| 116 |
+
1774091839.6918592,train_step,100,1,perf/logical_token_count,55764.0
|
| 117 |
+
1774091839.6918592,train_step,100,1,perf/gradient_accumulation_steps,8.0
|
| 118 |
+
1774091839.6918592,train_step,100,1,system/cuda_memory_allocated_gb,17.816345691680908
|
| 119 |
+
1774091839.6918592,train_step,100,1,system/cuda_max_memory_allocated_gb,88.38552379608154
|
| 120 |
+
1774091866.8768253,eval_step,100,1,eval/loss,0.7754902612717671
|
| 121 |
+
1774091866.8768253,eval_step,100,1,eval/duration_sec,27.182465851306915
|
| 122 |
+
1774091950.1670244,train_step,110,1,train/step_loss,0.9567192720644402
|
| 123 |
+
1774091950.1670244,train_step,110,1,train/step_real_loss,0.860762245953083
|
| 124 |
+
1774091950.1670244,train_step,110,1,train/lr,8.787063849045118e-05
|
| 125 |
+
1774091950.1670244,train_step,110,1,train/step_canary_loss,4.02734375
|
| 126 |
+
1774091950.1670244,train_step,110,1,perf/step_duration_sec,8.036363879218698
|
| 127 |
+
1774091950.1670244,train_step,110,1,perf/samples_per_sec,8.21266943507498
|
| 128 |
+
1774091950.1670244,train_step,110,1,perf/tokens_per_sec,6478.054102878841
|
| 129 |
+
1774091950.1670244,train_step,110,1,perf/logical_batch_size,66.0
|
| 130 |
+
1774091950.1670244,train_step,110,1,perf/logical_token_count,52060.0
|
| 131 |
+
1774091950.1670244,train_step,110,1,perf/gradient_accumulation_steps,8.0
|
| 132 |
+
1774091950.1670244,train_step,110,1,system/cuda_memory_allocated_gb,17.816345691680908
|
| 133 |
+
1774091950.1670244,train_step,110,1,system/cuda_max_memory_allocated_gb,88.38552379608154
|
| 134 |
+
1774092031.8751383,train_step,120,1,train/step_loss,0.9170611575245857
|
| 135 |
+
1774092031.8751383,train_step,120,1,train/step_real_loss,0.9170611575245857
|
| 136 |
+
1774092031.8751383,train_step,120,1,train/lr,8.5142735303366e-05
|
| 137 |
+
1774092031.8751383,train_step,120,1,perf/step_duration_sec,7.79720464348793
|
| 138 |
+
1774092031.8751383,train_step,120,1,perf/samples_per_sec,8.208069805305357
|
| 139 |
+
1774092031.8751383,train_step,120,1,perf/tokens_per_sec,6600.4423832818775
|
| 140 |
+
1774092031.8751383,train_step,120,1,perf/logical_batch_size,64.0
|
| 141 |
+
1774092031.8751383,train_step,120,1,perf/logical_token_count,51465.0
|
| 142 |
+
1774092031.8751383,train_step,120,1,perf/gradient_accumulation_steps,8.0
|
| 143 |
+
1774092031.8751383,train_step,120,1,system/cuda_memory_allocated_gb,17.816345691680908
|
| 144 |
+
1774092031.8751383,train_step,120,1,system/cuda_max_memory_allocated_gb,88.38552379608154
|
| 145 |
+
1774092114.4179726,train_step,130,1,train/step_loss,0.9158061426697355
|
| 146 |
+
1774092114.4179726,train_step,130,1,train/step_real_loss,0.8711828961968422
|
| 147 |
+
1774092114.4179726,train_step,130,1,train/lr,8.219038247038819e-05
|
| 148 |
+
1774092114.4179726,train_step,130,1,train/step_canary_loss,2.34375
|
| 149 |
+
1774092114.4179726,train_step,130,1,perf/step_duration_sec,8.22744246572256
|
| 150 |
+
1774092114.4179726,train_step,130,1,perf/samples_per_sec,8.021933945447975
|
| 151 |
+
1774092114.4179726,train_step,130,1,perf/tokens_per_sec,6342.311139506375
|
| 152 |
+
1774092114.4179726,train_step,130,1,perf/logical_batch_size,66.0
|
| 153 |
+
1774092114.4179726,train_step,130,1,perf/logical_token_count,52181.0
|
| 154 |
+
1774092114.4179726,train_step,130,1,perf/gradient_accumulation_steps,8.0
|
| 155 |
+
1774092114.4179726,train_step,130,1,system/cuda_memory_allocated_gb,17.816345691680908
|
| 156 |
+
1774092114.4179726,train_step,130,1,system/cuda_max_memory_allocated_gb,88.38552379608154
|
| 157 |
+
1774092196.3902268,train_step,140,1,train/step_loss,0.9856266433542425
|
| 158 |
+
1774092196.3902268,train_step,140,1,train/step_real_loss,0.8489470109343529
|
| 159 |
+
1774092196.3902268,train_step,140,1,train/lr,7.903243608061246e-05
|
| 160 |
+
1774092196.3902268,train_step,140,1,train/step_canary_loss,5.359375
|
| 161 |
+
1774092196.3902268,train_step,140,1,perf/step_duration_sec,8.333684524521232
|
| 162 |
+
1774092196.3902268,train_step,140,1,perf/samples_per_sec,7.9196662419605675
|
| 163 |
+
1774092196.3902268,train_step,140,1,perf/tokens_per_sec,6461.007714123103
|
| 164 |
+
1774092196.3902268,train_step,140,1,perf/logical_batch_size,66.0
|
| 165 |
+
1774092196.3902268,train_step,140,1,perf/logical_token_count,53844.0
|
| 166 |
+
1774092196.3902268,train_step,140,1,perf/gradient_accumulation_steps,8.0
|
| 167 |
+
1774092196.3902268,train_step,140,1,system/cuda_memory_allocated_gb,17.816345691680908
|
| 168 |
+
1774092196.3902268,train_step,140,1,system/cuda_max_memory_allocated_gb,88.38552379608154
|
| 169 |
+
1774092278.982639,train_step,150,1,train/step_loss,0.94221530854702
|
| 170 |
+
1774092278.982639,train_step,150,1,train/step_real_loss,0.94221530854702
|
| 171 |
+
1774092278.982639,train_step,150,1,train/lr,7.568906530820282e-05
|
| 172 |
+
1774092278.982639,train_step,150,1,perf/step_duration_sec,8.388992108404636
|
| 173 |
+
1774092278.982639,train_step,150,1,perf/samples_per_sec,7.629045202686584
|
| 174 |
+
1774092278.982639,train_step,150,1,perf/tokens_per_sec,6081.421861022835
|
| 175 |
+
1774092278.982639,train_step,150,1,perf/logical_batch_size,64.0
|
| 176 |
+
1774092278.982639,train_step,150,1,perf/logical_token_count,51017.0
|
| 177 |
+
1774092278.982639,train_step,150,1,perf/gradient_accumulation_steps,8.0
|
| 178 |
+
1774092278.982639,train_step,150,1,system/cuda_memory_allocated_gb,17.816345691680908
|
| 179 |
+
1774092278.982639,train_step,150,1,system/cuda_max_memory_allocated_gb,88.38552379608154
|
| 180 |
+
1774092306.1516783,eval_step,150,1,eval/loss,0.7622188641422096
|
| 181 |
+
1774092306.1516783,eval_step,150,1,eval/duration_sec,27.166858203709126
|
| 182 |
+
1774092388.1349308,train_step,160,1,train/step_loss,0.9667639210820198
|
| 183 |
+
1774092388.1349308,train_step,160,1,train/step_real_loss,0.9667639210820198
|
| 184 |
+
1774092388.1349308,train_step,160,1,train/lr,7.21816235958972e-05
|
| 185 |
+
1774092388.1349308,train_step,160,1,perf/step_duration_sec,8.235355000942945
|
| 186 |
+
1774092388.1349308,train_step,160,1,perf/samples_per_sec,7.771371117902267
|
| 187 |
+
1774092388.1349308,train_step,160,1,perf/tokens_per_sec,6008.24129552819
|
| 188 |
+
1774092388.1349308,train_step,160,1,perf/logical_batch_size,64.0
|
| 189 |
+
1774092388.1349308,train_step,160,1,perf/logical_token_count,49480.0
|
| 190 |
+
1774092388.1349308,train_step,160,1,perf/gradient_accumulation_steps,8.0
|
| 191 |
+
1774092388.1349308,train_step,160,1,system/cuda_memory_allocated_gb,17.816345691680908
|
| 192 |
+
1774092388.1349308,train_step,160,1,system/cuda_max_memory_allocated_gb,88.38552379608154
|
| 193 |
+
1774092469.9282126,train_step,170,1,train/step_loss,0.9262635037302971
|
| 194 |
+
1774092469.9282126,train_step,170,1,train/step_real_loss,0.9262635037302971
|
| 195 |
+
1774092469.9282126,train_step,170,1,train/lr,6.853251227482479e-05
|
| 196 |
+
1774092469.9282126,train_step,170,1,perf/step_duration_sec,7.9527543764561415
|
| 197 |
+
1774092469.9282126,train_step,170,1,perf/samples_per_sec,8.047526299752175
|
| 198 |
+
1774092469.9282126,train_step,170,1,perf/tokens_per_sec,6280.842791759705
|
| 199 |
+
1774092469.9282126,train_step,170,1,perf/logical_batch_size,64.0
|
| 200 |
+
1774092469.9282126,train_step,170,1,perf/logical_token_count,49950.0
|
| 201 |
+
1774092469.9282126,train_step,170,1,perf/gradient_accumulation_steps,8.0
|
| 202 |
+
1774092469.9282126,train_step,170,1,system/cuda_memory_allocated_gb,17.816345691680908
|
| 203 |
+
1774092469.9282126,train_step,170,1,system/cuda_max_memory_allocated_gb,88.38552379608154
|
| 204 |
+
1774092551.943189,train_step,180,1,train/step_loss,0.8705217910535408
|
| 205 |
+
1774092551.943189,train_step,180,1,train/step_real_loss,0.8467002063989639
|
| 206 |
+
1774092551.943189,train_step,180,1,train/lr,6.476503749166904e-05
|
| 207 |
+
1774092551.943189,train_step,180,1,train/step_canary_loss,1.6328125
|
| 208 |
+
1774092551.943189,train_step,180,1,perf/step_duration_sec,8.540382800623775
|
| 209 |
+
1774092551.943189,train_step,180,1,perf/samples_per_sec,7.7279908337574135
|
| 210 |
+
1774092551.943189,train_step,180,1,perf/tokens_per_sec,6385.076790236751
|
| 211 |
+
1774092551.943189,train_step,180,1,perf/logical_batch_size,66.0
|
| 212 |
+
1774092551.943189,train_step,180,1,perf/logical_token_count,54531.0
|
| 213 |
+
1774092551.943189,train_step,180,1,perf/gradient_accumulation_steps,8.0
|
| 214 |
+
1774092551.943189,train_step,180,1,system/cuda_memory_allocated_gb,17.816345691680908
|
| 215 |
+
1774092551.943189,train_step,180,1,system/cuda_max_memory_allocated_gb,88.38552379608154
|
| 216 |
+
1774092634.1860914,train_step,190,1,train/step_loss,0.8329041105050307
|
| 217 |
+
1774092634.1860914,train_step,190,1,train/step_real_loss,0.830781526863575
|
| 218 |
+
1774092634.1860914,train_step,190,1,train/lr,6.090326135695403e-05
|
| 219 |
+
1774092634.1860914,train_step,190,1,train/step_canary_loss,0.96875
|
| 220 |
+
1774092634.1860914,train_step,190,1,perf/step_duration_sec,7.938739079982042
|
| 221 |
+
1774092634.1860914,train_step,190,1,perf/samples_per_sec,8.18769824088324
|
| 222 |
+
1774092634.1860914,train_step,190,1,perf/tokens_per_sec,6963.07051322375
|
| 223 |
+
1774092634.1860914,train_step,190,1,perf/logical_batch_size,65.0
|
| 224 |
+
1774092634.1860914,train_step,190,1,perf/logical_token_count,55278.0
|
| 225 |
+
1774092634.1860914,train_step,190,1,perf/gradient_accumulation_steps,8.0
|
| 226 |
+
1774092634.1860914,train_step,190,1,system/cuda_memory_allocated_gb,17.816345691680908
|
| 227 |
+
1774092634.1860914,train_step,190,1,system/cuda_max_memory_allocated_gb,88.38552379608154
|
| 228 |
+
1774092717.69117,train_step,200,1,train/step_loss,0.9225348509274997
|
| 229 |
+
1774092717.69117,train_step,200,1,train/step_real_loss,0.8622424304485321
|
| 230 |
+
1774092717.69117,train_step,200,1,train/lr,5.697184826514057e-05
|
| 231 |
+
1774092717.69117,train_step,200,1,train/step_canary_loss,4.78125
|
| 232 |
+
1774092717.69117,train_step,200,1,perf/step_duration_sec,8.257816776633263
|
| 233 |
+
1774092717.69117,train_step,200,1,perf/samples_per_sec,7.871329887571168
|
| 234 |
+
1774092717.69117,train_step,200,1,perf/tokens_per_sec,6772.4922352662325
|
| 235 |
+
1774092717.69117,train_step,200,1,perf/logical_batch_size,65.0
|
| 236 |
+
1774092717.69117,train_step,200,1,perf/logical_token_count,55926.0
|
| 237 |
+
1774092717.69117,train_step,200,1,perf/gradient_accumulation_steps,8.0
|
| 238 |
+
1774092717.69117,train_step,200,1,system/cuda_memory_allocated_gb,17.816345691680908
|
| 239 |
+
1774092717.69117,train_step,200,1,system/cuda_max_memory_allocated_gb,88.38552379608154
|
| 240 |
+
1774092744.8800561,eval_step,200,1,eval/loss,0.7524948388166811
|
| 241 |
+
1774092744.8800561,eval_step,200,1,eval/duration_sec,27.18629032932222
|
| 242 |
+
1774092830.2611096,train_epoch,207,1,train/epoch_loss,1.036294524908722
|
| 243 |
+
1774092830.2611096,train_epoch,207,1,train/epoch_real_loss,0.957591299003593
|
| 244 |
+
1774092830.2611096,train_epoch,207,1,train/epoch_canary_loss,5.154934052784185
|
| 245 |
+
1774092830.2611096,train_epoch,207,1,perf/epoch_duration_sec,1814.8832574244589
|
| 246 |
+
1774092830.2611096,train_epoch,207,1,perf/epoch_samples_per_sec,29.762244915220545
|
| 247 |
+
1774092830.2611096,train_epoch,207,1,perf/epoch_tokens_per_sec,24175.732417227537
|
| 248 |
+
1774092830.2611096,train_epoch,207,1,perf/epoch_samples,54015.0
|
| 249 |
+
1774092830.2611096,train_epoch,207,1,perf/epoch_tokens,43876132.0
|
| 250 |
+
1774092830.2611096,train_epoch,207,1,system/cuda_epoch_peak_memory_gb,88.38552379608154
|
| 251 |
+
1774092830.2611096,train_epoch,207,1,eval/loss,0.7516984458284817
|
| 252 |
+
1774092830.2611096,train_epoch,207,1,eval/duration_sec,27.281295781955123
|
| 253 |
+
1774092840.935869,audit_epoch,207,1,audit/delta,1e-05
|
| 254 |
+
1774092840.935869,audit_epoch,207,1,audit/num_canaries,500.0
|
| 255 |
+
1774092840.935869,audit_epoch,207,1,audit/num_members,250.0
|
| 256 |
+
1774092840.935869,audit_epoch,207,1,audit/paper_guess_fraction,0.2
|
| 257 |
+
1774092840.935869,audit_epoch,207,1,audit/paper_guess_steps,20.0
|
| 258 |
+
1774092840.935869,audit_epoch,207,1,audit/loss/auc,0.998072
|
| 259 |
+
1774092840.935869,audit_epoch,207,1,audit/loss/empirical_epsilon/0.05,3.4791953936219215
|
| 260 |
+
1774092840.935869,audit_epoch,207,1,audit/loss/empirical_epsilon/0.01,3.023197554051876
|
| 261 |
+
1774092840.935869,audit_epoch,207,1,audit/loss/empirical_epsilon_details/0.05/epsilon,3.4791953936219215
|
| 262 |
+
1774092840.935869,audit_epoch,207,1,audit/loss/empirical_epsilon_details/0.05/num_guesses,100.0
|
| 263 |
+
1774092840.935869,audit_epoch,207,1,audit/loss/empirical_epsilon_details/0.05/correct_guesses,100.0
|
| 264 |
+
1774092840.935869,audit_epoch,207,1,audit/loss/empirical_epsilon_details/0.01/epsilon,3.023197554051876
|
| 265 |
+
1774092840.935869,audit_epoch,207,1,audit/loss/empirical_epsilon_details/0.01/num_guesses,100.0
|
| 266 |
+
1774092840.935869,audit_epoch,207,1,audit/loss/empirical_epsilon_details/0.01/correct_guesses,100.0
|
| 267 |
+
1774092840.935869,audit_epoch,207,1,audit/embedding/auc,0.96
|
| 268 |
+
1774092840.935869,audit_epoch,207,1,audit/embedding/empirical_epsilon/0.05,3.4791953936219215
|
| 269 |
+
1774092840.935869,audit_epoch,207,1,audit/embedding/empirical_epsilon/0.01,3.023197554051876
|
| 270 |
+
1774092840.935869,audit_epoch,207,1,audit/embedding/empirical_epsilon_details/0.05/epsilon,3.4791953936219215
|
| 271 |
+
1774092840.935869,audit_epoch,207,1,audit/embedding/empirical_epsilon_details/0.05/num_guesses,100.0
|
| 272 |
+
1774092840.935869,audit_epoch,207,1,audit/embedding/empirical_epsilon_details/0.05/correct_guesses,100.0
|
| 273 |
+
1774092840.935869,audit_epoch,207,1,audit/embedding/empirical_epsilon_details/0.01/epsilon,3.023197554051876
|
| 274 |
+
1774092840.935869,audit_epoch,207,1,audit/embedding/empirical_epsilon_details/0.01/num_guesses,100.0
|
| 275 |
+
1774092840.935869,audit_epoch,207,1,audit/embedding/empirical_epsilon_details/0.01/correct_guesses,100.0
|
| 276 |
+
1774092840.935869,audit_epoch,207,1,perf/audit_duration_sec,6.3129072319716215
|
| 277 |
+
1774092866.0795767,train_step,210,2,train/step_loss,0.8786039505944108
|
| 278 |
+
1774092866.0795767,train_step,210,2,train/step_real_loss,0.8291560411453247
|
| 279 |
+
1774092866.0795767,train_step,210,2,train/lr,5.29959073680547e-05
|
| 280 |
+
1774092866.0795767,train_step,210,2,train/step_canary_loss,2.4609375
|
| 281 |
+
1774092866.0795767,train_step,210,2,perf/step_duration_sec,8.105620637536049
|
| 282 |
+
1774092866.0795767,train_step,210,2,perf/samples_per_sec,8.142498020985931
|
| 283 |
+
1774092866.0795767,train_step,210,2,perf/tokens_per_sec,6755.805933775599
|
| 284 |
+
1774092866.0795767,train_step,210,2,perf/logical_batch_size,66.0
|
| 285 |
+
1774092866.0795767,train_step,210,2,perf/logical_token_count,54760.0
|
| 286 |
+
1774092866.0795767,train_step,210,2,perf/gradient_accumulation_steps,8.0
|
| 287 |
+
1774092866.0795767,train_step,210,2,system/cuda_memory_allocated_gb,17.911073684692383
|
| 288 |
+
1774092866.0795767,train_step,210,2,system/cuda_max_memory_allocated_gb,75.98184061050415
|
| 289 |
+
1774092947.7881637,train_step,220,2,train/step_loss,0.8338463682394761
|
| 290 |
+
1774092947.7881637,train_step,220,2,train/step_real_loss,0.8286867365241051
|
| 291 |
+
1774092947.7881637,train_step,220,2,train/lr,4.9000832207739676e-05
|
| 292 |
+
1774092947.7881637,train_step,220,2,train/step_canary_loss,1.1640625
|
| 293 |
+
1774092947.7881637,train_step,220,2,perf/step_duration_sec,8.100939376279712
|
| 294 |
+
1774092947.7881637,train_step,220,2,perf/samples_per_sec,8.023760823383757
|
| 295 |
+
1774092947.7881637,train_step,220,2,perf/tokens_per_sec,6522.083124668936
|
| 296 |
+
1774092947.7881637,train_step,220,2,perf/logical_batch_size,65.0
|
| 297 |
+
1774092947.7881637,train_step,220,2,perf/logical_token_count,52835.0
|
| 298 |
+
1774092947.7881637,train_step,220,2,perf/gradient_accumulation_steps,8.0
|
| 299 |
+
1774092947.7881637,train_step,220,2,system/cuda_memory_allocated_gb,17.816345691680908
|
| 300 |
+
1774092947.7881637,train_step,220,2,system/cuda_max_memory_allocated_gb,82.1704511642456
|
| 301 |
+
1774093029.3670137,train_step,230,2,train/step_loss,0.7926744809517494
|
| 302 |
+
1774093029.3670137,train_step,230,2,train/step_real_loss,0.8034120723605156
|
| 303 |
+
1774093029.3670137,train_step,230,2,train/lr,4.501213853296425e-05
|
| 304 |
+
1774093029.3670137,train_step,230,2,train/step_canary_loss,0.10546875
|
| 305 |
+
1774093029.3670137,train_step,230,2,perf/step_duration_sec,8.09845926053822
|
| 306 |
+
1774093029.3670137,train_step,230,2,perf/samples_per_sec,8.026218063073905
|
| 307 |
+
1774093029.3670137,train_step,230,2,perf/tokens_per_sec,6708.683497950849
|
| 308 |
+
1774093029.3670137,train_step,230,2,perf/logical_batch_size,65.0
|
| 309 |
+
1774093029.3670137,train_step,230,2,perf/logical_token_count,54330.0
|
| 310 |
+
1774093029.3670137,train_step,230,2,perf/gradient_accumulation_steps,8.0
|
| 311 |
+
1774093029.3670137,train_step,230,2,system/cuda_memory_allocated_gb,17.816345691680908
|
| 312 |
+
1774093029.3670137,train_step,230,2,system/cuda_max_memory_allocated_gb,82.1704511642456
|
| 313 |
+
1774093111.4853563,train_step,240,2,train/step_loss,0.8129148510369387
|
| 314 |
+
1774093111.4853563,train_step,240,2,train/step_real_loss,0.7906041964888573
|
| 315 |
+
1774093111.4853563,train_step,240,2,train/lr,4.1055301335220955e-05
|
| 316 |
+
1774093111.4853563,train_step,240,2,train/step_canary_loss,1.52685546875
|
| 317 |
+
1774093111.4853563,train_step,240,2,perf/step_duration_sec,8.505769751966
|
| 318 |
+
1774093111.4853563,train_step,240,2,perf/samples_per_sec,7.75943881913156
|
| 319 |
+
1774093111.4853563,train_step,240,2,perf/tokens_per_sec,6383.901937558238
|
| 320 |
+
1774093111.4853563,train_step,240,2,perf/logical_batch_size,66.0
|
| 321 |
+
1774093111.4853563,train_step,240,2,perf/logical_token_count,54300.0
|
| 322 |
+
1774093111.4853563,train_step,240,2,perf/gradient_accumulation_steps,8.0
|
| 323 |
+
1774093111.4853563,train_step,240,2,system/cuda_memory_allocated_gb,17.816345691680908
|
| 324 |
+
1774093111.4853563,train_step,240,2,system/cuda_max_memory_allocated_gb,82.1704511642456
|
| 325 |
+
1774093192.88336,train_step,250,2,train/step_loss,0.8410789425556476
|
| 326 |
+
1774093192.88336,train_step,250,2,train/step_real_loss,0.850802831351757
|
| 327 |
+
1774093192.88336,train_step,250,2,train/lr,3.715559214503298e-05
|
| 328 |
+
1774093192.88336,train_step,250,2,train/step_canary_loss,0.21875
|
| 329 |
+
1774093192.88336,train_step,250,2,perf/step_duration_sec,8.232771871611476
|
| 330 |
+
1774093192.88336,train_step,250,2,perf/samples_per_sec,7.895275250385015
|
| 331 |
+
1774093192.88336,train_step,250,2,perf/tokens_per_sec,6654.745309889904
|
| 332 |
+
1774093192.88336,train_step,250,2,perf/logical_batch_size,65.0
|
| 333 |
+
1774093192.88336,train_step,250,2,perf/logical_token_count,54787.0
|
| 334 |
+
1774093192.88336,train_step,250,2,perf/gradient_accumulation_steps,8.0
|
| 335 |
+
1774093192.88336,train_step,250,2,system/cuda_memory_allocated_gb,17.816345691680908
|
| 336 |
+
1774093192.88336,train_step,250,2,system/cuda_max_memory_allocated_gb,82.1704511642456
|
| 337 |
+
1774093220.0732572,eval_step,250,2,eval/loss,0.748522541389383
|
| 338 |
+
1774093220.0732572,eval_step,250,2,eval/duration_sec,27.187080297619104
|
| 339 |
+
1774093302.2765276,train_step,260,2,train/step_loss,0.8724462010643699
|
| 340 |
+
1774093302.2765276,train_step,260,2,train/step_real_loss,0.8488068357110023
|
| 341 |
+
1774093302.2765276,train_step,260,2,train/lr,3.33379176277258e-05
|
| 342 |
+
1774093302.2765276,train_step,260,2,train/step_canary_loss,1.62890625
|
| 343 |
+
1774093302.2765276,train_step,260,2,perf/step_duration_sec,8.470736568793654
|
| 344 |
+
1774093302.2765276,train_step,260,2,perf/samples_per_sec,7.7915302245551095
|
| 345 |
+
1774093302.2765276,train_step,260,2,perf/tokens_per_sec,6277.49423773815
|
| 346 |
+
1774093302.2765276,train_step,260,2,perf/logical_batch_size,66.0
|
| 347 |
+
1774093302.2765276,train_step,260,2,perf/logical_token_count,53175.0
|
| 348 |
+
1774093302.2765276,train_step,260,2,perf/gradient_accumulation_steps,8.0
|
| 349 |
+
1774093302.2765276,train_step,260,2,system/cuda_memory_allocated_gb,17.911073684692383
|
| 350 |
+
1774093302.2765276,train_step,260,2,system/cuda_max_memory_allocated_gb,82.1704511642456
|
| 351 |
+
1774093384.391679,train_step,270,2,train/step_loss,0.8640658855438232
|
| 352 |
+
1774093384.391679,train_step,270,2,train/step_real_loss,0.7926059737801552
|
| 353 |
+
1774093384.391679,train_step,270,2,train/lr,2.962666050951997e-05
|
| 354 |
+
1774093384.391679,train_step,270,2,train/step_canary_loss,5.4375
|
| 355 |
+
1774093384.391679,train_step,270,2,perf/step_duration_sec,7.947357261553407
|
| 356 |
+
1774093384.391679,train_step,270,2,perf/samples_per_sec,8.178819431516908
|
| 357 |
+
1774093384.391679,train_step,270,2,perf/tokens_per_sec,6407.161314659091
|
| 358 |
+
1774093384.391679,train_step,270,2,perf/logical_batch_size,65.0
|
| 359 |
+
1774093384.391679,train_step,270,2,perf/logical_token_count,50920.0
|
| 360 |
+
1774093384.391679,train_step,270,2,perf/gradient_accumulation_steps,8.0
|
| 361 |
+
1774093384.391679,train_step,270,2,system/cuda_memory_allocated_gb,17.816345691680908
|
| 362 |
+
1774093384.391679,train_step,270,2,system/cuda_max_memory_allocated_gb,82.1704511642456
|
| 363 |
+
1774093468.0585487,train_step,280,2,train/step_loss,0.7910019425245431
|
| 364 |
+
1774093468.0585487,train_step,280,2,train/step_real_loss,0.8018507286906242
|
| 365 |
+
1774093468.0585487,train_step,280,2,train/lr,2.604552384991855e-05
|
| 366 |
+
1774093468.0585487,train_step,280,2,train/step_canary_loss,0.0966796875
|
| 367 |
+
1774093468.0585487,train_step,280,2,perf/step_duration_sec,8.62639987282455
|
| 368 |
+
1774093468.0585487,train_step,280,2,perf/samples_per_sec,7.53500892124967
|
| 369 |
+
1774093468.0585487,train_step,280,2,perf/tokens_per_sec,6275.039506402552
|
| 370 |
+
1774093468.0585487,train_step,280,2,perf/logical_batch_size,65.0
|
| 371 |
+
1774093468.0585487,train_step,280,2,perf/logical_token_count,54131.0
|
| 372 |
+
1774093468.0585487,train_step,280,2,perf/gradient_accumulation_steps,8.0
|
| 373 |
+
1774093468.0585487,train_step,280,2,system/cuda_memory_allocated_gb,17.816345691680908
|
| 374 |
+
1774093468.0585487,train_step,280,2,system/cuda_max_memory_allocated_gb,82.1704511642456
|
| 375 |
+
1774093550.7132335,train_step,290,2,train/step_loss,0.8430270507480159
|
| 376 |
+
1774093550.7132335,train_step,290,2,train/step_real_loss,0.8613760396838188
|
| 377 |
+
1774093550.7132335,train_step,290,2,train/lr,2.2617379654990623e-05
|
| 378 |
+
1774093550.7132335,train_step,290,2,train/step_canary_loss,0.255859375
|
| 379 |
+
1774093550.7132335,train_step,290,2,perf/step_duration_sec,8.335046991705894
|
| 380 |
+
1774093550.7132335,train_step,290,2,perf/samples_per_sec,7.9183716739300705
|
| 381 |
+
1774093550.7132335,train_step,290,2,perf/tokens_per_sec,6261.392413496007
|
| 382 |
+
1774093550.7132335,train_step,290,2,perf/logical_batch_size,66.0
|
| 383 |
+
1774093550.7132335,train_step,290,2,perf/logical_token_count,52189.0
|
| 384 |
+
1774093550.7132335,train_step,290,2,perf/gradient_accumulation_steps,8.0
|
| 385 |
+
1774093550.7132335,train_step,290,2,system/cuda_memory_allocated_gb,17.816345691680908
|
| 386 |
+
1774093550.7132335,train_step,290,2,system/cuda_max_memory_allocated_gb,82.1704511642456
|
| 387 |
+
1774093632.5524132,train_step,300,2,train/step_loss,0.7691241015087474
|
| 388 |
+
1774093632.5524132,train_step,300,2,train/step_real_loss,0.784599058330059
|
| 389 |
+
1774093632.5524132,train_step,300,2,train/lr,1.936412279842705e-05
|
| 390 |
+
1774093632.5524132,train_step,300,2,train/step_canary_loss,0.27392578125
|
| 391 |
+
1774093632.5524132,train_step,300,2,perf/step_duration_sec,8.091827008873224
|
| 392 |
+
1774093632.5524132,train_step,300,2,perf/samples_per_sec,8.156378025336753
|
| 393 |
+
1774093632.5524132,train_step,300,2,perf/tokens_per_sec,5984.062677922065
|
| 394 |
+
1774093632.5524132,train_step,300,2,perf/logical_batch_size,66.0
|
| 395 |
+
1774093632.5524132,train_step,300,2,perf/logical_token_count,48422.0
|
| 396 |
+
1774093632.5524132,train_step,300,2,perf/gradient_accumulation_steps,8.0
|
| 397 |
+
1774093632.5524132,train_step,300,2,system/cuda_memory_allocated_gb,17.911073684692383
|
| 398 |
+
1774093632.5524132,train_step,300,2,system/cuda_max_memory_allocated_gb,82.1704511642456
|
| 399 |
+
1774093659.719083,eval_step,300,2,eval/loss,0.7462242173737493
|
| 400 |
+
1774093659.719083,eval_step,300,2,eval/duration_sec,27.16416385397315
|
| 401 |
+
1774093742.5612013,train_step,310,2,train/step_loss,0.7514625552928809
|
| 402 |
+
1774093742.5612013,train_step,310,2,train/step_real_loss,0.7710395082831383
|
| 403 |
+
1774093742.5612013,train_step,310,2,train/lr,1.6306531183346385e-05
|
| 404 |
+
1774093742.5612013,train_step,310,2,train/step_canary_loss,0.125
|
| 405 |
+
1774093742.5612013,train_step,310,2,perf/step_duration_sec,8.240195842459798
|
| 406 |
+
1774093742.5612013,train_step,310,2,perf/samples_per_sec,8.009518373327666
|
| 407 |
+
1774093742.5612013,train_step,310,2,perf/tokens_per_sec,6369.751520897312
|
| 408 |
+
1774093742.5612013,train_step,310,2,perf/logical_batch_size,66.0
|
| 409 |
+
1774093742.5612013,train_step,310,2,perf/logical_token_count,52488.0
|
| 410 |
+
1774093742.5612013,train_step,310,2,perf/gradient_accumulation_steps,8.0
|
| 411 |
+
1774093742.5612013,train_step,310,2,system/cuda_memory_allocated_gb,17.816345691680908
|
| 412 |
+
1774093742.5612013,train_step,310,2,system/cuda_max_memory_allocated_gb,82.1704511642456
|
| 413 |
+
1774093824.76255,train_step,320,2,train/step_loss,0.8317307600608239
|
| 414 |
+
1774093824.76255,train_step,320,2,train/step_real_loss,0.8416290208697319
|
| 415 |
+
1774093824.76255,train_step,320,2,train/lr,1.3464133037968912e-05
|
| 416 |
+
1774093824.76255,train_step,320,2,train/step_canary_loss,0.1982421875
|
| 417 |
+
1774093824.76255,train_step,320,2,perf/step_duration_sec,7.953533122316003
|
| 418 |
+
1774093824.76255,train_step,320,2,perf/samples_per_sec,8.172468637569782
|
| 419 |
+
1774093824.76255,train_step,320,2,perf/tokens_per_sec,6226.666720107783
|
| 420 |
+
1774093824.76255,train_step,320,2,perf/logical_batch_size,65.0
|
| 421 |
+
1774093824.76255,train_step,320,2,perf/logical_token_count,49524.0
|
| 422 |
+
1774093824.76255,train_step,320,2,perf/gradient_accumulation_steps,8.0
|
| 423 |
+
1774093824.76255,train_step,320,2,system/cuda_memory_allocated_gb,17.816345691680908
|
| 424 |
+
1774093824.76255,train_step,320,2,system/cuda_max_memory_allocated_gb,82.1704511642456
|
| 425 |
+
1774093907.919859,train_step,330,2,train/step_loss,0.8719267258277306
|
| 426 |
+
1774093907.919859,train_step,330,2,train/step_real_loss,0.8635779246687889
|
| 427 |
+
1774093907.919859,train_step,330,2,train/lr,1.0855082192715294e-05
|
| 428 |
+
1774093907.919859,train_step,330,2,train/step_canary_loss,1.40625
|
| 429 |
+
1774093907.919859,train_step,330,2,perf/step_duration_sec,8.101611092686653
|
| 430 |
+
1774093907.919859,train_step,330,2,perf/samples_per_sec,8.023095561656334
|
| 431 |
+
1774093907.919859,train_step,330,2,perf/tokens_per_sec,5623.572827524039
|
| 432 |
+
1774093907.919859,train_step,330,2,perf/logical_batch_size,65.0
|
| 433 |
+
1774093907.919859,train_step,330,2,perf/logical_token_count,45560.0
|
| 434 |
+
1774093907.919859,train_step,330,2,perf/gradient_accumulation_steps,8.0
|
| 435 |
+
1774093907.919859,train_step,330,2,system/cuda_memory_allocated_gb,17.816345691680908
|
| 436 |
+
1774093907.919859,train_step,330,2,system/cuda_max_memory_allocated_gb,82.1704511642456
|
| 437 |
+
1774093989.7239172,train_step,340,2,train/step_loss,0.8476224361043988
|
| 438 |
+
1774093989.7239172,train_step,340,2,train/step_real_loss,0.8228105828166008
|
| 439 |
+
1774093989.7239172,train_step,340,2,train/lr,8.49604213531004e-06
|
| 440 |
+
1774093989.7239172,train_step,340,2,train/step_canary_loss,1.6416015625
|
| 441 |
+
1774093989.7239172,train_step,340,2,perf/step_duration_sec,8.274557610973716
|
| 442 |
+
1774093989.7239172,train_step,340,2,perf/samples_per_sec,7.976257233676254
|
| 443 |
+
1774093989.7239172,train_step,340,2,perf/tokens_per_sec,6674.435371234426
|
| 444 |
+
1774093989.7239172,train_step,340,2,perf/logical_batch_size,66.0
|
| 445 |
+
1774093989.7239172,train_step,340,2,perf/logical_token_count,55228.0
|
| 446 |
+
1774093989.7239172,train_step,340,2,perf/gradient_accumulation_steps,8.0
|
| 447 |
+
1774093989.7239172,train_step,340,2,system/cuda_memory_allocated_gb,17.816345691680908
|
| 448 |
+
1774093989.7239172,train_step,340,2,system/cuda_max_memory_allocated_gb,82.1704511642456
|
| 449 |
+
1774094071.9116077,train_step,350,2,train/step_loss,0.8573094416355741
|
| 450 |
+
1774094071.9116077,train_step,350,2,train/step_real_loss,0.816773310303688
|
| 451 |
+
1774094071.9116077,train_step,350,2,train/lr,6.402079584406673e-06
|
| 452 |
+
1774094071.9116077,train_step,350,2,train/step_canary_loss,1.376171875
|
| 453 |
+
1774094071.9116077,train_step,350,2,perf/step_duration_sec,8.637357847765088
|
| 454 |
+
1774094071.9116077,train_step,350,2,perf/samples_per_sec,7.988554048140279
|
| 455 |
+
1774094071.9116077,train_step,350,2,perf/tokens_per_sec,6212.547974249382
|
| 456 |
+
1774094071.9116077,train_step,350,2,perf/logical_batch_size,69.0
|
| 457 |
+
1774094071.9116077,train_step,350,2,perf/logical_token_count,53660.0
|
| 458 |
+
1774094071.9116077,train_step,350,2,perf/gradient_accumulation_steps,8.0
|
| 459 |
+
1774094071.9116077,train_step,350,2,system/cuda_memory_allocated_gb,17.816345691680908
|
| 460 |
+
1774094071.9116077,train_step,350,2,system/cuda_max_memory_allocated_gb,82.1704511642456
|
| 461 |
+
1774094099.091823,eval_step,350,2,eval/loss,0.7456388631942629
|
| 462 |
+
1774094099.091823,eval_step,350,2,eval/duration_sec,27.17722495831549
|
| 463 |
+
1774094181.5654633,train_step,360,2,train/step_loss,0.8446246770712046
|
| 464 |
+
1774094181.5654633,train_step,360,2,train/step_real_loss,0.8453097268939018
|
| 465 |
+
1774094181.5654633,train_step,360,2,train/lr,4.586568261458729e-06
|
| 466 |
+
1774094181.5654633,train_step,360,2,train/step_canary_loss,0.80078125
|
| 467 |
+
1774094181.5654633,train_step,360,2,perf/step_duration_sec,7.976690696552396
|
| 468 |
+
1774094181.5654633,train_step,360,2,perf/samples_per_sec,8.148742689508275
|
| 469 |
+
1774094181.5654633,train_step,360,2,perf/tokens_per_sec,6948.119478163339
|
| 470 |
+
1774094181.5654633,train_step,360,2,perf/logical_batch_size,65.0
|
| 471 |
+
1774094181.5654633,train_step,360,2,perf/logical_token_count,55423.0
|
| 472 |
+
1774094181.5654633,train_step,360,2,perf/gradient_accumulation_steps,8.0
|
| 473 |
+
1774094181.5654633,train_step,360,2,system/cuda_memory_allocated_gb,17.816345691680908
|
| 474 |
+
1774094181.5654633,train_step,360,2,system/cuda_max_memory_allocated_gb,82.1704511642456
|
| 475 |
+
1774094263.5117958,train_step,370,2,train/step_loss,0.8721330687403679
|
| 476 |
+
1774094263.5117958,train_step,370,2,train/step_real_loss,0.8721330687403679
|
| 477 |
+
1774094263.5117958,train_step,370,2,train/lr,3.06110347542643e-06
|
| 478 |
+
1774094263.5117958,train_step,370,2,perf/step_duration_sec,8.37641635723412
|
| 479 |
+
1774094263.5117958,train_step,370,2,perf/samples_per_sec,7.640498904370688
|
| 480 |
+
1774094263.5117958,train_step,370,2,perf/tokens_per_sec,5830.058812420977
|
| 481 |
+
1774094263.5117958,train_step,370,2,perf/logical_batch_size,64.0
|
| 482 |
+
1774094263.5117958,train_step,370,2,perf/logical_token_count,48835.0
|
| 483 |
+
1774094263.5117958,train_step,370,2,perf/gradient_accumulation_steps,8.0
|
| 484 |
+
1774094263.5117958,train_step,370,2,system/cuda_memory_allocated_gb,17.816345691680908
|
| 485 |
+
1774094263.5117958,train_step,370,2,system/cuda_max_memory_allocated_gb,82.1704511642456
|
| 486 |
+
1774094346.1630964,train_step,380,2,train/step_loss,0.81929341004445
|
| 487 |
+
1774094346.1630964,train_step,380,2,train/step_real_loss,0.8280055150389671
|
| 488 |
+
1774094346.1630964,train_step,380,2,train/lr,1.8354280658494649e-06
|
| 489 |
+
1774094346.1630964,train_step,380,2,train/step_canary_loss,0.26171875
|
| 490 |
+
1774094346.1630964,train_step,380,2,perf/step_duration_sec,7.9715049508959055
|
| 491 |
+
1774094346.1630964,train_step,380,2,perf/samples_per_sec,8.154043734576712
|
| 492 |
+
1774094346.1630964,train_step,380,2,perf/tokens_per_sec,6787.676898315027
|
| 493 |
+
1774094346.1630964,train_step,380,2,perf/logical_batch_size,65.0
|
| 494 |
+
1774094346.1630964,train_step,380,2,perf/logical_token_count,54108.0
|
| 495 |
+
1774094346.1630964,train_step,380,2,perf/gradient_accumulation_steps,8.0
|
| 496 |
+
1774094346.1630964,train_step,380,2,system/cuda_memory_allocated_gb,17.816345691680908
|
| 497 |
+
1774094346.1630964,train_step,380,2,system/cuda_max_memory_allocated_gb,82.1704511642456
|
| 498 |
+
1774094429.2324018,train_step,390,2,train/step_loss,0.8554338663816452
|
| 499 |
+
1774094429.2324018,train_step,390,2,train/step_real_loss,0.8554338663816452
|
| 500 |
+
1774094429.2324018,train_step,390,2,train/lr,9.17370177272775e-07
|
| 501 |
+
1774094429.2324018,train_step,390,2,perf/step_duration_sec,8.24701151996851
|
| 502 |
+
1774094429.2324018,train_step,390,2,perf/samples_per_sec,7.760386880148844
|
| 503 |
+
1774094429.2324018,train_step,390,2,perf/tokens_per_sec,6142.34621563781
|
| 504 |
+
1774094429.2324018,train_step,390,2,perf/logical_batch_size,64.0
|
| 505 |
+
1774094429.2324018,train_step,390,2,perf/logical_token_count,50656.0
|
| 506 |
+
1774094429.2324018,train_step,390,2,perf/gradient_accumulation_steps,8.0
|
| 507 |
+
1774094429.2324018,train_step,390,2,system/cuda_memory_allocated_gb,17.816345691680908
|
| 508 |
+
1774094429.2324018,train_step,390,2,system/cuda_max_memory_allocated_gb,82.1704511642456
|
| 509 |
+
1774094512.2904398,train_step,400,2,train/step_loss,0.7998756021261215
|
| 510 |
+
1774094512.2904398,train_step,400,2,train/step_real_loss,0.7998756021261215
|
| 511 |
+
1774094512.2904398,train_step,400,2,train/lr,3.127932624475638e-07
|
| 512 |
+
1774094512.2904398,train_step,400,2,perf/step_duration_sec,7.943094424903393
|
| 513 |
+
1774094512.2904398,train_step,400,2,perf/samples_per_sec,8.0573132555677
|
| 514 |
+
1774094512.2904398,train_step,400,2,perf/tokens_per_sec,6869.86671452841
|
| 515 |
+
1774094512.2904398,train_step,400,2,perf/logical_batch_size,64.0
|
| 516 |
+
1774094512.2904398,train_step,400,2,perf/logical_token_count,54568.0
|
| 517 |
+
1774094512.2904398,train_step,400,2,perf/gradient_accumulation_steps,8.0
|
| 518 |
+
1774094512.2904398,train_step,400,2,system/cuda_memory_allocated_gb,17.816345691680908
|
| 519 |
+
1774094512.2904398,train_step,400,2,system/cuda_max_memory_allocated_gb,82.1704511642456
|
| 520 |
+
1774094539.4831598,eval_step,400,2,eval/loss,0.7455017728538349
|
| 521 |
+
1774094539.4831598,eval_step,400,2,eval/duration_sec,27.190383709967136
|
| 522 |
+
1774094621.4459202,train_step,410,2,train/step_loss,0.854436350591255
|
| 523 |
+
1774094621.4459202,train_step,410,2,train/step_real_loss,0.8517185375094414
|
| 524 |
+
1774094621.4459202,train_step,410,2,train/lr,2.5558633627303928e-08
|
| 525 |
+
1774094621.4459202,train_step,410,2,train/step_canary_loss,0.94140625
|
| 526 |
+
1774094621.4459202,train_step,410,2,perf/step_duration_sec,8.378037076443434
|
| 527 |
+
1774094621.4459202,train_step,410,2,perf/samples_per_sec,7.877740262760654
|
| 528 |
+
1774094621.4459202,train_step,410,2,perf/tokens_per_sec,6647.619184760483
|
| 529 |
+
1774094621.4459202,train_step,410,2,perf/logical_batch_size,66.0
|
| 530 |
+
1774094621.4459202,train_step,410,2,perf/logical_token_count,55694.0
|
| 531 |
+
1774094621.4459202,train_step,410,2,perf/gradient_accumulation_steps,8.0
|
| 532 |
+
1774094621.4459202,train_step,410,2,system/cuda_memory_allocated_gb,17.816345691680908
|
| 533 |
+
1774094621.4459202,train_step,410,2,system/cuda_max_memory_allocated_gb,88.38552379608154
|
| 534 |
+
1774094682.1799755,train_epoch,414,2,train/epoch_loss,0.840315752633972
|
| 535 |
+
1774094682.1799755,train_epoch,414,2,train/epoch_real_loss,0.8362452851041504
|
| 536 |
+
1774094682.1799755,train_epoch,414,2,train/epoch_canary_loss,1.0832555509625612
|
| 537 |
+
1774094682.1799755,train_epoch,414,2,perf/epoch_duration_sec,1814.1021996028721
|
| 538 |
+
1774094682.1799755,train_epoch,414,2,perf/epoch_samples_per_sec,29.765136722628178
|
| 539 |
+
1774094682.1799755,train_epoch,414,2,perf/epoch_tokens_per_sec,24186.19579955584
|
| 540 |
+
1774094682.1799755,train_epoch,414,2,perf/epoch_samples,53997.0
|
| 541 |
+
1774094682.1799755,train_epoch,414,2,perf/epoch_tokens,43876231.0
|
| 542 |
+
1774094682.1799755,train_epoch,414,2,system/cuda_epoch_peak_memory_gb,88.38552379608154
|
| 543 |
+
1774094682.1799755,train_epoch,414,2,eval/loss,0.7454980848164394
|
| 544 |
+
1774094682.1799755,train_epoch,414,2,eval/duration_sec,27.13153049722314
|
| 545 |
+
1774094693.161264,audit_epoch,414,2,audit/delta,1e-05
|
| 546 |
+
1774094693.161264,audit_epoch,414,2,audit/num_canaries,500.0
|
| 547 |
+
1774094693.161264,audit_epoch,414,2,audit/num_members,250.0
|
| 548 |
+
1774094693.161264,audit_epoch,414,2,audit/paper_guess_fraction,0.2
|
| 549 |
+
1774094693.161264,audit_epoch,414,2,audit/paper_guess_steps,20.0
|
| 550 |
+
1774094693.161264,audit_epoch,414,2,audit/loss/auc,1.0
|
| 551 |
+
1774094693.161264,audit_epoch,414,2,audit/loss/empirical_epsilon/0.05,3.4791953936219215
|
| 552 |
+
1774094693.161264,audit_epoch,414,2,audit/loss/empirical_epsilon/0.01,3.023197554051876
|
| 553 |
+
1774094693.161264,audit_epoch,414,2,audit/loss/empirical_epsilon_details/0.05/epsilon,3.4791953936219215
|
| 554 |
+
1774094693.161264,audit_epoch,414,2,audit/loss/empirical_epsilon_details/0.05/num_guesses,100.0
|
| 555 |
+
1774094693.161264,audit_epoch,414,2,audit/loss/empirical_epsilon_details/0.05/correct_guesses,100.0
|
| 556 |
+
1774094693.161264,audit_epoch,414,2,audit/loss/empirical_epsilon_details/0.01/epsilon,3.023197554051876
|
| 557 |
+
1774094693.161264,audit_epoch,414,2,audit/loss/empirical_epsilon_details/0.01/num_guesses,100.0
|
| 558 |
+
1774094693.161264,audit_epoch,414,2,audit/loss/empirical_epsilon_details/0.01/correct_guesses,100.0
|
| 559 |
+
1774094693.161264,audit_epoch,414,2,audit/embedding/auc,0.916224
|
| 560 |
+
1774094693.161264,audit_epoch,414,2,audit/embedding/empirical_epsilon/0.05,3.4791953936219215
|
| 561 |
+
1774094693.161264,audit_epoch,414,2,audit/embedding/empirical_epsilon/0.01,3.023197554051876
|
| 562 |
+
1774094693.161264,audit_epoch,414,2,audit/embedding/empirical_epsilon_details/0.05/epsilon,3.4791953936219215
|
| 563 |
+
1774094693.161264,audit_epoch,414,2,audit/embedding/empirical_epsilon_details/0.05/num_guesses,100.0
|
| 564 |
+
1774094693.161264,audit_epoch,414,2,audit/embedding/empirical_epsilon_details/0.05/correct_guesses,100.0
|
| 565 |
+
1774094693.161264,audit_epoch,414,2,audit/embedding/empirical_epsilon_details/0.01/epsilon,3.023197554051876
|
| 566 |
+
1774094693.161264,audit_epoch,414,2,audit/embedding/empirical_epsilon_details/0.01/num_guesses,100.0
|
| 567 |
+
1774094693.161264,audit_epoch,414,2,audit/embedding/empirical_epsilon_details/0.01/correct_guesses,100.0
|
| 568 |
+
1774094693.161264,audit_epoch,414,2,perf/audit_duration_sec,6.722778998315334
|
| 569 |
+
1774094704.3854914,audit_final,414,2,audit/delta,1e-05
|
| 570 |
+
1774094704.3854914,audit_final,414,2,audit/num_canaries,500.0
|
| 571 |
+
1774094704.3854914,audit_final,414,2,audit/num_members,250.0
|
| 572 |
+
1774094704.3854914,audit_final,414,2,audit/paper_guess_fraction,0.2
|
| 573 |
+
1774094704.3854914,audit_final,414,2,audit/paper_guess_steps,20.0
|
| 574 |
+
1774094704.3854914,audit_final,414,2,audit/loss/auc,1.0
|
| 575 |
+
1774094704.3854914,audit_final,414,2,audit/loss/empirical_epsilon/0.05,3.4791953936219215
|
| 576 |
+
1774094704.3854914,audit_final,414,2,audit/loss/empirical_epsilon/0.01,3.023197554051876
|
| 577 |
+
1774094704.3854914,audit_final,414,2,audit/loss/empirical_epsilon_details/0.05/epsilon,3.4791953936219215
|
| 578 |
+
1774094704.3854914,audit_final,414,2,audit/loss/empirical_epsilon_details/0.05/num_guesses,100.0
|
| 579 |
+
1774094704.3854914,audit_final,414,2,audit/loss/empirical_epsilon_details/0.05/correct_guesses,100.0
|
| 580 |
+
1774094704.3854914,audit_final,414,2,audit/loss/empirical_epsilon_details/0.01/epsilon,3.023197554051876
|
| 581 |
+
1774094704.3854914,audit_final,414,2,audit/loss/empirical_epsilon_details/0.01/num_guesses,100.0
|
| 582 |
+
1774094704.3854914,audit_final,414,2,audit/loss/empirical_epsilon_details/0.01/correct_guesses,100.0
|
| 583 |
+
1774094704.3854914,audit_final,414,2,audit/embedding/auc,0.916224
|
| 584 |
+
1774094704.3854914,audit_final,414,2,audit/embedding/empirical_epsilon/0.05,3.4791953936219215
|
| 585 |
+
1774094704.3854914,audit_final,414,2,audit/embedding/empirical_epsilon/0.01,3.023197554051876
|
| 586 |
+
1774094704.3854914,audit_final,414,2,audit/embedding/empirical_epsilon_details/0.05/epsilon,3.4791953936219215
|
| 587 |
+
1774094704.3854914,audit_final,414,2,audit/embedding/empirical_epsilon_details/0.05/num_guesses,100.0
|
| 588 |
+
1774094704.3854914,audit_final,414,2,audit/embedding/empirical_epsilon_details/0.05/correct_guesses,100.0
|
| 589 |
+
1774094704.3854914,audit_final,414,2,audit/embedding/empirical_epsilon_details/0.01/epsilon,3.023197554051876
|
| 590 |
+
1774094704.3854914,audit_final,414,2,audit/embedding/empirical_epsilon_details/0.01/num_guesses,100.0
|
| 591 |
+
1774094704.3854914,audit_final,414,2,audit/embedding/empirical_epsilon_details/0.01/correct_guesses,100.0
|
| 592 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/duration,3820.65162669681
|
| 593 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/emissions,0.5568149811716148
|
| 594 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/emissions_rate,0.0001457382236267942
|
| 595 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/cpu_power,179.3439937196561
|
| 596 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/gpu_power,2316.658494927477
|
| 597 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/ram_power,70.0
|
| 598 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/cpu_energy,0.18368298514637418
|
| 599 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/gpu_energy,2.457235179675422
|
| 600 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/ram_energy,0.07169108754482326
|
| 601 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/energy_consumed,2.7126092523666183
|
| 602 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/water_consumed,0.0
|
| 603 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/cpu_count,224.0
|
| 604 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/gpu_count,4.0
|
| 605 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/longitude,-121.9552
|
| 606 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/latitude,37.3541
|
| 607 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/ram_total_size,2015.5625190734863
|
| 608 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/cpu_utilization_percent,2.2497114375655825
|
| 609 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/gpu_utilization_percent,96.45330535152151
|
| 610 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/ram_utilization_percent,2.3057974816369358
|
| 611 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/ram_used_gb,46.78091819539025
|
| 612 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/pue,1.0
|
| 613 |
+
1774094704.914049,energy_final,414,,energy/codecarbon/wue,0.0
|
starcoder2-7b/base/summary.json
ADDED
|
@@ -0,0 +1,71 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"audit/delta": 1e-05,
|
| 3 |
+
"audit/embedding/auc": 0.916224,
|
| 4 |
+
"audit/embedding/empirical_epsilon/0.01": 3.023197554051876,
|
| 5 |
+
"audit/embedding/empirical_epsilon/0.05": 3.4791953936219215,
|
| 6 |
+
"audit/embedding/empirical_epsilon_details/0.01/correct_guesses": 100.0,
|
| 7 |
+
"audit/embedding/empirical_epsilon_details/0.01/epsilon": 3.023197554051876,
|
| 8 |
+
"audit/embedding/empirical_epsilon_details/0.01/num_guesses": 100.0,
|
| 9 |
+
"audit/embedding/empirical_epsilon_details/0.05/correct_guesses": 100.0,
|
| 10 |
+
"audit/embedding/empirical_epsilon_details/0.05/epsilon": 3.4791953936219215,
|
| 11 |
+
"audit/embedding/empirical_epsilon_details/0.05/num_guesses": 100.0,
|
| 12 |
+
"audit/loss/auc": 1.0,
|
| 13 |
+
"audit/loss/empirical_epsilon/0.01": 3.023197554051876,
|
| 14 |
+
"audit/loss/empirical_epsilon/0.05": 3.4791953936219215,
|
| 15 |
+
"audit/loss/empirical_epsilon_details/0.01/correct_guesses": 100.0,
|
| 16 |
+
"audit/loss/empirical_epsilon_details/0.01/epsilon": 3.023197554051876,
|
| 17 |
+
"audit/loss/empirical_epsilon_details/0.01/num_guesses": 100.0,
|
| 18 |
+
"audit/loss/empirical_epsilon_details/0.05/correct_guesses": 100.0,
|
| 19 |
+
"audit/loss/empirical_epsilon_details/0.05/epsilon": 3.4791953936219215,
|
| 20 |
+
"audit/loss/empirical_epsilon_details/0.05/num_guesses": 100.0,
|
| 21 |
+
"audit/num_canaries": 500.0,
|
| 22 |
+
"audit/num_members": 250.0,
|
| 23 |
+
"audit/paper_guess_fraction": 0.2,
|
| 24 |
+
"audit/paper_guess_steps": 20.0,
|
| 25 |
+
"energy/codecarbon/cpu_count": 224.0,
|
| 26 |
+
"energy/codecarbon/cpu_energy": 0.18368298514637418,
|
| 27 |
+
"energy/codecarbon/cpu_power": 179.3439937196561,
|
| 28 |
+
"energy/codecarbon/cpu_utilization_percent": 2.2497114375655825,
|
| 29 |
+
"energy/codecarbon/duration": 3820.65162669681,
|
| 30 |
+
"energy/codecarbon/emissions": 0.5568149811716148,
|
| 31 |
+
"energy/codecarbon/emissions_rate": 0.0001457382236267942,
|
| 32 |
+
"energy/codecarbon/energy_consumed": 2.7126092523666183,
|
| 33 |
+
"energy/codecarbon/gpu_count": 4.0,
|
| 34 |
+
"energy/codecarbon/gpu_energy": 2.457235179675422,
|
| 35 |
+
"energy/codecarbon/gpu_power": 2316.658494927477,
|
| 36 |
+
"energy/codecarbon/gpu_utilization_percent": 96.45330535152151,
|
| 37 |
+
"energy/codecarbon/latitude": 37.3541,
|
| 38 |
+
"energy/codecarbon/longitude": -121.9552,
|
| 39 |
+
"energy/codecarbon/pue": 1.0,
|
| 40 |
+
"energy/codecarbon/ram_energy": 0.07169108754482326,
|
| 41 |
+
"energy/codecarbon/ram_power": 70.0,
|
| 42 |
+
"energy/codecarbon/ram_total_size": 2015.5625190734863,
|
| 43 |
+
"energy/codecarbon/ram_used_gb": 46.78091819539025,
|
| 44 |
+
"energy/codecarbon/ram_utilization_percent": 2.3057974816369358,
|
| 45 |
+
"energy/codecarbon/water_consumed": 0.0,
|
| 46 |
+
"energy/codecarbon/wue": 0.0,
|
| 47 |
+
"eval/duration_sec": 27.13153049722314,
|
| 48 |
+
"eval/loss": 0.7454980848164394,
|
| 49 |
+
"perf/audit_duration_sec": 6.722778998315334,
|
| 50 |
+
"perf/epoch_duration_sec": 1814.1021996028721,
|
| 51 |
+
"perf/epoch_samples": 53997.0,
|
| 52 |
+
"perf/epoch_samples_per_sec": 29.765136722628178,
|
| 53 |
+
"perf/epoch_tokens": 43876231.0,
|
| 54 |
+
"perf/epoch_tokens_per_sec": 24186.19579955584,
|
| 55 |
+
"perf/gradient_accumulation_steps": 8.0,
|
| 56 |
+
"perf/logical_batch_size": 66.0,
|
| 57 |
+
"perf/logical_token_count": 55694.0,
|
| 58 |
+
"perf/samples_per_sec": 7.877740262760654,
|
| 59 |
+
"perf/step_duration_sec": 8.378037076443434,
|
| 60 |
+
"perf/tokens_per_sec": 6647.619184760483,
|
| 61 |
+
"system/cuda_epoch_peak_memory_gb": 88.38552379608154,
|
| 62 |
+
"system/cuda_max_memory_allocated_gb": 88.38552379608154,
|
| 63 |
+
"system/cuda_memory_allocated_gb": 17.816345691680908,
|
| 64 |
+
"train/epoch_canary_loss": 1.0832555509625612,
|
| 65 |
+
"train/epoch_loss": 0.840315752633972,
|
| 66 |
+
"train/epoch_real_loss": 0.8362452851041504,
|
| 67 |
+
"train/lr": 2.5558633627303928e-08,
|
| 68 |
+
"train/step_canary_loss": 0.94140625,
|
| 69 |
+
"train/step_loss": 0.854436350591255,
|
| 70 |
+
"train/step_real_loss": 0.8517185375094414
|
| 71 |
+
}
|
starcoder2-7b/base/tensorboard/events.out.tfevents.1774090870.364c4f8de9dd.6145.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:da29212bc469bf3917a2aa87b8ff1081b72b125dbbaefea65a979c0832e71373
|
| 3 |
+
size 41006
|
starcoder2-7b/base/tokenizer/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
starcoder2-7b/base/tokenizer/tokenizer_config.json
ADDED
|
@@ -0,0 +1,516 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_prefix_space": false,
|
| 3 |
+
"backend": "tokenizers",
|
| 4 |
+
"bos_token": "<|endoftext|>",
|
| 5 |
+
"clean_up_tokenization_spaces": true,
|
| 6 |
+
"eos_token": "<|endoftext|>",
|
| 7 |
+
"errors": "replace",
|
| 8 |
+
"extra_special_tokens": {
|
| 9 |
+
"extra_special_token_0": "865331112869",
|
| 10 |
+
"extra_special_token_1": "569765693871",
|
| 11 |
+
"extra_special_token_2": "485177821815",
|
| 12 |
+
"extra_special_token_3": "135441121756",
|
| 13 |
+
"extra_special_token_4": "367459894796",
|
| 14 |
+
"extra_special_token_5": "877482678543",
|
| 15 |
+
"extra_special_token_6": "457919547633",
|
| 16 |
+
"extra_special_token_7": "765474393376",
|
| 17 |
+
"extra_special_token_8": "114848338811",
|
| 18 |
+
"extra_special_token_9": "746285987371",
|
| 19 |
+
"extra_special_token_10": "649291669397",
|
| 20 |
+
"extra_special_token_11": "927914615679",
|
| 21 |
+
"extra_special_token_12": "445925149649",
|
| 22 |
+
"extra_special_token_13": "691587454538",
|
| 23 |
+
"extra_special_token_14": "143777992227",
|
| 24 |
+
"extra_special_token_15": "997981281989",
|
| 25 |
+
"extra_special_token_16": "425949483533",
|
| 26 |
+
"extra_special_token_17": "982993456429",
|
| 27 |
+
"extra_special_token_18": "718726519731",
|
| 28 |
+
"extra_special_token_19": "172599315861",
|
| 29 |
+
"extra_special_token_20": "643489267333",
|
| 30 |
+
"extra_special_token_21": "282322838685",
|
| 31 |
+
"extra_special_token_22": "781653545886",
|
| 32 |
+
"extra_special_token_23": "796415361892",
|
| 33 |
+
"extra_special_token_24": "841991688488",
|
| 34 |
+
"extra_special_token_25": "211411365397",
|
| 35 |
+
"extra_special_token_26": "698218415444",
|
| 36 |
+
"extra_special_token_27": "355977139358",
|
| 37 |
+
"extra_special_token_28": "682564697312",
|
| 38 |
+
"extra_special_token_29": "383837596997",
|
| 39 |
+
"extra_special_token_30": "689362171782",
|
| 40 |
+
"extra_special_token_31": "749966767285",
|
| 41 |
+
"extra_special_token_32": "753159165157",
|
| 42 |
+
"extra_special_token_33": "795693824762",
|
| 43 |
+
"extra_special_token_34": "669689115557",
|
| 44 |
+
"extra_special_token_35": "327491773134",
|
| 45 |
+
"extra_special_token_36": "983569279932",
|
| 46 |
+
"extra_special_token_37": "612128769512",
|
| 47 |
+
"extra_special_token_38": "374327157578",
|
| 48 |
+
"extra_special_token_39": "311632789559",
|
| 49 |
+
"extra_special_token_40": "523918658846",
|
| 50 |
+
"extra_special_token_41": "765981581453",
|
| 51 |
+
"extra_special_token_42": "794825141891",
|
| 52 |
+
"extra_special_token_43": "873898736873",
|
| 53 |
+
"extra_special_token_44": "447445629421",
|
| 54 |
+
"extra_special_token_45": "473822473819",
|
| 55 |
+
"extra_special_token_46": "181439694557",
|
| 56 |
+
"extra_special_token_47": "592538279337",
|
| 57 |
+
"extra_special_token_48": "668134915514",
|
| 58 |
+
"extra_special_token_49": "643692393748",
|
| 59 |
+
"extra_special_token_50": "696651276628",
|
| 60 |
+
"extra_special_token_51": "853859348234",
|
| 61 |
+
"extra_special_token_52": "778466723723",
|
| 62 |
+
"extra_special_token_53": "929826356991",
|
| 63 |
+
"extra_special_token_54": "272362973463",
|
| 64 |
+
"extra_special_token_55": "694235616268",
|
| 65 |
+
"extra_special_token_56": "281673864127",
|
| 66 |
+
"extra_special_token_57": "479676316326",
|
| 67 |
+
"extra_special_token_58": "646979124677",
|
| 68 |
+
"extra_special_token_59": "922327493433",
|
| 69 |
+
"extra_special_token_60": "883685933161",
|
| 70 |
+
"extra_special_token_61": "264259917554",
|
| 71 |
+
"extra_special_token_62": "836746273134",
|
| 72 |
+
"extra_special_token_63": "658481324922",
|
| 73 |
+
"extra_special_token_64": "481884157827",
|
| 74 |
+
"extra_special_token_65": "587787496812",
|
| 75 |
+
"extra_special_token_66": "579184949249",
|
| 76 |
+
"extra_special_token_67": "912193598348",
|
| 77 |
+
"extra_special_token_68": "529679678956",
|
| 78 |
+
"extra_special_token_69": "795838284624",
|
| 79 |
+
"extra_special_token_70": "159337222655",
|
| 80 |
+
"extra_special_token_71": "173781362446",
|
| 81 |
+
"extra_special_token_72": "773687856563",
|
| 82 |
+
"extra_special_token_73": "535787224917",
|
| 83 |
+
"extra_special_token_74": "351885857332",
|
| 84 |
+
"extra_special_token_75": "578827344666",
|
| 85 |
+
"extra_special_token_76": "198462689911",
|
| 86 |
+
"extra_special_token_77": "722618266242",
|
| 87 |
+
"extra_special_token_78": "952872416512",
|
| 88 |
+
"extra_special_token_79": "517778845323",
|
| 89 |
+
"extra_special_token_80": "749665846687",
|
| 90 |
+
"extra_special_token_81": "661436365453",
|
| 91 |
+
"extra_special_token_82": "259666844669",
|
| 92 |
+
"extra_special_token_83": "242851284913",
|
| 93 |
+
"extra_special_token_84": "514532995959",
|
| 94 |
+
"extra_special_token_85": "161588262349",
|
| 95 |
+
"extra_special_token_86": "742765629356",
|
| 96 |
+
"extra_special_token_87": "225164373623",
|
| 97 |
+
"extra_special_token_88": "676539973863",
|
| 98 |
+
"extra_special_token_89": "826214551218",
|
| 99 |
+
"extra_special_token_90": "182345464792",
|
| 100 |
+
"extra_special_token_91": "232776999554",
|
| 101 |
+
"extra_special_token_92": "337326533813",
|
| 102 |
+
"extra_special_token_93": "676676697292",
|
| 103 |
+
"extra_special_token_94": "929185622831",
|
| 104 |
+
"extra_special_token_95": "545512344383",
|
| 105 |
+
"extra_special_token_96": "499444466686",
|
| 106 |
+
"extra_special_token_97": "314697386682",
|
| 107 |
+
"extra_special_token_98": "517379856925",
|
| 108 |
+
"extra_special_token_99": "379557332953",
|
| 109 |
+
"extra_special_token_100": "614797267726",
|
| 110 |
+
"extra_special_token_101": "429781429464",
|
| 111 |
+
"extra_special_token_102": "922466849763",
|
| 112 |
+
"extra_special_token_103": "721737645236",
|
| 113 |
+
"extra_special_token_104": "479227349997",
|
| 114 |
+
"extra_special_token_105": "136931728327",
|
| 115 |
+
"extra_special_token_106": "259533577263",
|
| 116 |
+
"extra_special_token_107": "488538864842",
|
| 117 |
+
"extra_special_token_108": "937495658852",
|
| 118 |
+
"extra_special_token_109": "489991411364",
|
| 119 |
+
"extra_special_token_110": "499148455254",
|
| 120 |
+
"extra_special_token_111": "441373944925",
|
| 121 |
+
"extra_special_token_112": "899151413682",
|
| 122 |
+
"extra_special_token_113": "467893531755",
|
| 123 |
+
"extra_special_token_114": "527117488925",
|
| 124 |
+
"extra_special_token_115": "928335588653",
|
| 125 |
+
"extra_special_token_116": "374439448821",
|
| 126 |
+
"extra_special_token_117": "879425227932",
|
| 127 |
+
"extra_special_token_118": "867678158885",
|
| 128 |
+
"extra_special_token_119": "399749397872",
|
| 129 |
+
"extra_special_token_120": "129693547287",
|
| 130 |
+
"extra_special_token_121": "689285841825",
|
| 131 |
+
"extra_special_token_122": "771619544974",
|
| 132 |
+
"extra_special_token_123": "724883568652",
|
| 133 |
+
"extra_special_token_124": "516968424863",
|
| 134 |
+
"extra_special_token_125": "733737988257",
|
| 135 |
+
"extra_special_token_126": "852347289392",
|
| 136 |
+
"extra_special_token_127": "296953381169",
|
| 137 |
+
"extra_special_token_128": "377273562477",
|
| 138 |
+
"extra_special_token_129": "262296912232",
|
| 139 |
+
"extra_special_token_130": "547149832394",
|
| 140 |
+
"extra_special_token_131": "298464134954",
|
| 141 |
+
"extra_special_token_132": "216667245274",
|
| 142 |
+
"extra_special_token_133": "843998562287",
|
| 143 |
+
"extra_special_token_134": "572154333646",
|
| 144 |
+
"extra_special_token_135": "124589118494",
|
| 145 |
+
"extra_special_token_136": "841824384614",
|
| 146 |
+
"extra_special_token_137": "232896526252",
|
| 147 |
+
"extra_special_token_138": "295448593321",
|
| 148 |
+
"extra_special_token_139": "123741461297",
|
| 149 |
+
"extra_special_token_140": "653573457168",
|
| 150 |
+
"extra_special_token_141": "196735786156",
|
| 151 |
+
"extra_special_token_142": "377338713663",
|
| 152 |
+
"extra_special_token_143": "964342468552",
|
| 153 |
+
"extra_special_token_144": "586855179568",
|
| 154 |
+
"extra_special_token_145": "484773717614",
|
| 155 |
+
"extra_special_token_146": "894885246797",
|
| 156 |
+
"extra_special_token_147": "677896358599",
|
| 157 |
+
"extra_special_token_148": "848845611563",
|
| 158 |
+
"extra_special_token_149": "851852651677",
|
| 159 |
+
"extra_special_token_150": "398549545767",
|
| 160 |
+
"extra_special_token_151": "454244839926",
|
| 161 |
+
"extra_special_token_152": "799364566435",
|
| 162 |
+
"extra_special_token_153": "967114116556",
|
| 163 |
+
"extra_special_token_154": "817378986438",
|
| 164 |
+
"extra_special_token_155": "233795848681",
|
| 165 |
+
"extra_special_token_156": "824387273757",
|
| 166 |
+
"extra_special_token_157": "916198946615",
|
| 167 |
+
"extra_special_token_158": "563117729724",
|
| 168 |
+
"extra_special_token_159": "951794811935",
|
| 169 |
+
"extra_special_token_160": "374598961236",
|
| 170 |
+
"extra_special_token_161": "922867396683",
|
| 171 |
+
"extra_special_token_162": "765737843639",
|
| 172 |
+
"extra_special_token_163": "175469284871",
|
| 173 |
+
"extra_special_token_164": "231853711778",
|
| 174 |
+
"extra_special_token_165": "662426712668",
|
| 175 |
+
"extra_special_token_166": "711412347158",
|
| 176 |
+
"extra_special_token_167": "753466987363",
|
| 177 |
+
"extra_special_token_168": "513361312532",
|
| 178 |
+
"extra_special_token_169": "712992815957",
|
| 179 |
+
"extra_special_token_170": "971621888444",
|
| 180 |
+
"extra_special_token_171": "829235161526",
|
| 181 |
+
"extra_special_token_172": "585544633356",
|
| 182 |
+
"extra_special_token_173": "582471228164",
|
| 183 |
+
"extra_special_token_174": "678666359123",
|
| 184 |
+
"extra_special_token_175": "557533689478",
|
| 185 |
+
"extra_special_token_176": "632962475133",
|
| 186 |
+
"extra_special_token_177": "484489193824",
|
| 187 |
+
"extra_special_token_178": "489562189822",
|
| 188 |
+
"extra_special_token_179": "589547936288",
|
| 189 |
+
"extra_special_token_180": "363214487524",
|
| 190 |
+
"extra_special_token_181": "244885399387",
|
| 191 |
+
"extra_special_token_182": "431751228368",
|
| 192 |
+
"extra_special_token_183": "433581868192",
|
| 193 |
+
"extra_special_token_184": "486391569221",
|
| 194 |
+
"extra_special_token_185": "185438575221",
|
| 195 |
+
"extra_special_token_186": "126574388585",
|
| 196 |
+
"extra_special_token_187": "741757479784",
|
| 197 |
+
"extra_special_token_188": "529854679937",
|
| 198 |
+
"extra_special_token_189": "996116119839",
|
| 199 |
+
"extra_special_token_190": "616248973917",
|
| 200 |
+
"extra_special_token_191": "763531783491",
|
| 201 |
+
"extra_special_token_192": "955456118295",
|
| 202 |
+
"extra_special_token_193": "364196983365",
|
| 203 |
+
"extra_special_token_194": "195792996468",
|
| 204 |
+
"extra_special_token_195": "151859598873",
|
| 205 |
+
"extra_special_token_196": "399223169721",
|
| 206 |
+
"extra_special_token_197": "938488813964",
|
| 207 |
+
"extra_special_token_198": "961981959227",
|
| 208 |
+
"extra_special_token_199": "183368827562",
|
| 209 |
+
"extra_special_token_200": "533417736566",
|
| 210 |
+
"extra_special_token_201": "786391632558",
|
| 211 |
+
"extra_special_token_202": "665661658354",
|
| 212 |
+
"extra_special_token_203": "693281533643",
|
| 213 |
+
"extra_special_token_204": "475794684356",
|
| 214 |
+
"extra_special_token_205": "652154162978",
|
| 215 |
+
"extra_special_token_206": "753233719644",
|
| 216 |
+
"extra_special_token_207": "668514843129",
|
| 217 |
+
"extra_special_token_208": "819162623892",
|
| 218 |
+
"extra_special_token_209": "941169431859",
|
| 219 |
+
"extra_special_token_210": "877385381798",
|
| 220 |
+
"extra_special_token_211": "752644929761",
|
| 221 |
+
"extra_special_token_212": "881136466196",
|
| 222 |
+
"extra_special_token_213": "275597777299",
|
| 223 |
+
"extra_special_token_214": "731681792655",
|
| 224 |
+
"extra_special_token_215": "961133895172",
|
| 225 |
+
"extra_special_token_216": "864718285734",
|
| 226 |
+
"extra_special_token_217": "963852916563",
|
| 227 |
+
"extra_special_token_218": "319584985416",
|
| 228 |
+
"extra_special_token_219": "563365646341",
|
| 229 |
+
"extra_special_token_220": "811371928234",
|
| 230 |
+
"extra_special_token_221": "837131396371",
|
| 231 |
+
"extra_special_token_222": "267514771964",
|
| 232 |
+
"extra_special_token_223": "944513428457",
|
| 233 |
+
"extra_special_token_224": "117298239631",
|
| 234 |
+
"extra_special_token_225": "158142752582",
|
| 235 |
+
"extra_special_token_226": "252867443568",
|
| 236 |
+
"extra_special_token_227": "839269684865",
|
| 237 |
+
"extra_special_token_228": "612788593128",
|
| 238 |
+
"extra_special_token_229": "145669731981",
|
| 239 |
+
"extra_special_token_230": "121557291859",
|
| 240 |
+
"extra_special_token_231": "245416776926",
|
| 241 |
+
"extra_special_token_232": "799417897197",
|
| 242 |
+
"extra_special_token_233": "997958836435",
|
| 243 |
+
"extra_special_token_234": "892336777248",
|
| 244 |
+
"extra_special_token_235": "158929292238",
|
| 245 |
+
"extra_special_token_236": "581976444672",
|
| 246 |
+
"extra_special_token_237": "897784492783",
|
| 247 |
+
"extra_special_token_238": "492373714791",
|
| 248 |
+
"extra_special_token_239": "512659818733",
|
| 249 |
+
"extra_special_token_240": "881112998642",
|
| 250 |
+
"extra_special_token_241": "619454958782",
|
| 251 |
+
"extra_special_token_242": "431149748713",
|
| 252 |
+
"extra_special_token_243": "624221476921",
|
| 253 |
+
"extra_special_token_244": "125866399464",
|
| 254 |
+
"extra_special_token_245": "339882449689",
|
| 255 |
+
"extra_special_token_246": "186198784585",
|
| 256 |
+
"extra_special_token_247": "943193294691",
|
| 257 |
+
"extra_special_token_248": "955668961269",
|
| 258 |
+
"extra_special_token_249": "232787996724",
|
| 259 |
+
"extra_special_token_250": "215671314196",
|
| 260 |
+
"extra_special_token_251": "286173241916",
|
| 261 |
+
"extra_special_token_252": "745977673725",
|
| 262 |
+
"extra_special_token_253": "556976448182",
|
| 263 |
+
"extra_special_token_254": "599961512792",
|
| 264 |
+
"extra_special_token_255": "766294538337",
|
| 265 |
+
"extra_special_token_256": "934912591213",
|
| 266 |
+
"extra_special_token_257": "295118729589",
|
| 267 |
+
"extra_special_token_258": "529455466433",
|
| 268 |
+
"extra_special_token_259": "196119929397",
|
| 269 |
+
"extra_special_token_260": "379571934299",
|
| 270 |
+
"extra_special_token_261": "251789649997",
|
| 271 |
+
"extra_special_token_262": "564544131355",
|
| 272 |
+
"extra_special_token_263": "244371196654",
|
| 273 |
+
"extra_special_token_264": "384598329253",
|
| 274 |
+
"extra_special_token_265": "887753195844",
|
| 275 |
+
"extra_special_token_266": "364947325679",
|
| 276 |
+
"extra_special_token_267": "655517954651",
|
| 277 |
+
"extra_special_token_268": "673948786567",
|
| 278 |
+
"extra_special_token_269": "857231548835",
|
| 279 |
+
"extra_special_token_270": "816115936673",
|
| 280 |
+
"extra_special_token_271": "644234165531",
|
| 281 |
+
"extra_special_token_272": "182782912224",
|
| 282 |
+
"extra_special_token_273": "234316622259",
|
| 283 |
+
"extra_special_token_274": "421369185549",
|
| 284 |
+
"extra_special_token_275": "434632855397",
|
| 285 |
+
"extra_special_token_276": "921889371893",
|
| 286 |
+
"extra_special_token_277": "415956914763",
|
| 287 |
+
"extra_special_token_278": "598916996413",
|
| 288 |
+
"extra_special_token_279": "773671349113",
|
| 289 |
+
"extra_special_token_280": "952465217972",
|
| 290 |
+
"extra_special_token_281": "117657531962",
|
| 291 |
+
"extra_special_token_282": "729825168745",
|
| 292 |
+
"extra_special_token_283": "691315125346",
|
| 293 |
+
"extra_special_token_284": "768461952319",
|
| 294 |
+
"extra_special_token_285": "664847713559",
|
| 295 |
+
"extra_special_token_286": "953267689786",
|
| 296 |
+
"extra_special_token_287": "886464195129",
|
| 297 |
+
"extra_special_token_288": "824488329416",
|
| 298 |
+
"extra_special_token_289": "837873762491",
|
| 299 |
+
"extra_special_token_290": "532833541879",
|
| 300 |
+
"extra_special_token_291": "669183782449",
|
| 301 |
+
"extra_special_token_292": "941976537588",
|
| 302 |
+
"extra_special_token_293": "739394546916",
|
| 303 |
+
"extra_special_token_294": "267954879268",
|
| 304 |
+
"extra_special_token_295": "637551427887",
|
| 305 |
+
"extra_special_token_296": "217756494954",
|
| 306 |
+
"extra_special_token_297": "524444658383",
|
| 307 |
+
"extra_special_token_298": "117783274348",
|
| 308 |
+
"extra_special_token_299": "138218735276",
|
| 309 |
+
"extra_special_token_300": "814611949491",
|
| 310 |
+
"extra_special_token_301": "711641973413",
|
| 311 |
+
"extra_special_token_302": "499156317423",
|
| 312 |
+
"extra_special_token_303": "515856611931",
|
| 313 |
+
"extra_special_token_304": "454164859837",
|
| 314 |
+
"extra_special_token_305": "345271433112",
|
| 315 |
+
"extra_special_token_306": "462294118988",
|
| 316 |
+
"extra_special_token_307": "511785788222",
|
| 317 |
+
"extra_special_token_308": "497294727353",
|
| 318 |
+
"extra_special_token_309": "866519986723",
|
| 319 |
+
"extra_special_token_310": "334513529294",
|
| 320 |
+
"extra_special_token_311": "549946382131",
|
| 321 |
+
"extra_special_token_312": "284445431422",
|
| 322 |
+
"extra_special_token_313": "396521188476",
|
| 323 |
+
"extra_special_token_314": "421435255895",
|
| 324 |
+
"extra_special_token_315": "133373659361",
|
| 325 |
+
"extra_special_token_316": "322683334381",
|
| 326 |
+
"extra_special_token_317": "228358422847",
|
| 327 |
+
"extra_special_token_318": "291762694874",
|
| 328 |
+
"extra_special_token_319": "143182978129",
|
| 329 |
+
"extra_special_token_320": "511923256573",
|
| 330 |
+
"extra_special_token_321": "327158398268",
|
| 331 |
+
"extra_special_token_322": "879764613759",
|
| 332 |
+
"extra_special_token_323": "564395222747",
|
| 333 |
+
"extra_special_token_324": "451161679736",
|
| 334 |
+
"extra_special_token_325": "538631466654",
|
| 335 |
+
"extra_special_token_326": "221762325616",
|
| 336 |
+
"extra_special_token_327": "218391991184",
|
| 337 |
+
"extra_special_token_328": "322589379462",
|
| 338 |
+
"extra_special_token_329": "876537814263",
|
| 339 |
+
"extra_special_token_330": "152676556624",
|
| 340 |
+
"extra_special_token_331": "332522971941",
|
| 341 |
+
"extra_special_token_332": "884354318946",
|
| 342 |
+
"extra_special_token_333": "513349618943",
|
| 343 |
+
"extra_special_token_334": "116639746413",
|
| 344 |
+
"extra_special_token_335": "635185846287",
|
| 345 |
+
"extra_special_token_336": "993832498489",
|
| 346 |
+
"extra_special_token_337": "813981174797",
|
| 347 |
+
"extra_special_token_338": "438745114173",
|
| 348 |
+
"extra_special_token_339": "983493951323",
|
| 349 |
+
"extra_special_token_340": "724492262421",
|
| 350 |
+
"extra_special_token_341": "622553389126",
|
| 351 |
+
"extra_special_token_342": "889965243135",
|
| 352 |
+
"extra_special_token_343": "364492359246",
|
| 353 |
+
"extra_special_token_344": "154962668224",
|
| 354 |
+
"extra_special_token_345": "179564995814",
|
| 355 |
+
"extra_special_token_346": "418412875665",
|
| 356 |
+
"extra_special_token_347": "718951851413",
|
| 357 |
+
"extra_special_token_348": "699446724178",
|
| 358 |
+
"extra_special_token_349": "624266421831",
|
| 359 |
+
"extra_special_token_350": "815458725125",
|
| 360 |
+
"extra_special_token_351": "455423278865",
|
| 361 |
+
"extra_special_token_352": "393741199486",
|
| 362 |
+
"extra_special_token_353": "328552864359",
|
| 363 |
+
"extra_special_token_354": "211662639865",
|
| 364 |
+
"extra_special_token_355": "218784516525",
|
| 365 |
+
"extra_special_token_356": "762486672996",
|
| 366 |
+
"extra_special_token_357": "142799718159",
|
| 367 |
+
"extra_special_token_358": "858146415154",
|
| 368 |
+
"extra_special_token_359": "767858144912",
|
| 369 |
+
"extra_special_token_360": "571317457151",
|
| 370 |
+
"extra_special_token_361": "635127952696",
|
| 371 |
+
"extra_special_token_362": "116427191984",
|
| 372 |
+
"extra_special_token_363": "268921994538",
|
| 373 |
+
"extra_special_token_364": "523937669294",
|
| 374 |
+
"extra_special_token_365": "165429152138",
|
| 375 |
+
"extra_special_token_366": "739246183345",
|
| 376 |
+
"extra_special_token_367": "591464355756",
|
| 377 |
+
"extra_special_token_368": "212985874612",
|
| 378 |
+
"extra_special_token_369": "191887635211",
|
| 379 |
+
"extra_special_token_370": "967214577653",
|
| 380 |
+
"extra_special_token_371": "119342152414",
|
| 381 |
+
"extra_special_token_372": "946444632795",
|
| 382 |
+
"extra_special_token_373": "618423867817",
|
| 383 |
+
"extra_special_token_374": "228565148417",
|
| 384 |
+
"extra_special_token_375": "729116422489",
|
| 385 |
+
"extra_special_token_376": "527874729936",
|
| 386 |
+
"extra_special_token_377": "739784153482",
|
| 387 |
+
"extra_special_token_378": "387763951128",
|
| 388 |
+
"extra_special_token_379": "331369926711",
|
| 389 |
+
"extra_special_token_380": "562716493614",
|
| 390 |
+
"extra_special_token_381": "739667844957",
|
| 391 |
+
"extra_special_token_382": "562389434565",
|
| 392 |
+
"extra_special_token_383": "256497188281",
|
| 393 |
+
"extra_special_token_384": "859927364588",
|
| 394 |
+
"extra_special_token_385": "417668946583",
|
| 395 |
+
"extra_special_token_386": "357621613582",
|
| 396 |
+
"extra_special_token_387": "438435178228",
|
| 397 |
+
"extra_special_token_388": "485692541169",
|
| 398 |
+
"extra_special_token_389": "825815739116",
|
| 399 |
+
"extra_special_token_390": "342221452223",
|
| 400 |
+
"extra_special_token_391": "697747991249",
|
| 401 |
+
"extra_special_token_392": "716763689965",
|
| 402 |
+
"extra_special_token_393": "141499982867",
|
| 403 |
+
"extra_special_token_394": "818479319499",
|
| 404 |
+
"extra_special_token_395": "336813343298",
|
| 405 |
+
"extra_special_token_396": "594688742928",
|
| 406 |
+
"extra_special_token_397": "472129283475",
|
| 407 |
+
"extra_special_token_398": "514354144759",
|
| 408 |
+
"extra_special_token_399": "349249721685",
|
| 409 |
+
"extra_special_token_400": "546276298359",
|
| 410 |
+
"extra_special_token_401": "353755529131",
|
| 411 |
+
"extra_special_token_402": "315534574435",
|
| 412 |
+
"extra_special_token_403": "523723475786",
|
| 413 |
+
"extra_special_token_404": "215826764872",
|
| 414 |
+
"extra_special_token_405": "367968398551",
|
| 415 |
+
"extra_special_token_406": "569853653352",
|
| 416 |
+
"extra_special_token_407": "389715484387",
|
| 417 |
+
"extra_special_token_408": "293847485454",
|
| 418 |
+
"extra_special_token_409": "714738141818",
|
| 419 |
+
"extra_special_token_410": "178478368922",
|
| 420 |
+
"extra_special_token_411": "581493616981",
|
| 421 |
+
"extra_special_token_412": "589439538674",
|
| 422 |
+
"extra_special_token_413": "846657726193",
|
| 423 |
+
"extra_special_token_414": "722339992679",
|
| 424 |
+
"extra_special_token_415": "138154781148",
|
| 425 |
+
"extra_special_token_416": "757785319772",
|
| 426 |
+
"extra_special_token_417": "492516914298",
|
| 427 |
+
"extra_special_token_418": "919181521716",
|
| 428 |
+
"extra_special_token_419": "985781138935",
|
| 429 |
+
"extra_special_token_420": "476969195485",
|
| 430 |
+
"extra_special_token_421": "313145133463",
|
| 431 |
+
"extra_special_token_422": "758963111966",
|
| 432 |
+
"extra_special_token_423": "147541537162",
|
| 433 |
+
"extra_special_token_424": "557163366873",
|
| 434 |
+
"extra_special_token_425": "144373897488",
|
| 435 |
+
"extra_special_token_426": "522515164754",
|
| 436 |
+
"extra_special_token_427": "724964923582",
|
| 437 |
+
"extra_special_token_428": "284776712475",
|
| 438 |
+
"extra_special_token_429": "375429755114",
|
| 439 |
+
"extra_special_token_430": "181233596124",
|
| 440 |
+
"extra_special_token_431": "948585673431",
|
| 441 |
+
"extra_special_token_432": "243165586174",
|
| 442 |
+
"extra_special_token_433": "396847976144",
|
| 443 |
+
"extra_special_token_434": "997724962668",
|
| 444 |
+
"extra_special_token_435": "558837194455",
|
| 445 |
+
"extra_special_token_436": "163165456396",
|
| 446 |
+
"extra_special_token_437": "378749551722",
|
| 447 |
+
"extra_special_token_438": "161238482259",
|
| 448 |
+
"extra_special_token_439": "754978243758",
|
| 449 |
+
"extra_special_token_440": "195388849133",
|
| 450 |
+
"extra_special_token_441": "229775525672",
|
| 451 |
+
"extra_special_token_442": "262437452884",
|
| 452 |
+
"extra_special_token_443": "441377892146",
|
| 453 |
+
"extra_special_token_444": "451885565366",
|
| 454 |
+
"extra_special_token_445": "981277526855",
|
| 455 |
+
"extra_special_token_446": "762495822823",
|
| 456 |
+
"extra_special_token_447": "368763327262",
|
| 457 |
+
"extra_special_token_448": "757422791351",
|
| 458 |
+
"extra_special_token_449": "636324136426",
|
| 459 |
+
"extra_special_token_450": "214193645583",
|
| 460 |
+
"extra_special_token_451": "412843856172",
|
| 461 |
+
"extra_special_token_452": "179386156569",
|
| 462 |
+
"extra_special_token_453": "756916173536",
|
| 463 |
+
"extra_special_token_454": "892697125149",
|
| 464 |
+
"extra_special_token_455": "625334487352",
|
| 465 |
+
"extra_special_token_456": "941861857715",
|
| 466 |
+
"extra_special_token_457": "887417525236",
|
| 467 |
+
"extra_special_token_458": "649516938598",
|
| 468 |
+
"extra_special_token_459": "717628619782",
|
| 469 |
+
"extra_special_token_460": "438124184139",
|
| 470 |
+
"extra_special_token_461": "547563892268",
|
| 471 |
+
"extra_special_token_462": "856317483891",
|
| 472 |
+
"extra_special_token_463": "313313831273",
|
| 473 |
+
"extra_special_token_464": "371496153876",
|
| 474 |
+
"extra_special_token_465": "587541149322",
|
| 475 |
+
"extra_special_token_466": "265847332563",
|
| 476 |
+
"extra_special_token_467": "449549215429",
|
| 477 |
+
"extra_special_token_468": "163497196769",
|
| 478 |
+
"extra_special_token_469": "861342291298",
|
| 479 |
+
"extra_special_token_470": "268433315926",
|
| 480 |
+
"extra_special_token_471": "774679513717",
|
| 481 |
+
"extra_special_token_472": "851254219729",
|
| 482 |
+
"extra_special_token_473": "583527834464",
|
| 483 |
+
"extra_special_token_474": "488496781997",
|
| 484 |
+
"extra_special_token_475": "556814553861",
|
| 485 |
+
"extra_special_token_476": "482829231639",
|
| 486 |
+
"extra_special_token_477": "618878266619",
|
| 487 |
+
"extra_special_token_478": "147444452794",
|
| 488 |
+
"extra_special_token_479": "949235426629",
|
| 489 |
+
"extra_special_token_480": "357299947518",
|
| 490 |
+
"extra_special_token_481": "175528632226",
|
| 491 |
+
"extra_special_token_482": "645527857972",
|
| 492 |
+
"extra_special_token_483": "186872457894",
|
| 493 |
+
"extra_special_token_484": "552738847828",
|
| 494 |
+
"extra_special_token_485": "626748382482",
|
| 495 |
+
"extra_special_token_486": "921894985642",
|
| 496 |
+
"extra_special_token_487": "943878645871",
|
| 497 |
+
"extra_special_token_488": "859289776479",
|
| 498 |
+
"extra_special_token_489": "614583493135",
|
| 499 |
+
"extra_special_token_490": "933775286797",
|
| 500 |
+
"extra_special_token_491": "332234613346",
|
| 501 |
+
"extra_special_token_492": "325196781219",
|
| 502 |
+
"extra_special_token_493": "142526557681",
|
| 503 |
+
"extra_special_token_494": "356722692178",
|
| 504 |
+
"extra_special_token_495": "449318681694",
|
| 505 |
+
"extra_special_token_496": "687284547244",
|
| 506 |
+
"extra_special_token_497": "947262995132",
|
| 507 |
+
"extra_special_token_498": "893974619684",
|
| 508 |
+
"extra_special_token_499": "797238311233"
|
| 509 |
+
},
|
| 510 |
+
"is_local": false,
|
| 511 |
+
"model_max_length": 1000000000000000019884624838656,
|
| 512 |
+
"pad_token": "<|endoftext|>",
|
| 513 |
+
"tokenizer_class": "GPT2Tokenizer",
|
| 514 |
+
"unk_token": "<|endoftext|>",
|
| 515 |
+
"vocab_size": 49152
|
| 516 |
+
}
|
starcoder2-7b/base/train.log
ADDED
|
@@ -0,0 +1,49 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2026-03-21 11:04:31,986 [INFO] new_opacus_codex.train_steps: epoch=1 step=10 loss=1.5025
|
| 2 |
+
2026-03-21 11:05:54,380 [INFO] new_opacus_codex.train_steps: epoch=1 step=20 loss=1.3712
|
| 3 |
+
2026-03-21 11:07:15,652 [INFO] new_opacus_codex.train_steps: epoch=1 step=30 loss=1.1814
|
| 4 |
+
2026-03-21 11:08:38,661 [INFO] new_opacus_codex.train_steps: epoch=1 step=40 loss=1.1834
|
| 5 |
+
2026-03-21 11:10:02,035 [INFO] new_opacus_codex.train_steps: epoch=1 step=50 loss=1.1336
|
| 6 |
+
2026-03-21 11:10:29,195 [INFO] new_opacus_codex.train_steps: eval event=eval_step epoch=1 step=50 eval_loss=0.8034 duration_sec=27.16
|
| 7 |
+
2026-03-21 11:11:51,908 [INFO] new_opacus_codex.train_steps: epoch=1 step=60 loss=1.1219
|
| 8 |
+
2026-03-21 11:13:13,533 [INFO] new_opacus_codex.train_steps: epoch=1 step=70 loss=1.0407
|
| 9 |
+
2026-03-21 11:14:36,548 [INFO] new_opacus_codex.train_steps: epoch=1 step=80 loss=1.0129
|
| 10 |
+
2026-03-21 11:15:57,878 [INFO] new_opacus_codex.train_steps: epoch=1 step=90 loss=1.0161
|
| 11 |
+
2026-03-21 11:17:19,691 [INFO] new_opacus_codex.train_steps: epoch=1 step=100 loss=1.0200
|
| 12 |
+
2026-03-21 11:17:46,876 [INFO] new_opacus_codex.train_steps: eval event=eval_step epoch=1 step=100 eval_loss=0.7755 duration_sec=27.18
|
| 13 |
+
2026-03-21 11:19:10,166 [INFO] new_opacus_codex.train_steps: epoch=1 step=110 loss=1.0177
|
| 14 |
+
2026-03-21 11:20:31,874 [INFO] new_opacus_codex.train_steps: epoch=1 step=120 loss=0.9631
|
| 15 |
+
2026-03-21 11:21:54,417 [INFO] new_opacus_codex.train_steps: epoch=1 step=130 loss=0.9569
|
| 16 |
+
2026-03-21 11:23:16,389 [INFO] new_opacus_codex.train_steps: epoch=1 step=140 loss=0.9588
|
| 17 |
+
2026-03-21 11:24:38,982 [INFO] new_opacus_codex.train_steps: epoch=1 step=150 loss=0.9273
|
| 18 |
+
2026-03-21 11:25:06,151 [INFO] new_opacus_codex.train_steps: eval event=eval_step epoch=1 step=150 eval_loss=0.7622 duration_sec=27.17
|
| 19 |
+
2026-03-21 11:26:28,134 [INFO] new_opacus_codex.train_steps: epoch=1 step=160 loss=0.9226
|
| 20 |
+
2026-03-21 11:27:49,927 [INFO] new_opacus_codex.train_steps: epoch=1 step=170 loss=0.9244
|
| 21 |
+
2026-03-21 11:29:11,942 [INFO] new_opacus_codex.train_steps: epoch=1 step=180 loss=0.9250
|
| 22 |
+
2026-03-21 11:30:34,185 [INFO] new_opacus_codex.train_steps: epoch=1 step=190 loss=0.8691
|
| 23 |
+
2026-03-21 11:31:57,690 [INFO] new_opacus_codex.train_steps: epoch=1 step=200 loss=0.8916
|
| 24 |
+
2026-03-21 11:32:24,879 [INFO] new_opacus_codex.train_steps: eval event=eval_step epoch=1 step=200 eval_loss=0.7525 duration_sec=27.19
|
| 25 |
+
2026-03-21 11:34:26,079 [INFO] new_opacus_codex.train_steps: epoch=2 step=210 loss=0.9168
|
| 26 |
+
2026-03-21 11:35:47,787 [INFO] new_opacus_codex.train_steps: epoch=2 step=220 loss=0.8558
|
| 27 |
+
2026-03-21 11:37:09,366 [INFO] new_opacus_codex.train_steps: epoch=2 step=230 loss=0.8524
|
| 28 |
+
2026-03-21 11:38:31,484 [INFO] new_opacus_codex.train_steps: epoch=2 step=240 loss=0.8580
|
| 29 |
+
2026-03-21 11:39:52,882 [INFO] new_opacus_codex.train_steps: epoch=2 step=250 loss=0.8491
|
| 30 |
+
2026-03-21 11:40:20,073 [INFO] new_opacus_codex.train_steps: eval event=eval_step epoch=2 step=250 eval_loss=0.7485 duration_sec=27.19
|
| 31 |
+
2026-03-21 11:41:42,276 [INFO] new_opacus_codex.train_steps: epoch=2 step=260 loss=0.8442
|
| 32 |
+
2026-03-21 11:43:04,391 [INFO] new_opacus_codex.train_steps: epoch=2 step=270 loss=0.8571
|
| 33 |
+
2026-03-21 11:44:28,058 [INFO] new_opacus_codex.train_steps: epoch=2 step=280 loss=0.8479
|
| 34 |
+
2026-03-21 11:45:50,712 [INFO] new_opacus_codex.train_steps: epoch=2 step=290 loss=0.8434
|
| 35 |
+
2026-03-21 11:47:12,551 [INFO] new_opacus_codex.train_steps: epoch=2 step=300 loss=0.8405
|
| 36 |
+
2026-03-21 11:47:39,718 [INFO] new_opacus_codex.train_steps: eval event=eval_step epoch=2 step=300 eval_loss=0.7462 duration_sec=27.16
|
| 37 |
+
2026-03-21 11:49:02,560 [INFO] new_opacus_codex.train_steps: epoch=2 step=310 loss=0.8507
|
| 38 |
+
2026-03-21 11:50:24,762 [INFO] new_opacus_codex.train_steps: epoch=2 step=320 loss=0.8479
|
| 39 |
+
2026-03-21 11:51:47,919 [INFO] new_opacus_codex.train_steps: epoch=2 step=330 loss=0.8466
|
| 40 |
+
2026-03-21 11:53:09,723 [INFO] new_opacus_codex.train_steps: epoch=2 step=340 loss=0.8369
|
| 41 |
+
2026-03-21 11:54:31,911 [INFO] new_opacus_codex.train_steps: epoch=2 step=350 loss=0.8474
|
| 42 |
+
2026-03-21 11:54:59,091 [INFO] new_opacus_codex.train_steps: eval event=eval_step epoch=2 step=350 eval_loss=0.7456 duration_sec=27.18
|
| 43 |
+
2026-03-21 11:56:21,565 [INFO] new_opacus_codex.train_steps: epoch=2 step=360 loss=0.8568
|
| 44 |
+
2026-03-21 11:57:43,511 [INFO] new_opacus_codex.train_steps: epoch=2 step=370 loss=0.8227
|
| 45 |
+
2026-03-21 11:59:06,162 [INFO] new_opacus_codex.train_steps: epoch=2 step=380 loss=0.8342
|
| 46 |
+
2026-03-21 12:00:29,231 [INFO] new_opacus_codex.train_steps: epoch=2 step=390 loss=0.8265
|
| 47 |
+
2026-03-21 12:01:52,289 [INFO] new_opacus_codex.train_steps: epoch=2 step=400 loss=0.8197
|
| 48 |
+
2026-03-21 12:02:19,482 [INFO] new_opacus_codex.train_steps: eval event=eval_step epoch=2 step=400 eval_loss=0.7455 duration_sec=27.19
|
| 49 |
+
2026-03-21 12:03:41,445 [INFO] new_opacus_codex.train_steps: epoch=2 step=410 loss=0.8286
|
starcoder2-7b/dp3/adapter/README.md
ADDED
|
@@ -0,0 +1,207 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
base_model: bigcode/starcoder2-7b
|
| 3 |
+
library_name: peft
|
| 4 |
+
pipeline_tag: text-generation
|
| 5 |
+
tags:
|
| 6 |
+
- base_model:adapter:bigcode/starcoder2-7b
|
| 7 |
+
- lora
|
| 8 |
+
- transformers
|
| 9 |
+
---
|
| 10 |
+
|
| 11 |
+
# Model Card for Model ID
|
| 12 |
+
|
| 13 |
+
<!-- Provide a quick summary of what the model is/does. -->
|
| 14 |
+
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
## Model Details
|
| 18 |
+
|
| 19 |
+
### Model Description
|
| 20 |
+
|
| 21 |
+
<!-- Provide a longer summary of what this model is. -->
|
| 22 |
+
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
- **Developed by:** [More Information Needed]
|
| 26 |
+
- **Funded by [optional]:** [More Information Needed]
|
| 27 |
+
- **Shared by [optional]:** [More Information Needed]
|
| 28 |
+
- **Model type:** [More Information Needed]
|
| 29 |
+
- **Language(s) (NLP):** [More Information Needed]
|
| 30 |
+
- **License:** [More Information Needed]
|
| 31 |
+
- **Finetuned from model [optional]:** [More Information Needed]
|
| 32 |
+
|
| 33 |
+
### Model Sources [optional]
|
| 34 |
+
|
| 35 |
+
<!-- Provide the basic links for the model. -->
|
| 36 |
+
|
| 37 |
+
- **Repository:** [More Information Needed]
|
| 38 |
+
- **Paper [optional]:** [More Information Needed]
|
| 39 |
+
- **Demo [optional]:** [More Information Needed]
|
| 40 |
+
|
| 41 |
+
## Uses
|
| 42 |
+
|
| 43 |
+
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
| 44 |
+
|
| 45 |
+
### Direct Use
|
| 46 |
+
|
| 47 |
+
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
| 48 |
+
|
| 49 |
+
[More Information Needed]
|
| 50 |
+
|
| 51 |
+
### Downstream Use [optional]
|
| 52 |
+
|
| 53 |
+
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
|
| 54 |
+
|
| 55 |
+
[More Information Needed]
|
| 56 |
+
|
| 57 |
+
### Out-of-Scope Use
|
| 58 |
+
|
| 59 |
+
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
|
| 60 |
+
|
| 61 |
+
[More Information Needed]
|
| 62 |
+
|
| 63 |
+
## Bias, Risks, and Limitations
|
| 64 |
+
|
| 65 |
+
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
|
| 66 |
+
|
| 67 |
+
[More Information Needed]
|
| 68 |
+
|
| 69 |
+
### Recommendations
|
| 70 |
+
|
| 71 |
+
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
|
| 72 |
+
|
| 73 |
+
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
|
| 74 |
+
|
| 75 |
+
## How to Get Started with the Model
|
| 76 |
+
|
| 77 |
+
Use the code below to get started with the model.
|
| 78 |
+
|
| 79 |
+
[More Information Needed]
|
| 80 |
+
|
| 81 |
+
## Training Details
|
| 82 |
+
|
| 83 |
+
### Training Data
|
| 84 |
+
|
| 85 |
+
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
| 86 |
+
|
| 87 |
+
[More Information Needed]
|
| 88 |
+
|
| 89 |
+
### Training Procedure
|
| 90 |
+
|
| 91 |
+
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
|
| 92 |
+
|
| 93 |
+
#### Preprocessing [optional]
|
| 94 |
+
|
| 95 |
+
[More Information Needed]
|
| 96 |
+
|
| 97 |
+
|
| 98 |
+
#### Training Hyperparameters
|
| 99 |
+
|
| 100 |
+
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
| 101 |
+
|
| 102 |
+
#### Speeds, Sizes, Times [optional]
|
| 103 |
+
|
| 104 |
+
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|
| 105 |
+
|
| 106 |
+
[More Information Needed]
|
| 107 |
+
|
| 108 |
+
## Evaluation
|
| 109 |
+
|
| 110 |
+
<!-- This section describes the evaluation protocols and provides the results. -->
|
| 111 |
+
|
| 112 |
+
### Testing Data, Factors & Metrics
|
| 113 |
+
|
| 114 |
+
#### Testing Data
|
| 115 |
+
|
| 116 |
+
<!-- This should link to a Dataset Card if possible. -->
|
| 117 |
+
|
| 118 |
+
[More Information Needed]
|
| 119 |
+
|
| 120 |
+
#### Factors
|
| 121 |
+
|
| 122 |
+
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
|
| 123 |
+
|
| 124 |
+
[More Information Needed]
|
| 125 |
+
|
| 126 |
+
#### Metrics
|
| 127 |
+
|
| 128 |
+
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
|
| 129 |
+
|
| 130 |
+
[More Information Needed]
|
| 131 |
+
|
| 132 |
+
### Results
|
| 133 |
+
|
| 134 |
+
[More Information Needed]
|
| 135 |
+
|
| 136 |
+
#### Summary
|
| 137 |
+
|
| 138 |
+
|
| 139 |
+
|
| 140 |
+
## Model Examination [optional]
|
| 141 |
+
|
| 142 |
+
<!-- Relevant interpretability work for the model goes here -->
|
| 143 |
+
|
| 144 |
+
[More Information Needed]
|
| 145 |
+
|
| 146 |
+
## Environmental Impact
|
| 147 |
+
|
| 148 |
+
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
|
| 149 |
+
|
| 150 |
+
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
| 151 |
+
|
| 152 |
+
- **Hardware Type:** [More Information Needed]
|
| 153 |
+
- **Hours used:** [More Information Needed]
|
| 154 |
+
- **Cloud Provider:** [More Information Needed]
|
| 155 |
+
- **Compute Region:** [More Information Needed]
|
| 156 |
+
- **Carbon Emitted:** [More Information Needed]
|
| 157 |
+
|
| 158 |
+
## Technical Specifications [optional]
|
| 159 |
+
|
| 160 |
+
### Model Architecture and Objective
|
| 161 |
+
|
| 162 |
+
[More Information Needed]
|
| 163 |
+
|
| 164 |
+
### Compute Infrastructure
|
| 165 |
+
|
| 166 |
+
[More Information Needed]
|
| 167 |
+
|
| 168 |
+
#### Hardware
|
| 169 |
+
|
| 170 |
+
[More Information Needed]
|
| 171 |
+
|
| 172 |
+
#### Software
|
| 173 |
+
|
| 174 |
+
[More Information Needed]
|
| 175 |
+
|
| 176 |
+
## Citation [optional]
|
| 177 |
+
|
| 178 |
+
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
| 179 |
+
|
| 180 |
+
**BibTeX:**
|
| 181 |
+
|
| 182 |
+
[More Information Needed]
|
| 183 |
+
|
| 184 |
+
**APA:**
|
| 185 |
+
|
| 186 |
+
[More Information Needed]
|
| 187 |
+
|
| 188 |
+
## Glossary [optional]
|
| 189 |
+
|
| 190 |
+
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
|
| 191 |
+
|
| 192 |
+
[More Information Needed]
|
| 193 |
+
|
| 194 |
+
## More Information [optional]
|
| 195 |
+
|
| 196 |
+
[More Information Needed]
|
| 197 |
+
|
| 198 |
+
## Model Card Authors [optional]
|
| 199 |
+
|
| 200 |
+
[More Information Needed]
|
| 201 |
+
|
| 202 |
+
## Model Card Contact
|
| 203 |
+
|
| 204 |
+
[More Information Needed]
|
| 205 |
+
### Framework versions
|
| 206 |
+
|
| 207 |
+
- PEFT 0.18.1
|
starcoder2-7b/dp3/adapter/adapter_config.json
ADDED
|
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"alora_invocation_tokens": null,
|
| 3 |
+
"alpha_pattern": {},
|
| 4 |
+
"arrow_config": null,
|
| 5 |
+
"auto_mapping": null,
|
| 6 |
+
"base_model_name_or_path": "bigcode/starcoder2-7b",
|
| 7 |
+
"bias": "none",
|
| 8 |
+
"corda_config": null,
|
| 9 |
+
"ensure_weight_tying": true,
|
| 10 |
+
"eva_config": null,
|
| 11 |
+
"exclude_modules": null,
|
| 12 |
+
"fan_in_fan_out": false,
|
| 13 |
+
"inference_mode": true,
|
| 14 |
+
"init_lora_weights": true,
|
| 15 |
+
"layer_replication": null,
|
| 16 |
+
"layers_pattern": null,
|
| 17 |
+
"layers_to_transform": null,
|
| 18 |
+
"loftq_config": {},
|
| 19 |
+
"lora_alpha": 32,
|
| 20 |
+
"lora_bias": false,
|
| 21 |
+
"lora_dropout": 0.05,
|
| 22 |
+
"megatron_config": null,
|
| 23 |
+
"megatron_core": "megatron.core",
|
| 24 |
+
"modules_to_save": [
|
| 25 |
+
"lm_head",
|
| 26 |
+
"embed_tokens"
|
| 27 |
+
],
|
| 28 |
+
"peft_type": "LORA",
|
| 29 |
+
"peft_version": "0.18.1",
|
| 30 |
+
"qalora_group_size": 16,
|
| 31 |
+
"r": 16,
|
| 32 |
+
"rank_pattern": {},
|
| 33 |
+
"revision": null,
|
| 34 |
+
"target_modules": [
|
| 35 |
+
"k_proj",
|
| 36 |
+
"q_proj",
|
| 37 |
+
"v_proj",
|
| 38 |
+
"o_proj"
|
| 39 |
+
],
|
| 40 |
+
"target_parameters": null,
|
| 41 |
+
"task_type": "CAUSAL_LM",
|
| 42 |
+
"trainable_token_indices": null,
|
| 43 |
+
"use_dora": false,
|
| 44 |
+
"use_qalora": false,
|
| 45 |
+
"use_rslora": false
|
| 46 |
+
}
|
starcoder2-7b/dp3/adapter/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:853fac2b8cce7de6abb5b5d69b51e3a5f349fcbb8c2008b2dfeca74d3f065d9a
|
| 3 |
+
size 2804312360
|
starcoder2-7b/dp3/audit_results.json
ADDED
|
@@ -0,0 +1,137 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"delta": 1e-05,
|
| 3 |
+
"num_canaries": 500,
|
| 4 |
+
"num_members": 250,
|
| 5 |
+
"paper_guess_fraction": 0.2,
|
| 6 |
+
"paper_guess_steps": 20,
|
| 7 |
+
"loss": {
|
| 8 |
+
"auc": 0.525808,
|
| 9 |
+
"empirical_epsilon": {
|
| 10 |
+
"0.05": 0.05073561053723097,
|
| 11 |
+
"0.01": 0.0
|
| 12 |
+
},
|
| 13 |
+
"empirical_epsilon_details": {
|
| 14 |
+
"0.05": {
|
| 15 |
+
"epsilon": 0.05073561053723097,
|
| 16 |
+
"num_guesses": 100,
|
| 17 |
+
"correct_guesses": 60,
|
| 18 |
+
"candidate_num_guesses": [
|
| 19 |
+
5,
|
| 20 |
+
10,
|
| 21 |
+
15,
|
| 22 |
+
20,
|
| 23 |
+
25,
|
| 24 |
+
30,
|
| 25 |
+
35,
|
| 26 |
+
40,
|
| 27 |
+
45,
|
| 28 |
+
50,
|
| 29 |
+
55,
|
| 30 |
+
60,
|
| 31 |
+
65,
|
| 32 |
+
70,
|
| 33 |
+
75,
|
| 34 |
+
80,
|
| 35 |
+
85,
|
| 36 |
+
90,
|
| 37 |
+
95,
|
| 38 |
+
100
|
| 39 |
+
],
|
| 40 |
+
"direction": "higher"
|
| 41 |
+
},
|
| 42 |
+
"0.01": {
|
| 43 |
+
"epsilon": 0.0,
|
| 44 |
+
"num_guesses": 0,
|
| 45 |
+
"correct_guesses": 0,
|
| 46 |
+
"candidate_num_guesses": [
|
| 47 |
+
5,
|
| 48 |
+
10,
|
| 49 |
+
15,
|
| 50 |
+
20,
|
| 51 |
+
25,
|
| 52 |
+
30,
|
| 53 |
+
35,
|
| 54 |
+
40,
|
| 55 |
+
45,
|
| 56 |
+
50,
|
| 57 |
+
55,
|
| 58 |
+
60,
|
| 59 |
+
65,
|
| 60 |
+
70,
|
| 61 |
+
75,
|
| 62 |
+
80,
|
| 63 |
+
85,
|
| 64 |
+
90,
|
| 65 |
+
95,
|
| 66 |
+
100
|
| 67 |
+
],
|
| 68 |
+
"direction": "lower"
|
| 69 |
+
}
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"embedding": {
|
| 73 |
+
"auc": 0.520928,
|
| 74 |
+
"empirical_epsilon": {
|
| 75 |
+
"0.05": 0.0,
|
| 76 |
+
"0.01": 0.0
|
| 77 |
+
},
|
| 78 |
+
"empirical_epsilon_details": {
|
| 79 |
+
"0.05": {
|
| 80 |
+
"epsilon": 0.0,
|
| 81 |
+
"num_guesses": 0,
|
| 82 |
+
"correct_guesses": 0,
|
| 83 |
+
"candidate_num_guesses": [
|
| 84 |
+
5,
|
| 85 |
+
10,
|
| 86 |
+
15,
|
| 87 |
+
20,
|
| 88 |
+
25,
|
| 89 |
+
30,
|
| 90 |
+
35,
|
| 91 |
+
40,
|
| 92 |
+
45,
|
| 93 |
+
50,
|
| 94 |
+
55,
|
| 95 |
+
60,
|
| 96 |
+
65,
|
| 97 |
+
70,
|
| 98 |
+
75,
|
| 99 |
+
80,
|
| 100 |
+
85,
|
| 101 |
+
90,
|
| 102 |
+
95,
|
| 103 |
+
100
|
| 104 |
+
],
|
| 105 |
+
"direction": "lower"
|
| 106 |
+
},
|
| 107 |
+
"0.01": {
|
| 108 |
+
"epsilon": 0.0,
|
| 109 |
+
"num_guesses": 0,
|
| 110 |
+
"correct_guesses": 0,
|
| 111 |
+
"candidate_num_guesses": [
|
| 112 |
+
5,
|
| 113 |
+
10,
|
| 114 |
+
15,
|
| 115 |
+
20,
|
| 116 |
+
25,
|
| 117 |
+
30,
|
| 118 |
+
35,
|
| 119 |
+
40,
|
| 120 |
+
45,
|
| 121 |
+
50,
|
| 122 |
+
55,
|
| 123 |
+
60,
|
| 124 |
+
65,
|
| 125 |
+
70,
|
| 126 |
+
75,
|
| 127 |
+
80,
|
| 128 |
+
85,
|
| 129 |
+
90,
|
| 130 |
+
95,
|
| 131 |
+
100
|
| 132 |
+
],
|
| 133 |
+
"direction": "lower"
|
| 134 |
+
}
|
| 135 |
+
}
|
| 136 |
+
}
|
| 137 |
+
}
|
starcoder2-7b/dp3/audit_scores.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4fe821d6aec2e4b9779349695c9909f92bfbca3b25bd240d418bb85f362550a5
|
| 3 |
+
size 12784
|
starcoder2-7b/dp3/canary_meta.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
starcoder2-7b/dp3/codecarbon.csv
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
timestamp,project_name,run_id,experiment_id,duration,emissions,emissions_rate,cpu_power,gpu_power,ram_power,cpu_energy,gpu_energy,ram_energy,energy_consumed,water_consumed,country_name,country_iso_code,region,cloud_provider,cloud_region,os,python_version,codecarbon_version,cpu_count,cpu_model,gpu_count,gpu_model,longitude,latitude,ram_total_size,tracking_mode,cpu_utilization_percent,gpu_utilization_percent,ram_utilization_percent,ram_used_gb,on_cloud,pue,wue
|
| 2 |
+
2026-03-21T13:36:12,codedp-starcoder2-7b-cpt-dp3,457cc088-8edf-418e-8c6d-170a67e73d63,5b0fa12a-3dd7-45bb-9766-cc326314d9f1,3938.0607158355415,0.5702959606126596,0.00014481644691749236,179.92904019353898,2298.550082608512,70.0,0.18975838183590477,2.514705143984827,0.07382037755078118,2.778283903371513,0.0,United States,USA,california,,,Linux-5.15.0-157-generic-x86_64-with-glibc2.39,3.12.13,3.2.5,224,Intel(R) Xeon(R) Platinum 8480C,4,4 x NVIDIA H200,-121.9552,37.3541,2015.5625190734863,machine,2.4580389144905275,95.87682458386683,2.36410650281618,47.93522234837092,N,1.0,0.0
|
starcoder2-7b/dp3/epochs/epoch_001/adapter/README.md
ADDED
|
@@ -0,0 +1,207 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
base_model: bigcode/starcoder2-7b
|
| 3 |
+
library_name: peft
|
| 4 |
+
pipeline_tag: text-generation
|
| 5 |
+
tags:
|
| 6 |
+
- base_model:adapter:bigcode/starcoder2-7b
|
| 7 |
+
- lora
|
| 8 |
+
- transformers
|
| 9 |
+
---
|
| 10 |
+
|
| 11 |
+
# Model Card for Model ID
|
| 12 |
+
|
| 13 |
+
<!-- Provide a quick summary of what the model is/does. -->
|
| 14 |
+
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
## Model Details
|
| 18 |
+
|
| 19 |
+
### Model Description
|
| 20 |
+
|
| 21 |
+
<!-- Provide a longer summary of what this model is. -->
|
| 22 |
+
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
- **Developed by:** [More Information Needed]
|
| 26 |
+
- **Funded by [optional]:** [More Information Needed]
|
| 27 |
+
- **Shared by [optional]:** [More Information Needed]
|
| 28 |
+
- **Model type:** [More Information Needed]
|
| 29 |
+
- **Language(s) (NLP):** [More Information Needed]
|
| 30 |
+
- **License:** [More Information Needed]
|
| 31 |
+
- **Finetuned from model [optional]:** [More Information Needed]
|
| 32 |
+
|
| 33 |
+
### Model Sources [optional]
|
| 34 |
+
|
| 35 |
+
<!-- Provide the basic links for the model. -->
|
| 36 |
+
|
| 37 |
+
- **Repository:** [More Information Needed]
|
| 38 |
+
- **Paper [optional]:** [More Information Needed]
|
| 39 |
+
- **Demo [optional]:** [More Information Needed]
|
| 40 |
+
|
| 41 |
+
## Uses
|
| 42 |
+
|
| 43 |
+
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
| 44 |
+
|
| 45 |
+
### Direct Use
|
| 46 |
+
|
| 47 |
+
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
| 48 |
+
|
| 49 |
+
[More Information Needed]
|
| 50 |
+
|
| 51 |
+
### Downstream Use [optional]
|
| 52 |
+
|
| 53 |
+
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
|
| 54 |
+
|
| 55 |
+
[More Information Needed]
|
| 56 |
+
|
| 57 |
+
### Out-of-Scope Use
|
| 58 |
+
|
| 59 |
+
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
|
| 60 |
+
|
| 61 |
+
[More Information Needed]
|
| 62 |
+
|
| 63 |
+
## Bias, Risks, and Limitations
|
| 64 |
+
|
| 65 |
+
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
|
| 66 |
+
|
| 67 |
+
[More Information Needed]
|
| 68 |
+
|
| 69 |
+
### Recommendations
|
| 70 |
+
|
| 71 |
+
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
|
| 72 |
+
|
| 73 |
+
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
|
| 74 |
+
|
| 75 |
+
## How to Get Started with the Model
|
| 76 |
+
|
| 77 |
+
Use the code below to get started with the model.
|
| 78 |
+
|
| 79 |
+
[More Information Needed]
|
| 80 |
+
|
| 81 |
+
## Training Details
|
| 82 |
+
|
| 83 |
+
### Training Data
|
| 84 |
+
|
| 85 |
+
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
| 86 |
+
|
| 87 |
+
[More Information Needed]
|
| 88 |
+
|
| 89 |
+
### Training Procedure
|
| 90 |
+
|
| 91 |
+
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
|
| 92 |
+
|
| 93 |
+
#### Preprocessing [optional]
|
| 94 |
+
|
| 95 |
+
[More Information Needed]
|
| 96 |
+
|
| 97 |
+
|
| 98 |
+
#### Training Hyperparameters
|
| 99 |
+
|
| 100 |
+
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
| 101 |
+
|
| 102 |
+
#### Speeds, Sizes, Times [optional]
|
| 103 |
+
|
| 104 |
+
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|
| 105 |
+
|
| 106 |
+
[More Information Needed]
|
| 107 |
+
|
| 108 |
+
## Evaluation
|
| 109 |
+
|
| 110 |
+
<!-- This section describes the evaluation protocols and provides the results. -->
|
| 111 |
+
|
| 112 |
+
### Testing Data, Factors & Metrics
|
| 113 |
+
|
| 114 |
+
#### Testing Data
|
| 115 |
+
|
| 116 |
+
<!-- This should link to a Dataset Card if possible. -->
|
| 117 |
+
|
| 118 |
+
[More Information Needed]
|
| 119 |
+
|
| 120 |
+
#### Factors
|
| 121 |
+
|
| 122 |
+
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
|
| 123 |
+
|
| 124 |
+
[More Information Needed]
|
| 125 |
+
|
| 126 |
+
#### Metrics
|
| 127 |
+
|
| 128 |
+
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
|
| 129 |
+
|
| 130 |
+
[More Information Needed]
|
| 131 |
+
|
| 132 |
+
### Results
|
| 133 |
+
|
| 134 |
+
[More Information Needed]
|
| 135 |
+
|
| 136 |
+
#### Summary
|
| 137 |
+
|
| 138 |
+
|
| 139 |
+
|
| 140 |
+
## Model Examination [optional]
|
| 141 |
+
|
| 142 |
+
<!-- Relevant interpretability work for the model goes here -->
|
| 143 |
+
|
| 144 |
+
[More Information Needed]
|
| 145 |
+
|
| 146 |
+
## Environmental Impact
|
| 147 |
+
|
| 148 |
+
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
|
| 149 |
+
|
| 150 |
+
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
| 151 |
+
|
| 152 |
+
- **Hardware Type:** [More Information Needed]
|
| 153 |
+
- **Hours used:** [More Information Needed]
|
| 154 |
+
- **Cloud Provider:** [More Information Needed]
|
| 155 |
+
- **Compute Region:** [More Information Needed]
|
| 156 |
+
- **Carbon Emitted:** [More Information Needed]
|
| 157 |
+
|
| 158 |
+
## Technical Specifications [optional]
|
| 159 |
+
|
| 160 |
+
### Model Architecture and Objective
|
| 161 |
+
|
| 162 |
+
[More Information Needed]
|
| 163 |
+
|
| 164 |
+
### Compute Infrastructure
|
| 165 |
+
|
| 166 |
+
[More Information Needed]
|
| 167 |
+
|
| 168 |
+
#### Hardware
|
| 169 |
+
|
| 170 |
+
[More Information Needed]
|
| 171 |
+
|
| 172 |
+
#### Software
|
| 173 |
+
|
| 174 |
+
[More Information Needed]
|
| 175 |
+
|
| 176 |
+
## Citation [optional]
|
| 177 |
+
|
| 178 |
+
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
| 179 |
+
|
| 180 |
+
**BibTeX:**
|
| 181 |
+
|
| 182 |
+
[More Information Needed]
|
| 183 |
+
|
| 184 |
+
**APA:**
|
| 185 |
+
|
| 186 |
+
[More Information Needed]
|
| 187 |
+
|
| 188 |
+
## Glossary [optional]
|
| 189 |
+
|
| 190 |
+
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
|
| 191 |
+
|
| 192 |
+
[More Information Needed]
|
| 193 |
+
|
| 194 |
+
## More Information [optional]
|
| 195 |
+
|
| 196 |
+
[More Information Needed]
|
| 197 |
+
|
| 198 |
+
## Model Card Authors [optional]
|
| 199 |
+
|
| 200 |
+
[More Information Needed]
|
| 201 |
+
|
| 202 |
+
## Model Card Contact
|
| 203 |
+
|
| 204 |
+
[More Information Needed]
|
| 205 |
+
### Framework versions
|
| 206 |
+
|
| 207 |
+
- PEFT 0.18.1
|
starcoder2-7b/dp3/epochs/epoch_001/adapter/adapter_config.json
ADDED
|
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"alora_invocation_tokens": null,
|
| 3 |
+
"alpha_pattern": {},
|
| 4 |
+
"arrow_config": null,
|
| 5 |
+
"auto_mapping": null,
|
| 6 |
+
"base_model_name_or_path": "bigcode/starcoder2-7b",
|
| 7 |
+
"bias": "none",
|
| 8 |
+
"corda_config": null,
|
| 9 |
+
"ensure_weight_tying": true,
|
| 10 |
+
"eva_config": null,
|
| 11 |
+
"exclude_modules": null,
|
| 12 |
+
"fan_in_fan_out": false,
|
| 13 |
+
"inference_mode": true,
|
| 14 |
+
"init_lora_weights": true,
|
| 15 |
+
"layer_replication": null,
|
| 16 |
+
"layers_pattern": null,
|
| 17 |
+
"layers_to_transform": null,
|
| 18 |
+
"loftq_config": {},
|
| 19 |
+
"lora_alpha": 32,
|
| 20 |
+
"lora_bias": false,
|
| 21 |
+
"lora_dropout": 0.05,
|
| 22 |
+
"megatron_config": null,
|
| 23 |
+
"megatron_core": "megatron.core",
|
| 24 |
+
"modules_to_save": [
|
| 25 |
+
"lm_head",
|
| 26 |
+
"embed_tokens"
|
| 27 |
+
],
|
| 28 |
+
"peft_type": "LORA",
|
| 29 |
+
"peft_version": "0.18.1",
|
| 30 |
+
"qalora_group_size": 16,
|
| 31 |
+
"r": 16,
|
| 32 |
+
"rank_pattern": {},
|
| 33 |
+
"revision": null,
|
| 34 |
+
"target_modules": [
|
| 35 |
+
"k_proj",
|
| 36 |
+
"q_proj",
|
| 37 |
+
"v_proj",
|
| 38 |
+
"o_proj"
|
| 39 |
+
],
|
| 40 |
+
"target_parameters": null,
|
| 41 |
+
"task_type": "CAUSAL_LM",
|
| 42 |
+
"trainable_token_indices": null,
|
| 43 |
+
"use_dora": false,
|
| 44 |
+
"use_qalora": false,
|
| 45 |
+
"use_rslora": false
|
| 46 |
+
}
|
starcoder2-7b/dp3/epochs/epoch_001/adapter/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1eb0abf981ab5eb12b4b55a3c95748d96446558dfe488188b7929dae1d80c5b4
|
| 3 |
+
size 2804312360
|
starcoder2-7b/dp3/epochs/epoch_001/audit_results.json
ADDED
|
@@ -0,0 +1,137 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"delta": 1e-05,
|
| 3 |
+
"num_canaries": 500,
|
| 4 |
+
"num_members": 250,
|
| 5 |
+
"paper_guess_fraction": 0.2,
|
| 6 |
+
"paper_guess_steps": 20,
|
| 7 |
+
"loss": {
|
| 8 |
+
"auc": 0.527168,
|
| 9 |
+
"empirical_epsilon": {
|
| 10 |
+
"0.05": 0.09125839080661535,
|
| 11 |
+
"0.01": 0.0
|
| 12 |
+
},
|
| 13 |
+
"empirical_epsilon_details": {
|
| 14 |
+
"0.05": {
|
| 15 |
+
"epsilon": 0.09125839080661535,
|
| 16 |
+
"num_guesses": 100,
|
| 17 |
+
"correct_guesses": 61,
|
| 18 |
+
"candidate_num_guesses": [
|
| 19 |
+
5,
|
| 20 |
+
10,
|
| 21 |
+
15,
|
| 22 |
+
20,
|
| 23 |
+
25,
|
| 24 |
+
30,
|
| 25 |
+
35,
|
| 26 |
+
40,
|
| 27 |
+
45,
|
| 28 |
+
50,
|
| 29 |
+
55,
|
| 30 |
+
60,
|
| 31 |
+
65,
|
| 32 |
+
70,
|
| 33 |
+
75,
|
| 34 |
+
80,
|
| 35 |
+
85,
|
| 36 |
+
90,
|
| 37 |
+
95,
|
| 38 |
+
100
|
| 39 |
+
],
|
| 40 |
+
"direction": "higher"
|
| 41 |
+
},
|
| 42 |
+
"0.01": {
|
| 43 |
+
"epsilon": 0.0,
|
| 44 |
+
"num_guesses": 0,
|
| 45 |
+
"correct_guesses": 0,
|
| 46 |
+
"candidate_num_guesses": [
|
| 47 |
+
5,
|
| 48 |
+
10,
|
| 49 |
+
15,
|
| 50 |
+
20,
|
| 51 |
+
25,
|
| 52 |
+
30,
|
| 53 |
+
35,
|
| 54 |
+
40,
|
| 55 |
+
45,
|
| 56 |
+
50,
|
| 57 |
+
55,
|
| 58 |
+
60,
|
| 59 |
+
65,
|
| 60 |
+
70,
|
| 61 |
+
75,
|
| 62 |
+
80,
|
| 63 |
+
85,
|
| 64 |
+
90,
|
| 65 |
+
95,
|
| 66 |
+
100
|
| 67 |
+
],
|
| 68 |
+
"direction": "lower"
|
| 69 |
+
}
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"embedding": {
|
| 73 |
+
"auc": 0.514792,
|
| 74 |
+
"empirical_epsilon": {
|
| 75 |
+
"0.05": 0.0,
|
| 76 |
+
"0.01": 0.0
|
| 77 |
+
},
|
| 78 |
+
"empirical_epsilon_details": {
|
| 79 |
+
"0.05": {
|
| 80 |
+
"epsilon": 0.0,
|
| 81 |
+
"num_guesses": 0,
|
| 82 |
+
"correct_guesses": 0,
|
| 83 |
+
"candidate_num_guesses": [
|
| 84 |
+
5,
|
| 85 |
+
10,
|
| 86 |
+
15,
|
| 87 |
+
20,
|
| 88 |
+
25,
|
| 89 |
+
30,
|
| 90 |
+
35,
|
| 91 |
+
40,
|
| 92 |
+
45,
|
| 93 |
+
50,
|
| 94 |
+
55,
|
| 95 |
+
60,
|
| 96 |
+
65,
|
| 97 |
+
70,
|
| 98 |
+
75,
|
| 99 |
+
80,
|
| 100 |
+
85,
|
| 101 |
+
90,
|
| 102 |
+
95,
|
| 103 |
+
100
|
| 104 |
+
],
|
| 105 |
+
"direction": "lower"
|
| 106 |
+
},
|
| 107 |
+
"0.01": {
|
| 108 |
+
"epsilon": 0.0,
|
| 109 |
+
"num_guesses": 0,
|
| 110 |
+
"correct_guesses": 0,
|
| 111 |
+
"candidate_num_guesses": [
|
| 112 |
+
5,
|
| 113 |
+
10,
|
| 114 |
+
15,
|
| 115 |
+
20,
|
| 116 |
+
25,
|
| 117 |
+
30,
|
| 118 |
+
35,
|
| 119 |
+
40,
|
| 120 |
+
45,
|
| 121 |
+
50,
|
| 122 |
+
55,
|
| 123 |
+
60,
|
| 124 |
+
65,
|
| 125 |
+
70,
|
| 126 |
+
75,
|
| 127 |
+
80,
|
| 128 |
+
85,
|
| 129 |
+
90,
|
| 130 |
+
95,
|
| 131 |
+
100
|
| 132 |
+
],
|
| 133 |
+
"direction": "lower"
|
| 134 |
+
}
|
| 135 |
+
}
|
| 136 |
+
}
|
| 137 |
+
}
|
starcoder2-7b/dp3/epochs/epoch_001/audit_scores.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e3dedd61e7d2e9b43bbdd58fe90dba89fa6bab75a49f7c8b244e1a9d3a2cb374
|
| 3 |
+
size 12784
|
starcoder2-7b/dp3/epochs/epoch_002/adapter/README.md
ADDED
|
@@ -0,0 +1,207 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
base_model: bigcode/starcoder2-7b
|
| 3 |
+
library_name: peft
|
| 4 |
+
pipeline_tag: text-generation
|
| 5 |
+
tags:
|
| 6 |
+
- base_model:adapter:bigcode/starcoder2-7b
|
| 7 |
+
- lora
|
| 8 |
+
- transformers
|
| 9 |
+
---
|
| 10 |
+
|
| 11 |
+
# Model Card for Model ID
|
| 12 |
+
|
| 13 |
+
<!-- Provide a quick summary of what the model is/does. -->
|
| 14 |
+
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
## Model Details
|
| 18 |
+
|
| 19 |
+
### Model Description
|
| 20 |
+
|
| 21 |
+
<!-- Provide a longer summary of what this model is. -->
|
| 22 |
+
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
- **Developed by:** [More Information Needed]
|
| 26 |
+
- **Funded by [optional]:** [More Information Needed]
|
| 27 |
+
- **Shared by [optional]:** [More Information Needed]
|
| 28 |
+
- **Model type:** [More Information Needed]
|
| 29 |
+
- **Language(s) (NLP):** [More Information Needed]
|
| 30 |
+
- **License:** [More Information Needed]
|
| 31 |
+
- **Finetuned from model [optional]:** [More Information Needed]
|
| 32 |
+
|
| 33 |
+
### Model Sources [optional]
|
| 34 |
+
|
| 35 |
+
<!-- Provide the basic links for the model. -->
|
| 36 |
+
|
| 37 |
+
- **Repository:** [More Information Needed]
|
| 38 |
+
- **Paper [optional]:** [More Information Needed]
|
| 39 |
+
- **Demo [optional]:** [More Information Needed]
|
| 40 |
+
|
| 41 |
+
## Uses
|
| 42 |
+
|
| 43 |
+
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
| 44 |
+
|
| 45 |
+
### Direct Use
|
| 46 |
+
|
| 47 |
+
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
| 48 |
+
|
| 49 |
+
[More Information Needed]
|
| 50 |
+
|
| 51 |
+
### Downstream Use [optional]
|
| 52 |
+
|
| 53 |
+
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
|
| 54 |
+
|
| 55 |
+
[More Information Needed]
|
| 56 |
+
|
| 57 |
+
### Out-of-Scope Use
|
| 58 |
+
|
| 59 |
+
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
|
| 60 |
+
|
| 61 |
+
[More Information Needed]
|
| 62 |
+
|
| 63 |
+
## Bias, Risks, and Limitations
|
| 64 |
+
|
| 65 |
+
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
|
| 66 |
+
|
| 67 |
+
[More Information Needed]
|
| 68 |
+
|
| 69 |
+
### Recommendations
|
| 70 |
+
|
| 71 |
+
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
|
| 72 |
+
|
| 73 |
+
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
|
| 74 |
+
|
| 75 |
+
## How to Get Started with the Model
|
| 76 |
+
|
| 77 |
+
Use the code below to get started with the model.
|
| 78 |
+
|
| 79 |
+
[More Information Needed]
|
| 80 |
+
|
| 81 |
+
## Training Details
|
| 82 |
+
|
| 83 |
+
### Training Data
|
| 84 |
+
|
| 85 |
+
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
| 86 |
+
|
| 87 |
+
[More Information Needed]
|
| 88 |
+
|
| 89 |
+
### Training Procedure
|
| 90 |
+
|
| 91 |
+
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
|
| 92 |
+
|
| 93 |
+
#### Preprocessing [optional]
|
| 94 |
+
|
| 95 |
+
[More Information Needed]
|
| 96 |
+
|
| 97 |
+
|
| 98 |
+
#### Training Hyperparameters
|
| 99 |
+
|
| 100 |
+
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
| 101 |
+
|
| 102 |
+
#### Speeds, Sizes, Times [optional]
|
| 103 |
+
|
| 104 |
+
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|
| 105 |
+
|
| 106 |
+
[More Information Needed]
|
| 107 |
+
|
| 108 |
+
## Evaluation
|
| 109 |
+
|
| 110 |
+
<!-- This section describes the evaluation protocols and provides the results. -->
|
| 111 |
+
|
| 112 |
+
### Testing Data, Factors & Metrics
|
| 113 |
+
|
| 114 |
+
#### Testing Data
|
| 115 |
+
|
| 116 |
+
<!-- This should link to a Dataset Card if possible. -->
|
| 117 |
+
|
| 118 |
+
[More Information Needed]
|
| 119 |
+
|
| 120 |
+
#### Factors
|
| 121 |
+
|
| 122 |
+
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
|
| 123 |
+
|
| 124 |
+
[More Information Needed]
|
| 125 |
+
|
| 126 |
+
#### Metrics
|
| 127 |
+
|
| 128 |
+
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
|
| 129 |
+
|
| 130 |
+
[More Information Needed]
|
| 131 |
+
|
| 132 |
+
### Results
|
| 133 |
+
|
| 134 |
+
[More Information Needed]
|
| 135 |
+
|
| 136 |
+
#### Summary
|
| 137 |
+
|
| 138 |
+
|
| 139 |
+
|
| 140 |
+
## Model Examination [optional]
|
| 141 |
+
|
| 142 |
+
<!-- Relevant interpretability work for the model goes here -->
|
| 143 |
+
|
| 144 |
+
[More Information Needed]
|
| 145 |
+
|
| 146 |
+
## Environmental Impact
|
| 147 |
+
|
| 148 |
+
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
|
| 149 |
+
|
| 150 |
+
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
| 151 |
+
|
| 152 |
+
- **Hardware Type:** [More Information Needed]
|
| 153 |
+
- **Hours used:** [More Information Needed]
|
| 154 |
+
- **Cloud Provider:** [More Information Needed]
|
| 155 |
+
- **Compute Region:** [More Information Needed]
|
| 156 |
+
- **Carbon Emitted:** [More Information Needed]
|
| 157 |
+
|
| 158 |
+
## Technical Specifications [optional]
|
| 159 |
+
|
| 160 |
+
### Model Architecture and Objective
|
| 161 |
+
|
| 162 |
+
[More Information Needed]
|
| 163 |
+
|
| 164 |
+
### Compute Infrastructure
|
| 165 |
+
|
| 166 |
+
[More Information Needed]
|
| 167 |
+
|
| 168 |
+
#### Hardware
|
| 169 |
+
|
| 170 |
+
[More Information Needed]
|
| 171 |
+
|
| 172 |
+
#### Software
|
| 173 |
+
|
| 174 |
+
[More Information Needed]
|
| 175 |
+
|
| 176 |
+
## Citation [optional]
|
| 177 |
+
|
| 178 |
+
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
| 179 |
+
|
| 180 |
+
**BibTeX:**
|
| 181 |
+
|
| 182 |
+
[More Information Needed]
|
| 183 |
+
|
| 184 |
+
**APA:**
|
| 185 |
+
|
| 186 |
+
[More Information Needed]
|
| 187 |
+
|
| 188 |
+
## Glossary [optional]
|
| 189 |
+
|
| 190 |
+
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
|
| 191 |
+
|
| 192 |
+
[More Information Needed]
|
| 193 |
+
|
| 194 |
+
## More Information [optional]
|
| 195 |
+
|
| 196 |
+
[More Information Needed]
|
| 197 |
+
|
| 198 |
+
## Model Card Authors [optional]
|
| 199 |
+
|
| 200 |
+
[More Information Needed]
|
| 201 |
+
|
| 202 |
+
## Model Card Contact
|
| 203 |
+
|
| 204 |
+
[More Information Needed]
|
| 205 |
+
### Framework versions
|
| 206 |
+
|
| 207 |
+
- PEFT 0.18.1
|
starcoder2-7b/dp3/epochs/epoch_002/adapter/adapter_config.json
ADDED
|
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"alora_invocation_tokens": null,
|
| 3 |
+
"alpha_pattern": {},
|
| 4 |
+
"arrow_config": null,
|
| 5 |
+
"auto_mapping": null,
|
| 6 |
+
"base_model_name_or_path": "bigcode/starcoder2-7b",
|
| 7 |
+
"bias": "none",
|
| 8 |
+
"corda_config": null,
|
| 9 |
+
"ensure_weight_tying": true,
|
| 10 |
+
"eva_config": null,
|
| 11 |
+
"exclude_modules": null,
|
| 12 |
+
"fan_in_fan_out": false,
|
| 13 |
+
"inference_mode": true,
|
| 14 |
+
"init_lora_weights": true,
|
| 15 |
+
"layer_replication": null,
|
| 16 |
+
"layers_pattern": null,
|
| 17 |
+
"layers_to_transform": null,
|
| 18 |
+
"loftq_config": {},
|
| 19 |
+
"lora_alpha": 32,
|
| 20 |
+
"lora_bias": false,
|
| 21 |
+
"lora_dropout": 0.05,
|
| 22 |
+
"megatron_config": null,
|
| 23 |
+
"megatron_core": "megatron.core",
|
| 24 |
+
"modules_to_save": [
|
| 25 |
+
"lm_head",
|
| 26 |
+
"embed_tokens"
|
| 27 |
+
],
|
| 28 |
+
"peft_type": "LORA",
|
| 29 |
+
"peft_version": "0.18.1",
|
| 30 |
+
"qalora_group_size": 16,
|
| 31 |
+
"r": 16,
|
| 32 |
+
"rank_pattern": {},
|
| 33 |
+
"revision": null,
|
| 34 |
+
"target_modules": [
|
| 35 |
+
"k_proj",
|
| 36 |
+
"q_proj",
|
| 37 |
+
"v_proj",
|
| 38 |
+
"o_proj"
|
| 39 |
+
],
|
| 40 |
+
"target_parameters": null,
|
| 41 |
+
"task_type": "CAUSAL_LM",
|
| 42 |
+
"trainable_token_indices": null,
|
| 43 |
+
"use_dora": false,
|
| 44 |
+
"use_qalora": false,
|
| 45 |
+
"use_rslora": false
|
| 46 |
+
}
|
starcoder2-7b/dp3/epochs/epoch_002/adapter/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:853fac2b8cce7de6abb5b5d69b51e3a5f349fcbb8c2008b2dfeca74d3f065d9a
|
| 3 |
+
size 2804312360
|
starcoder2-7b/dp3/epochs/epoch_002/audit_results.json
ADDED
|
@@ -0,0 +1,137 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"delta": 1e-05,
|
| 3 |
+
"num_canaries": 500,
|
| 4 |
+
"num_members": 250,
|
| 5 |
+
"paper_guess_fraction": 0.2,
|
| 6 |
+
"paper_guess_steps": 20,
|
| 7 |
+
"loss": {
|
| 8 |
+
"auc": 0.525808,
|
| 9 |
+
"empirical_epsilon": {
|
| 10 |
+
"0.05": 0.05073561053723097,
|
| 11 |
+
"0.01": 0.0
|
| 12 |
+
},
|
| 13 |
+
"empirical_epsilon_details": {
|
| 14 |
+
"0.05": {
|
| 15 |
+
"epsilon": 0.05073561053723097,
|
| 16 |
+
"num_guesses": 100,
|
| 17 |
+
"correct_guesses": 60,
|
| 18 |
+
"candidate_num_guesses": [
|
| 19 |
+
5,
|
| 20 |
+
10,
|
| 21 |
+
15,
|
| 22 |
+
20,
|
| 23 |
+
25,
|
| 24 |
+
30,
|
| 25 |
+
35,
|
| 26 |
+
40,
|
| 27 |
+
45,
|
| 28 |
+
50,
|
| 29 |
+
55,
|
| 30 |
+
60,
|
| 31 |
+
65,
|
| 32 |
+
70,
|
| 33 |
+
75,
|
| 34 |
+
80,
|
| 35 |
+
85,
|
| 36 |
+
90,
|
| 37 |
+
95,
|
| 38 |
+
100
|
| 39 |
+
],
|
| 40 |
+
"direction": "higher"
|
| 41 |
+
},
|
| 42 |
+
"0.01": {
|
| 43 |
+
"epsilon": 0.0,
|
| 44 |
+
"num_guesses": 0,
|
| 45 |
+
"correct_guesses": 0,
|
| 46 |
+
"candidate_num_guesses": [
|
| 47 |
+
5,
|
| 48 |
+
10,
|
| 49 |
+
15,
|
| 50 |
+
20,
|
| 51 |
+
25,
|
| 52 |
+
30,
|
| 53 |
+
35,
|
| 54 |
+
40,
|
| 55 |
+
45,
|
| 56 |
+
50,
|
| 57 |
+
55,
|
| 58 |
+
60,
|
| 59 |
+
65,
|
| 60 |
+
70,
|
| 61 |
+
75,
|
| 62 |
+
80,
|
| 63 |
+
85,
|
| 64 |
+
90,
|
| 65 |
+
95,
|
| 66 |
+
100
|
| 67 |
+
],
|
| 68 |
+
"direction": "lower"
|
| 69 |
+
}
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"embedding": {
|
| 73 |
+
"auc": 0.520928,
|
| 74 |
+
"empirical_epsilon": {
|
| 75 |
+
"0.05": 0.0,
|
| 76 |
+
"0.01": 0.0
|
| 77 |
+
},
|
| 78 |
+
"empirical_epsilon_details": {
|
| 79 |
+
"0.05": {
|
| 80 |
+
"epsilon": 0.0,
|
| 81 |
+
"num_guesses": 0,
|
| 82 |
+
"correct_guesses": 0,
|
| 83 |
+
"candidate_num_guesses": [
|
| 84 |
+
5,
|
| 85 |
+
10,
|
| 86 |
+
15,
|
| 87 |
+
20,
|
| 88 |
+
25,
|
| 89 |
+
30,
|
| 90 |
+
35,
|
| 91 |
+
40,
|
| 92 |
+
45,
|
| 93 |
+
50,
|
| 94 |
+
55,
|
| 95 |
+
60,
|
| 96 |
+
65,
|
| 97 |
+
70,
|
| 98 |
+
75,
|
| 99 |
+
80,
|
| 100 |
+
85,
|
| 101 |
+
90,
|
| 102 |
+
95,
|
| 103 |
+
100
|
| 104 |
+
],
|
| 105 |
+
"direction": "lower"
|
| 106 |
+
},
|
| 107 |
+
"0.01": {
|
| 108 |
+
"epsilon": 0.0,
|
| 109 |
+
"num_guesses": 0,
|
| 110 |
+
"correct_guesses": 0,
|
| 111 |
+
"candidate_num_guesses": [
|
| 112 |
+
5,
|
| 113 |
+
10,
|
| 114 |
+
15,
|
| 115 |
+
20,
|
| 116 |
+
25,
|
| 117 |
+
30,
|
| 118 |
+
35,
|
| 119 |
+
40,
|
| 120 |
+
45,
|
| 121 |
+
50,
|
| 122 |
+
55,
|
| 123 |
+
60,
|
| 124 |
+
65,
|
| 125 |
+
70,
|
| 126 |
+
75,
|
| 127 |
+
80,
|
| 128 |
+
85,
|
| 129 |
+
90,
|
| 130 |
+
95,
|
| 131 |
+
100
|
| 132 |
+
],
|
| 133 |
+
"direction": "lower"
|
| 134 |
+
}
|
| 135 |
+
}
|
| 136 |
+
}
|
| 137 |
+
}
|
starcoder2-7b/dp3/epochs/epoch_002/audit_scores.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4fe821d6aec2e4b9779349695c9909f92bfbca3b25bd240d418bb85f362550a5
|
| 3 |
+
size 12784
|
starcoder2-7b/dp3/metrics.jsonl
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"timestamp": 1774096500.8353336, "event": "train_step", "step": 10, "epoch": 1, "metrics": {"train/step_loss": 1.6323321043555417, "train/step_real_loss": 1.330917976796627, "train/lr": 0.00018181818181818183, "train/step_canary_loss": 8.062500158945719, "perf/step_duration_sec": 17.560116425156593, "perf/samples_per_sec": 7.573981674145647, "perf/tokens_per_sec": 6185.665138551686, "perf/logical_batch_size": 133.0, "perf/logical_token_count": 108621.0, "perf/physical_batches": 18.0, "privacy/epsilon": 0.7131647471248268, "system/cuda_memory_allocated_gb": 16.63217782974243, "system/cuda_max_memory_allocated_gb": 74.69534254074097}}
|
| 2 |
+
{"timestamp": 1774096676.393492, "event": "train_step", "step": 20, "epoch": 1, "metrics": {"train/step_loss": 1.826857232699429, "train/step_real_loss": 1.3395854756236076, "train/lr": 0.00019897180218885507, "train/step_canary_loss": 8.756944444444445, "perf/step_duration_sec": 18.843991773203015, "perf/samples_per_sec": 7.217154498729827, "perf/tokens_per_sec": 5434.1989336686165, "perf/logical_batch_size": 136.0, "perf/logical_token_count": 102402.0, "perf/physical_batches": 18.0, "privacy/epsilon": 0.9542480237478311, "system/cuda_memory_allocated_gb": 16.916476249694824, "system/cuda_max_memory_allocated_gb": 74.69534397125244}}
|
| 3 |
+
{"timestamp": 1774096852.7017646, "event": "train_step", "step": 30, "epoch": 1, "metrics": {"train/step_loss": 1.5258880311792546, "train/step_real_loss": 1.3621462360024452, "train/lr": 0.00019544467510209388, "train/step_canary_loss": 6.765625476837158, "perf/step_duration_sec": 18.006671600043774, "perf/samples_per_sec": 7.275081309289916, "perf/tokens_per_sec": 5874.1561099910805, "perf/logical_batch_size": 131.0, "perf/logical_token_count": 105774.0, "perf/physical_batches": 18.0, "privacy/epsilon": 1.145336977893831, "system/cuda_memory_allocated_gb": 16.442769050598145, "system/cuda_max_memory_allocated_gb": 74.69534397125244}}
|
| 4 |
+
{"timestamp": 1774097029.0227332, "event": "train_step", "step": 40, "epoch": 1, "metrics": {"train/step_loss": 1.6870967794347693, "train/step_real_loss": 1.2007466033101082, "train/lr": 0.00018949541262593762, "train/step_canary_loss": 10.580357142857142, "perf/step_duration_sec": 17.532358843833208, "perf/samples_per_sec": 7.643010344106256, "perf/tokens_per_sec": 6436.156195815627, "perf/logical_batch_size": 134.0, "perf/logical_token_count": 112841.0, "perf/physical_batches": 18.0, "privacy/epsilon": 1.311193108872294, "system/cuda_memory_allocated_gb": 16.726882934570312, "system/cuda_max_memory_allocated_gb": 74.69534683227539}}
|
| 5 |
+
{"timestamp": 1774097204.3534977, "event": "train_step", "step": 50, "epoch": 1, "metrics": {"train/step_loss": 1.5818433902881763, "train/step_real_loss": 1.2425691708922386, "train/lr": 0.00018127499143005268, "train/step_canary_loss": 7.785714830671038, "perf/step_duration_sec": 17.701514894142747, "perf/samples_per_sec": 7.569973575783576, "perf/tokens_per_sec": 6216.078152520672, "perf/logical_batch_size": 134.0, "perf/logical_token_count": 110034.0, "perf/physical_batches": 18.0, "privacy/epsilon": 1.4583641061524852, "system/cuda_memory_allocated_gb": 16.726882934570312, "system/cuda_max_memory_allocated_gb": 74.69534683227539}}
|
| 6 |
+
{"timestamp": 1774097231.3920577, "event": "eval_step", "step": 50, "epoch": 1, "metrics": {"eval/loss": 0.8423196817266531, "eval/duration_sec": 27.033981669694185}}
|
| 7 |
+
{"timestamp": 1774097408.5970793, "event": "train_step", "step": 60, "epoch": 1, "metrics": {"train/step_loss": 1.6164713336112804, "train/step_real_loss": 1.3143803551793098, "train/lr": 0.0001709920242324663, "train/step_canary_loss": 9.350000381469727, "perf/step_duration_sec": 17.616566620767117, "perf/samples_per_sec": 7.549711749349289, "perf/tokens_per_sec": 5881.736335493052, "perf/logical_batch_size": 133.0, "perf/logical_token_count": 103616.0, "perf/physical_batches": 17.0, "privacy/epsilon": 1.5929146459722179, "system/cuda_memory_allocated_gb": 16.63217782974243, "system/cuda_max_memory_allocated_gb": 74.69534969329834}}
|
| 8 |
+
{"timestamp": 1774097584.0989442, "event": "train_step", "step": 70, "epoch": 1, "metrics": {"train/step_loss": 1.5537053346633911, "train/step_real_loss": 1.274133626371622, "train/lr": 0.00015890746575622231, "train/step_canary_loss": 10.5, "perf/step_duration_sec": 17.34652972780168, "perf/samples_per_sec": 7.609591201889827, "perf/tokens_per_sec": 6165.613622913036, "perf/logical_batch_size": 132.0, "perf/logical_token_count": 106952.0, "perf/physical_batches": 17.0, "privacy/epsilon": 1.7182200620091768, "system/cuda_memory_allocated_gb": 16.537474155426025, "system/cuda_max_memory_allocated_gb": 74.69534969329834}}
|
| 9 |
+
{"timestamp": 1774097758.6071105, "event": "train_step", "step": 80, "epoch": 1, "metrics": {"train/step_loss": 1.6813924115403254, "train/step_real_loss": 1.3261733427643776, "train/lr": 0.00014532799038330385, "train/step_canary_loss": 10.77500057220459, "perf/step_duration_sec": 16.975525552406907, "perf/samples_per_sec": 7.834808977748694, "perf/tokens_per_sec": 6001.228043602786, "perf/logical_batch_size": 133.0, "perf/logical_token_count": 101874.0, "perf/physical_batches": 17.0, "privacy/epsilon": 1.8360751937320303, "system/cuda_memory_allocated_gb": 16.63217782974243, "system/cuda_max_memory_allocated_gb": 74.69534969329834}}
|
| 10 |
+
{"timestamp": 1774097934.2494926, "event": "train_step", "step": 90, "epoch": 1, "metrics": {"train/step_loss": 1.8558794460157408, "train/step_real_loss": 1.2407655008137226, "train/lr": 0.00013059820956358998, "train/step_canary_loss": 10.604166666666666, "perf/step_duration_sec": 17.994612561538815, "perf/samples_per_sec": 7.557817626520207, "perf/tokens_per_sec": 6203.967971981345, "perf/logical_batch_size": 136.0, "perf/logical_token_count": 111638.0, "perf/physical_batches": 18.0, "privacy/epsilon": 1.9476934830264792, "system/cuda_memory_allocated_gb": 16.916476249694824, "system/cuda_max_memory_allocated_gb": 74.69534969329834}}
|
| 11 |
+
{"timestamp": 1774098110.4647262, "event": "train_step", "step": 100, "epoch": 1, "metrics": {"train/step_loss": 1.86213939777319, "train/step_real_loss": 1.2256366163492203, "train/lr": 0.00011509192648058249, "train/step_canary_loss": 10.009375, "perf/step_duration_sec": 17.502219082787633, "perf/samples_per_sec": 7.884714466619527, "perf/tokens_per_sec": 6282.689039593839, "perf/logical_batch_size": 138.0, "perf/logical_token_count": 109961.0, "perf/physical_batches": 18.0, "privacy/epsilon": 2.054057754637441, "system/cuda_memory_allocated_gb": 16.34807252883911, "system/cuda_max_memory_allocated_gb": 74.69534969329834}}
|
| 12 |
+
{"timestamp": 1774098137.4743512, "event": "eval_step", "step": 100, "epoch": 1, "metrics": {"eval/loss": 0.8425039929905157, "eval/duration_sec": 27.006801065057516}}
|
| 13 |
+
{"timestamp": 1774098228.5487797, "event": "train_epoch", "step": 104, "epoch": 1, "metrics": {"train/epoch_loss": 1.7094427731604633, "train/epoch_real_loss": 1.285309421313279, "train/epoch_canary_loss": 9.719154772066274, "perf/epoch_duration_sec": 1878.0538540109992, "perf/epoch_samples_per_sec": 29.64824458099586, "perf/epoch_tokens_per_sec": 23429.532069074678, "perf/epoch_samples": 55681.0, "perf/epoch_tokens": 44001923.0, "system/cuda_epoch_peak_memory_gb": 74.69534969329834, "eval/loss": 0.8425441932404178, "eval/duration_sec": 26.999734753742814, "privacy/epsilon": 2.0952814257505974}}
|
| 14 |
+
{"timestamp": 1774098238.4572744, "event": "audit_epoch", "step": 104, "epoch": 1, "metrics": {"audit/delta": 1e-05, "audit/num_canaries": 500.0, "audit/num_members": 250.0, "audit/paper_guess_fraction": 0.2, "audit/paper_guess_steps": 20.0, "audit/loss/auc": 0.527168, "audit/loss/empirical_epsilon/0.05": 0.09125839080661535, "audit/loss/empirical_epsilon/0.01": 0.0, "audit/loss/empirical_epsilon_details/0.05/epsilon": 0.09125839080661535, "audit/loss/empirical_epsilon_details/0.05/num_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.05/correct_guesses": 61.0, "audit/loss/empirical_epsilon_details/0.01/epsilon": 0.0, "audit/loss/empirical_epsilon_details/0.01/num_guesses": 0.0, "audit/loss/empirical_epsilon_details/0.01/correct_guesses": 0.0, "audit/embedding/auc": 0.514792, "audit/embedding/empirical_epsilon/0.05": 0.0, "audit/embedding/empirical_epsilon/0.01": 0.0, "audit/embedding/empirical_epsilon_details/0.05/epsilon": 0.0, "audit/embedding/empirical_epsilon_details/0.05/num_guesses": 0.0, "audit/embedding/empirical_epsilon_details/0.05/correct_guesses": 0.0, "audit/embedding/empirical_epsilon_details/0.01/epsilon": 0.0, "audit/embedding/empirical_epsilon_details/0.01/num_guesses": 0.0, "audit/embedding/empirical_epsilon_details/0.01/correct_guesses": 0.0, "perf/audit_duration_sec": 5.7471471559256315}}
|
| 15 |
+
{"timestamp": 1774098344.630848, "event": "train_step", "step": 110, "epoch": 2, "metrics": {"train/step_loss": 1.570878519940732, "train/step_real_loss": 1.2690251618623734, "train/lr": 9.920264990753837e-05, "train/step_canary_loss": 8.010416825612387, "perf/step_duration_sec": 17.30536743812263, "perf/samples_per_sec": 7.685476802243991, "perf/tokens_per_sec": 6186.173184867881, "perf/logical_batch_size": 133.0, "perf/logical_token_count": 107054.0, "perf/physical_batches": 18.0, "privacy/epsilon": 2.155895236118489, "system/cuda_memory_allocated_gb": 16.63217782974243, "system/cuda_max_memory_allocated_gb": 74.69534254074097}}
|
| 16 |
+
{"timestamp": 1774098520.7483177, "event": "train_step", "step": 120, "epoch": 2, "metrics": {"train/step_loss": 1.733701615437974, "train/step_real_loss": 1.2730829790234566, "train/lr": 8.333360798744496e-05, "train/step_canary_loss": 8.284722222222221, "perf/step_duration_sec": 18.148367216810584, "perf/samples_per_sec": 7.493787092539381, "perf/tokens_per_sec": 5505.233545608104, "perf/logical_batch_size": 136.0, "perf/logical_token_count": 99911.0, "perf/physical_batches": 18.0, "privacy/epsilon": 2.2538737991591966, "system/cuda_memory_allocated_gb": 16.916476249694824, "system/cuda_max_memory_allocated_gb": 74.69534254074097}}
|
| 17 |
+
{"timestamp": 1774098695.815589, "event": "train_step", "step": 130, "epoch": 2, "metrics": {"train/step_loss": 1.4695805564071194, "train/step_real_loss": 1.2108174487948418, "train/lr": 6.788751536089739e-05, "train/step_canary_loss": 9.75, "perf/step_duration_sec": 16.7934517133981, "perf/samples_per_sec": 7.860206600331495, "perf/tokens_per_sec": 6584.88807942771, "perf/logical_batch_size": 132.0, "perf/logical_token_count": 110583.0, "perf/physical_batches": 17.0, "privacy/epsilon": 2.34829265801524, "system/cuda_memory_allocated_gb": 16.537474155426025, "system/cuda_max_memory_allocated_gb": 74.69534254074097}}
|
| 18 |
+
{"timestamp": 1774098871.6479974, "event": "train_step", "step": 140, "epoch": 2, "metrics": {"train/step_loss": 1.409767813354958, "train/step_real_loss": 1.280211590230465, "train/lr": 5.325635332531864e-05, "train/step_canary_loss": 6.9375, "perf/step_duration_sec": 17.296851823106408, "perf/samples_per_sec": 7.515818562215838, "perf/tokens_per_sec": 5869.218343212228, "perf/logical_batch_size": 130.0, "perf/logical_token_count": 101519.0, "perf/physical_batches": 18.0, "privacy/epsilon": 2.4397183333948855, "system/cuda_memory_allocated_gb": 16.348063945770264, "system/cuda_max_memory_allocated_gb": 74.69534254074097}}
|
| 19 |
+
{"timestamp": 1774099047.329602, "event": "train_step", "step": 150, "epoch": 2, "metrics": {"train/step_loss": 1.470641034250041, "train/step_real_loss": 1.2892888560891151, "train/lr": 3.981142237826332e-05, "train/step_canary_loss": 9.208333969116211, "perf/step_duration_sec": 17.535365080460906, "perf/samples_per_sec": 7.4706172012334715, "perf/tokens_per_sec": 5992.347437184814, "perf/logical_batch_size": 131.0, "perf/logical_token_count": 105078.0, "perf/physical_batches": 17.0, "privacy/epsilon": 2.5283910517887054, "system/cuda_memory_allocated_gb": 16.442769050598145, "system/cuda_max_memory_allocated_gb": 74.69534254074097}}
|
| 20 |
+
{"timestamp": 1774099074.3730319, "event": "eval_step", "step": 150, "epoch": 2, "metrics": {"eval/loss": 0.842570746081999, "eval/duration_sec": 27.040043698623776}}
|
| 21 |
+
{"timestamp": 1774099249.8653634, "event": "train_step", "step": 160, "epoch": 2, "metrics": {"train/step_loss": 1.536541504646415, "train/step_real_loss": 1.2477270364761353, "train/lr": 2.789391958515183e-05, "train/step_canary_loss": 7.697916825612386, "perf/step_duration_sec": 17.38564824871719, "perf/samples_per_sec": 7.649987972683935, "perf/tokens_per_sec": 6056.662282222899, "perf/logical_batch_size": 133.0, "perf/logical_token_count": 105299.0, "perf/physical_batches": 18.0, "privacy/epsilon": 2.6145698431381854, "system/cuda_memory_allocated_gb": 16.63217782974243, "system/cuda_max_memory_allocated_gb": 74.69534540176392}}
|
| 22 |
+
{"timestamp": 1774099424.502899, "event": "train_step", "step": 170, "epoch": 2, "metrics": {"train/step_loss": 1.4997789441174223, "train/step_real_loss": 1.2844417318701744, "train/lr": 1.7806279893114875e-05, "train/step_canary_loss": 10.6875, "perf/step_duration_sec": 17.294984631240368, "perf/samples_per_sec": 7.574450211616343, "perf/tokens_per_sec": 5730.794337970551, "perf/logical_batch_size": 131.0, "perf/logical_token_count": 99114.0, "perf/physical_batches": 17.0, "privacy/epsilon": 2.6985357679843074, "system/cuda_memory_allocated_gb": 16.442769050598145, "system/cuda_max_memory_allocated_gb": 74.69534540176392}}
|
| 23 |
+
{"timestamp": 1774099602.7806635, "event": "train_step", "step": 180, "epoch": 2, "metrics": {"train/step_loss": 1.5966231734664351, "train/step_real_loss": 1.2403349727392197, "train/lr": 9.804501125681243e-06, "train/step_canary_loss": 8.111607415335518, "perf/step_duration_sec": 17.99831911176443, "perf/samples_per_sec": 7.445139691539982, "perf/tokens_per_sec": 5983.892125215344, "perf/logical_batch_size": 134.0, "perf/logical_token_count": 107700.0, "perf/physical_batches": 18.0, "privacy/epsilon": 2.780402267783889, "system/cuda_memory_allocated_gb": 16.726882934570312, "system/cuda_max_memory_allocated_gb": 74.69534540176392}}
|
| 24 |
+
{"timestamp": 1774099786.8860722, "event": "train_step", "step": 190, "epoch": 2, "metrics": {"train/step_loss": 1.301622830904447, "train/step_real_loss": 1.246277093887329, "train/lr": 4.091647429802869e-06, "train/step_canary_loss": 4.84375, "perf/step_duration_sec": 17.514903953298926, "perf/samples_per_sec": 7.3651560033649455, "perf/tokens_per_sec": 6243.596898480441, "perf/logical_batch_size": 129.0, "perf/logical_token_count": 109356.0, "perf/physical_batches": 18.0, "privacy/epsilon": 2.860377969759561, "system/cuda_memory_allocated_gb": 16.253358840942383, "system/cuda_max_memory_allocated_gb": 74.69534540176392}}
|
| 25 |
+
{"timestamp": 1774099963.4199688, "event": "train_step", "step": 200, "epoch": 2, "metrics": {"train/step_loss": 1.5310632604541201, "train/step_real_loss": 1.2371121123433113, "train/lr": 8.126960406835249e-07, "train/step_canary_loss": 10.9375, "perf/step_duration_sec": 16.793452875688672, "perf/samples_per_sec": 7.860206056319249, "perf/tokens_per_sec": 6343.96039865215, "perf/logical_batch_size": 132.0, "perf/logical_token_count": 106537.0, "perf/physical_batches": 17.0, "privacy/epsilon": 2.938565800812133, "system/cuda_memory_allocated_gb": 16.537474155426025, "system/cuda_max_memory_allocated_gb": 74.69534540176392}}
|
| 26 |
+
{"timestamp": 1774099990.6645026, "event": "eval_step", "step": 200, "epoch": 2, "metrics": {"eval/loss": 0.8425621900750303, "eval/duration_sec": 27.24077794700861}}
|
| 27 |
+
{"timestamp": 1774100151.6417675, "event": "train_epoch", "step": 208, "epoch": 2, "metrics": {"train/epoch_loss": 1.6631105757675884, "train/epoch_real_loss": 1.2492633356516678, "train/epoch_canary_loss": 9.74998009501868, "perf/epoch_duration_sec": 1886.0955602359027, "perf/epoch_samples_per_sec": 29.488961838732855, "perf/epoch_tokens_per_sec": 23327.959583605018, "perf/epoch_samples": 55619.0, "perf/epoch_tokens": 43998761.0, "system/cuda_epoch_peak_memory_gb": 74.69534540176392, "eval/loss": 0.8425623961221212, "eval/duration_sec": 27.067450830712914, "privacy/epsilon": 2.9999680995370417}}
|
| 28 |
+
{"timestamp": 1774100161.8989484, "event": "audit_epoch", "step": 208, "epoch": 2, "metrics": {"audit/delta": 1e-05, "audit/num_canaries": 500.0, "audit/num_members": 250.0, "audit/paper_guess_fraction": 0.2, "audit/paper_guess_steps": 20.0, "audit/loss/auc": 0.525808, "audit/loss/empirical_epsilon/0.05": 0.05073561053723097, "audit/loss/empirical_epsilon/0.01": 0.0, "audit/loss/empirical_epsilon_details/0.05/epsilon": 0.05073561053723097, "audit/loss/empirical_epsilon_details/0.05/num_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.05/correct_guesses": 60.0, "audit/loss/empirical_epsilon_details/0.01/epsilon": 0.0, "audit/loss/empirical_epsilon_details/0.01/num_guesses": 0.0, "audit/loss/empirical_epsilon_details/0.01/correct_guesses": 0.0, "audit/embedding/auc": 0.520928, "audit/embedding/empirical_epsilon/0.05": 0.0, "audit/embedding/empirical_epsilon/0.01": 0.0, "audit/embedding/empirical_epsilon_details/0.05/epsilon": 0.0, "audit/embedding/empirical_epsilon_details/0.05/num_guesses": 0.0, "audit/embedding/empirical_epsilon_details/0.05/correct_guesses": 0.0, "audit/embedding/empirical_epsilon_details/0.01/epsilon": 0.0, "audit/embedding/empirical_epsilon_details/0.01/num_guesses": 0.0, "audit/embedding/empirical_epsilon_details/0.01/correct_guesses": 0.0, "perf/audit_duration_sec": 5.841585950925946}}
|
| 29 |
+
{"timestamp": 1774100172.4001799, "event": "audit_final", "step": 208, "epoch": 2, "metrics": {"audit/delta": 1e-05, "audit/num_canaries": 500.0, "audit/num_members": 250.0, "audit/paper_guess_fraction": 0.2, "audit/paper_guess_steps": 20.0, "audit/loss/auc": 0.525808, "audit/loss/empirical_epsilon/0.05": 0.05073561053723097, "audit/loss/empirical_epsilon/0.01": 0.0, "audit/loss/empirical_epsilon_details/0.05/epsilon": 0.05073561053723097, "audit/loss/empirical_epsilon_details/0.05/num_guesses": 100.0, "audit/loss/empirical_epsilon_details/0.05/correct_guesses": 60.0, "audit/loss/empirical_epsilon_details/0.01/epsilon": 0.0, "audit/loss/empirical_epsilon_details/0.01/num_guesses": 0.0, "audit/loss/empirical_epsilon_details/0.01/correct_guesses": 0.0, "audit/embedding/auc": 0.520928, "audit/embedding/empirical_epsilon/0.05": 0.0, "audit/embedding/empirical_epsilon/0.01": 0.0, "audit/embedding/empirical_epsilon_details/0.05/epsilon": 0.0, "audit/embedding/empirical_epsilon_details/0.05/num_guesses": 0.0, "audit/embedding/empirical_epsilon_details/0.05/correct_guesses": 0.0, "audit/embedding/empirical_epsilon_details/0.01/epsilon": 0.0, "audit/embedding/empirical_epsilon_details/0.01/num_guesses": 0.0, "audit/embedding/empirical_epsilon_details/0.01/correct_guesses": 0.0}}
|
| 30 |
+
{"timestamp": 1774100172.924808, "event": "energy_final", "step": 208, "epoch": null, "metrics": {"energy/codecarbon/duration": 3938.0607158355415, "energy/codecarbon/emissions": 0.5702959606126596, "energy/codecarbon/emissions_rate": 0.00014481644691749236, "energy/codecarbon/cpu_power": 179.92904019353898, "energy/codecarbon/gpu_power": 2298.550082608512, "energy/codecarbon/ram_power": 70.0, "energy/codecarbon/cpu_energy": 0.18975838183590477, "energy/codecarbon/gpu_energy": 2.514705143984827, "energy/codecarbon/ram_energy": 0.07382037755078118, "energy/codecarbon/energy_consumed": 2.778283903371513, "energy/codecarbon/water_consumed": 0.0, "energy/codecarbon/cpu_count": 224.0, "energy/codecarbon/gpu_count": 4.0, "energy/codecarbon/longitude": -121.9552, "energy/codecarbon/latitude": 37.3541, "energy/codecarbon/ram_total_size": 2015.5625190734863, "energy/codecarbon/cpu_utilization_percent": 2.4580389144905275, "energy/codecarbon/gpu_utilization_percent": 95.87682458386683, "energy/codecarbon/ram_utilization_percent": 2.36410650281618, "energy/codecarbon/ram_used_gb": 47.93522234837092, "energy/codecarbon/pue": 1.0, "energy/codecarbon/wue": 0.0}}
|
starcoder2-7b/dp3/pretrain_lm_head.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c9a9b40bb4ecc60cfc3a7e2f7cd31be2121d8f32f92d1a4ecdac554172fb4e5b
|
| 3 |
+
size 457594472
|
starcoder2-7b/dp3/resolved_config.yaml
ADDED
|
@@ -0,0 +1,101 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
name: bigcode/starcoder2-7b
|
| 3 |
+
tokenizer_name: bigcode/starcoder2-7b
|
| 4 |
+
max_length: 1024
|
| 5 |
+
dtype: bfloat16
|
| 6 |
+
trust_remote_code: true
|
| 7 |
+
use_fast_tokenizer: true
|
| 8 |
+
cache_dir: null
|
| 9 |
+
local_files_only: false
|
| 10 |
+
low_cpu_mem_usage: true
|
| 11 |
+
tie_word_embeddings: true
|
| 12 |
+
gradient_checkpointing: false
|
| 13 |
+
use_chat_template: false
|
| 14 |
+
dataset:
|
| 15 |
+
name: melihcatal/codedp-cpt
|
| 16 |
+
split: train
|
| 17 |
+
mode: cpt
|
| 18 |
+
text_column: text
|
| 19 |
+
validation_ratio: 0.05
|
| 20 |
+
max_samples: -1
|
| 21 |
+
lora:
|
| 22 |
+
enabled: true
|
| 23 |
+
r: 16
|
| 24 |
+
alpha: 32
|
| 25 |
+
dropout: 0.05
|
| 26 |
+
target_modules:
|
| 27 |
+
- q_proj
|
| 28 |
+
- k_proj
|
| 29 |
+
- v_proj
|
| 30 |
+
- o_proj
|
| 31 |
+
modules_to_save:
|
| 32 |
+
- lm_head
|
| 33 |
+
bias: none
|
| 34 |
+
training:
|
| 35 |
+
seed: 42
|
| 36 |
+
epochs: 2
|
| 37 |
+
warmup_steps: null
|
| 38 |
+
warmup_ratio: 0.05
|
| 39 |
+
mixed_precision: false
|
| 40 |
+
mixed_precision_dtype: bfloat16
|
| 41 |
+
batch_size: 8
|
| 42 |
+
eval_batch_size: 8
|
| 43 |
+
eval_every_steps: 50
|
| 44 |
+
eval_every_epochs: 1
|
| 45 |
+
learning_rate: 0.0002
|
| 46 |
+
optimizer: adamw
|
| 47 |
+
lr_scheduler: cosine
|
| 48 |
+
adam_beta1: 0.9
|
| 49 |
+
adam_beta2: 0.999
|
| 50 |
+
adam_epsilon: 1.0e-08
|
| 51 |
+
sgd_momentum: 0.9
|
| 52 |
+
weight_decay: 0.01
|
| 53 |
+
max_grad_norm: 1.0
|
| 54 |
+
log_every: 10
|
| 55 |
+
gradient_accumulation_steps: 16
|
| 56 |
+
num_workers: 4
|
| 57 |
+
output_dir: runs/cpt/starcoder2-7b/dp3
|
| 58 |
+
distributed:
|
| 59 |
+
strategy: dpddp
|
| 60 |
+
backend: nccl
|
| 61 |
+
devices: null
|
| 62 |
+
dp:
|
| 63 |
+
module_validator: auto
|
| 64 |
+
target_delta: 1.0e-05
|
| 65 |
+
noise_multiplier: null
|
| 66 |
+
max_grad_norm: 1.0
|
| 67 |
+
grad_sample_mode: hooks
|
| 68 |
+
clipping: flat
|
| 69 |
+
secure_mode: false
|
| 70 |
+
enabled: true
|
| 71 |
+
target_epsilon: 3.0
|
| 72 |
+
audit:
|
| 73 |
+
enabled: true
|
| 74 |
+
run_every_epoch: true
|
| 75 |
+
epoch_device: cuda
|
| 76 |
+
q_canary: auto
|
| 77 |
+
num_canaries: 500
|
| 78 |
+
prefix_length: 49
|
| 79 |
+
num_digits: 12
|
| 80 |
+
batch_size: 32
|
| 81 |
+
delta: 1.0e-05
|
| 82 |
+
p_values:
|
| 83 |
+
- 0.05
|
| 84 |
+
- 0.01
|
| 85 |
+
paper_guess_fraction: 0.2
|
| 86 |
+
paper_guess_steps: 20
|
| 87 |
+
enable_holdout_empirical_epsilon: false
|
| 88 |
+
holdout_seed: 42
|
| 89 |
+
tie_seed: 42
|
| 90 |
+
tracking:
|
| 91 |
+
enabled: true
|
| 92 |
+
tensorboard: true
|
| 93 |
+
wandb: false
|
| 94 |
+
wandb_project: codedp-finetune-h200-audit
|
| 95 |
+
wandb_run_name: starcoder2-7b-cpt-dp3
|
| 96 |
+
wandb_mode: online
|
| 97 |
+
codecarbon: true
|
| 98 |
+
codecarbon_output_file: codecarbon.csv
|
| 99 |
+
codecarbon_measure_power_secs: 15
|
| 100 |
+
codecarbon_country_iso_code: null
|
| 101 |
+
codecarbon_project_name: codedp-starcoder2-7b-cpt-dp3
|
starcoder2-7b/dp3/scalars.csv
ADDED
|
@@ -0,0 +1,386 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
timestamp,event,step,epoch,key,value
|
| 2 |
+
1774096500.8353336,train_step,10,1,train/step_loss,1.6323321043555417
|
| 3 |
+
1774096500.8353336,train_step,10,1,train/step_real_loss,1.330917976796627
|
| 4 |
+
1774096500.8353336,train_step,10,1,train/lr,0.00018181818181818183
|
| 5 |
+
1774096500.8353336,train_step,10,1,train/step_canary_loss,8.062500158945719
|
| 6 |
+
1774096500.8353336,train_step,10,1,perf/step_duration_sec,17.560116425156593
|
| 7 |
+
1774096500.8353336,train_step,10,1,perf/samples_per_sec,7.573981674145647
|
| 8 |
+
1774096500.8353336,train_step,10,1,perf/tokens_per_sec,6185.665138551686
|
| 9 |
+
1774096500.8353336,train_step,10,1,perf/logical_batch_size,133.0
|
| 10 |
+
1774096500.8353336,train_step,10,1,perf/logical_token_count,108621.0
|
| 11 |
+
1774096500.8353336,train_step,10,1,perf/physical_batches,18.0
|
| 12 |
+
1774096500.8353336,train_step,10,1,privacy/epsilon,0.7131647471248268
|
| 13 |
+
1774096500.8353336,train_step,10,1,system/cuda_memory_allocated_gb,16.63217782974243
|
| 14 |
+
1774096500.8353336,train_step,10,1,system/cuda_max_memory_allocated_gb,74.69534254074097
|
| 15 |
+
1774096676.393492,train_step,20,1,train/step_loss,1.826857232699429
|
| 16 |
+
1774096676.393492,train_step,20,1,train/step_real_loss,1.3395854756236076
|
| 17 |
+
1774096676.393492,train_step,20,1,train/lr,0.00019897180218885507
|
| 18 |
+
1774096676.393492,train_step,20,1,train/step_canary_loss,8.756944444444445
|
| 19 |
+
1774096676.393492,train_step,20,1,perf/step_duration_sec,18.843991773203015
|
| 20 |
+
1774096676.393492,train_step,20,1,perf/samples_per_sec,7.217154498729827
|
| 21 |
+
1774096676.393492,train_step,20,1,perf/tokens_per_sec,5434.1989336686165
|
| 22 |
+
1774096676.393492,train_step,20,1,perf/logical_batch_size,136.0
|
| 23 |
+
1774096676.393492,train_step,20,1,perf/logical_token_count,102402.0
|
| 24 |
+
1774096676.393492,train_step,20,1,perf/physical_batches,18.0
|
| 25 |
+
1774096676.393492,train_step,20,1,privacy/epsilon,0.9542480237478311
|
| 26 |
+
1774096676.393492,train_step,20,1,system/cuda_memory_allocated_gb,16.916476249694824
|
| 27 |
+
1774096676.393492,train_step,20,1,system/cuda_max_memory_allocated_gb,74.69534397125244
|
| 28 |
+
1774096852.7017646,train_step,30,1,train/step_loss,1.5258880311792546
|
| 29 |
+
1774096852.7017646,train_step,30,1,train/step_real_loss,1.3621462360024452
|
| 30 |
+
1774096852.7017646,train_step,30,1,train/lr,0.00019544467510209388
|
| 31 |
+
1774096852.7017646,train_step,30,1,train/step_canary_loss,6.765625476837158
|
| 32 |
+
1774096852.7017646,train_step,30,1,perf/step_duration_sec,18.006671600043774
|
| 33 |
+
1774096852.7017646,train_step,30,1,perf/samples_per_sec,7.275081309289916
|
| 34 |
+
1774096852.7017646,train_step,30,1,perf/tokens_per_sec,5874.1561099910805
|
| 35 |
+
1774096852.7017646,train_step,30,1,perf/logical_batch_size,131.0
|
| 36 |
+
1774096852.7017646,train_step,30,1,perf/logical_token_count,105774.0
|
| 37 |
+
1774096852.7017646,train_step,30,1,perf/physical_batches,18.0
|
| 38 |
+
1774096852.7017646,train_step,30,1,privacy/epsilon,1.145336977893831
|
| 39 |
+
1774096852.7017646,train_step,30,1,system/cuda_memory_allocated_gb,16.442769050598145
|
| 40 |
+
1774096852.7017646,train_step,30,1,system/cuda_max_memory_allocated_gb,74.69534397125244
|
| 41 |
+
1774097029.0227332,train_step,40,1,train/step_loss,1.6870967794347693
|
| 42 |
+
1774097029.0227332,train_step,40,1,train/step_real_loss,1.2007466033101082
|
| 43 |
+
1774097029.0227332,train_step,40,1,train/lr,0.00018949541262593762
|
| 44 |
+
1774097029.0227332,train_step,40,1,train/step_canary_loss,10.580357142857142
|
| 45 |
+
1774097029.0227332,train_step,40,1,perf/step_duration_sec,17.532358843833208
|
| 46 |
+
1774097029.0227332,train_step,40,1,perf/samples_per_sec,7.643010344106256
|
| 47 |
+
1774097029.0227332,train_step,40,1,perf/tokens_per_sec,6436.156195815627
|
| 48 |
+
1774097029.0227332,train_step,40,1,perf/logical_batch_size,134.0
|
| 49 |
+
1774097029.0227332,train_step,40,1,perf/logical_token_count,112841.0
|
| 50 |
+
1774097029.0227332,train_step,40,1,perf/physical_batches,18.0
|
| 51 |
+
1774097029.0227332,train_step,40,1,privacy/epsilon,1.311193108872294
|
| 52 |
+
1774097029.0227332,train_step,40,1,system/cuda_memory_allocated_gb,16.726882934570312
|
| 53 |
+
1774097029.0227332,train_step,40,1,system/cuda_max_memory_allocated_gb,74.69534683227539
|
| 54 |
+
1774097204.3534977,train_step,50,1,train/step_loss,1.5818433902881763
|
| 55 |
+
1774097204.3534977,train_step,50,1,train/step_real_loss,1.2425691708922386
|
| 56 |
+
1774097204.3534977,train_step,50,1,train/lr,0.00018127499143005268
|
| 57 |
+
1774097204.3534977,train_step,50,1,train/step_canary_loss,7.785714830671038
|
| 58 |
+
1774097204.3534977,train_step,50,1,perf/step_duration_sec,17.701514894142747
|
| 59 |
+
1774097204.3534977,train_step,50,1,perf/samples_per_sec,7.569973575783576
|
| 60 |
+
1774097204.3534977,train_step,50,1,perf/tokens_per_sec,6216.078152520672
|
| 61 |
+
1774097204.3534977,train_step,50,1,perf/logical_batch_size,134.0
|
| 62 |
+
1774097204.3534977,train_step,50,1,perf/logical_token_count,110034.0
|
| 63 |
+
1774097204.3534977,train_step,50,1,perf/physical_batches,18.0
|
| 64 |
+
1774097204.3534977,train_step,50,1,privacy/epsilon,1.4583641061524852
|
| 65 |
+
1774097204.3534977,train_step,50,1,system/cuda_memory_allocated_gb,16.726882934570312
|
| 66 |
+
1774097204.3534977,train_step,50,1,system/cuda_max_memory_allocated_gb,74.69534683227539
|
| 67 |
+
1774097231.3920577,eval_step,50,1,eval/loss,0.8423196817266531
|
| 68 |
+
1774097231.3920577,eval_step,50,1,eval/duration_sec,27.033981669694185
|
| 69 |
+
1774097408.5970793,train_step,60,1,train/step_loss,1.6164713336112804
|
| 70 |
+
1774097408.5970793,train_step,60,1,train/step_real_loss,1.3143803551793098
|
| 71 |
+
1774097408.5970793,train_step,60,1,train/lr,0.0001709920242324663
|
| 72 |
+
1774097408.5970793,train_step,60,1,train/step_canary_loss,9.350000381469727
|
| 73 |
+
1774097408.5970793,train_step,60,1,perf/step_duration_sec,17.616566620767117
|
| 74 |
+
1774097408.5970793,train_step,60,1,perf/samples_per_sec,7.549711749349289
|
| 75 |
+
1774097408.5970793,train_step,60,1,perf/tokens_per_sec,5881.736335493052
|
| 76 |
+
1774097408.5970793,train_step,60,1,perf/logical_batch_size,133.0
|
| 77 |
+
1774097408.5970793,train_step,60,1,perf/logical_token_count,103616.0
|
| 78 |
+
1774097408.5970793,train_step,60,1,perf/physical_batches,17.0
|
| 79 |
+
1774097408.5970793,train_step,60,1,privacy/epsilon,1.5929146459722179
|
| 80 |
+
1774097408.5970793,train_step,60,1,system/cuda_memory_allocated_gb,16.63217782974243
|
| 81 |
+
1774097408.5970793,train_step,60,1,system/cuda_max_memory_allocated_gb,74.69534969329834
|
| 82 |
+
1774097584.0989442,train_step,70,1,train/step_loss,1.5537053346633911
|
| 83 |
+
1774097584.0989442,train_step,70,1,train/step_real_loss,1.274133626371622
|
| 84 |
+
1774097584.0989442,train_step,70,1,train/lr,0.00015890746575622231
|
| 85 |
+
1774097584.0989442,train_step,70,1,train/step_canary_loss,10.5
|
| 86 |
+
1774097584.0989442,train_step,70,1,perf/step_duration_sec,17.34652972780168
|
| 87 |
+
1774097584.0989442,train_step,70,1,perf/samples_per_sec,7.609591201889827
|
| 88 |
+
1774097584.0989442,train_step,70,1,perf/tokens_per_sec,6165.613622913036
|
| 89 |
+
1774097584.0989442,train_step,70,1,perf/logical_batch_size,132.0
|
| 90 |
+
1774097584.0989442,train_step,70,1,perf/logical_token_count,106952.0
|
| 91 |
+
1774097584.0989442,train_step,70,1,perf/physical_batches,17.0
|
| 92 |
+
1774097584.0989442,train_step,70,1,privacy/epsilon,1.7182200620091768
|
| 93 |
+
1774097584.0989442,train_step,70,1,system/cuda_memory_allocated_gb,16.537474155426025
|
| 94 |
+
1774097584.0989442,train_step,70,1,system/cuda_max_memory_allocated_gb,74.69534969329834
|
| 95 |
+
1774097758.6071105,train_step,80,1,train/step_loss,1.6813924115403254
|
| 96 |
+
1774097758.6071105,train_step,80,1,train/step_real_loss,1.3261733427643776
|
| 97 |
+
1774097758.6071105,train_step,80,1,train/lr,0.00014532799038330385
|
| 98 |
+
1774097758.6071105,train_step,80,1,train/step_canary_loss,10.77500057220459
|
| 99 |
+
1774097758.6071105,train_step,80,1,perf/step_duration_sec,16.975525552406907
|
| 100 |
+
1774097758.6071105,train_step,80,1,perf/samples_per_sec,7.834808977748694
|
| 101 |
+
1774097758.6071105,train_step,80,1,perf/tokens_per_sec,6001.228043602786
|
| 102 |
+
1774097758.6071105,train_step,80,1,perf/logical_batch_size,133.0
|
| 103 |
+
1774097758.6071105,train_step,80,1,perf/logical_token_count,101874.0
|
| 104 |
+
1774097758.6071105,train_step,80,1,perf/physical_batches,17.0
|
| 105 |
+
1774097758.6071105,train_step,80,1,privacy/epsilon,1.8360751937320303
|
| 106 |
+
1774097758.6071105,train_step,80,1,system/cuda_memory_allocated_gb,16.63217782974243
|
| 107 |
+
1774097758.6071105,train_step,80,1,system/cuda_max_memory_allocated_gb,74.69534969329834
|
| 108 |
+
1774097934.2494926,train_step,90,1,train/step_loss,1.8558794460157408
|
| 109 |
+
1774097934.2494926,train_step,90,1,train/step_real_loss,1.2407655008137226
|
| 110 |
+
1774097934.2494926,train_step,90,1,train/lr,0.00013059820956358998
|
| 111 |
+
1774097934.2494926,train_step,90,1,train/step_canary_loss,10.604166666666666
|
| 112 |
+
1774097934.2494926,train_step,90,1,perf/step_duration_sec,17.994612561538815
|
| 113 |
+
1774097934.2494926,train_step,90,1,perf/samples_per_sec,7.557817626520207
|
| 114 |
+
1774097934.2494926,train_step,90,1,perf/tokens_per_sec,6203.967971981345
|
| 115 |
+
1774097934.2494926,train_step,90,1,perf/logical_batch_size,136.0
|
| 116 |
+
1774097934.2494926,train_step,90,1,perf/logical_token_count,111638.0
|
| 117 |
+
1774097934.2494926,train_step,90,1,perf/physical_batches,18.0
|
| 118 |
+
1774097934.2494926,train_step,90,1,privacy/epsilon,1.9476934830264792
|
| 119 |
+
1774097934.2494926,train_step,90,1,system/cuda_memory_allocated_gb,16.916476249694824
|
| 120 |
+
1774097934.2494926,train_step,90,1,system/cuda_max_memory_allocated_gb,74.69534969329834
|
| 121 |
+
1774098110.4647262,train_step,100,1,train/step_loss,1.86213939777319
|
| 122 |
+
1774098110.4647262,train_step,100,1,train/step_real_loss,1.2256366163492203
|
| 123 |
+
1774098110.4647262,train_step,100,1,train/lr,0.00011509192648058249
|
| 124 |
+
1774098110.4647262,train_step,100,1,train/step_canary_loss,10.009375
|
| 125 |
+
1774098110.4647262,train_step,100,1,perf/step_duration_sec,17.502219082787633
|
| 126 |
+
1774098110.4647262,train_step,100,1,perf/samples_per_sec,7.884714466619527
|
| 127 |
+
1774098110.4647262,train_step,100,1,perf/tokens_per_sec,6282.689039593839
|
| 128 |
+
1774098110.4647262,train_step,100,1,perf/logical_batch_size,138.0
|
| 129 |
+
1774098110.4647262,train_step,100,1,perf/logical_token_count,109961.0
|
| 130 |
+
1774098110.4647262,train_step,100,1,perf/physical_batches,18.0
|
| 131 |
+
1774098110.4647262,train_step,100,1,privacy/epsilon,2.054057754637441
|
| 132 |
+
1774098110.4647262,train_step,100,1,system/cuda_memory_allocated_gb,16.34807252883911
|
| 133 |
+
1774098110.4647262,train_step,100,1,system/cuda_max_memory_allocated_gb,74.69534969329834
|
| 134 |
+
1774098137.4743512,eval_step,100,1,eval/loss,0.8425039929905157
|
| 135 |
+
1774098137.4743512,eval_step,100,1,eval/duration_sec,27.006801065057516
|
| 136 |
+
1774098228.5487797,train_epoch,104,1,train/epoch_loss,1.7094427731604633
|
| 137 |
+
1774098228.5487797,train_epoch,104,1,train/epoch_real_loss,1.285309421313279
|
| 138 |
+
1774098228.5487797,train_epoch,104,1,train/epoch_canary_loss,9.719154772066274
|
| 139 |
+
1774098228.5487797,train_epoch,104,1,perf/epoch_duration_sec,1878.0538540109992
|
| 140 |
+
1774098228.5487797,train_epoch,104,1,perf/epoch_samples_per_sec,29.64824458099586
|
| 141 |
+
1774098228.5487797,train_epoch,104,1,perf/epoch_tokens_per_sec,23429.532069074678
|
| 142 |
+
1774098228.5487797,train_epoch,104,1,perf/epoch_samples,55681.0
|
| 143 |
+
1774098228.5487797,train_epoch,104,1,perf/epoch_tokens,44001923.0
|
| 144 |
+
1774098228.5487797,train_epoch,104,1,system/cuda_epoch_peak_memory_gb,74.69534969329834
|
| 145 |
+
1774098228.5487797,train_epoch,104,1,eval/loss,0.8425441932404178
|
| 146 |
+
1774098228.5487797,train_epoch,104,1,eval/duration_sec,26.999734753742814
|
| 147 |
+
1774098228.5487797,train_epoch,104,1,privacy/epsilon,2.0952814257505974
|
| 148 |
+
1774098238.4572744,audit_epoch,104,1,audit/delta,1e-05
|
| 149 |
+
1774098238.4572744,audit_epoch,104,1,audit/num_canaries,500.0
|
| 150 |
+
1774098238.4572744,audit_epoch,104,1,audit/num_members,250.0
|
| 151 |
+
1774098238.4572744,audit_epoch,104,1,audit/paper_guess_fraction,0.2
|
| 152 |
+
1774098238.4572744,audit_epoch,104,1,audit/paper_guess_steps,20.0
|
| 153 |
+
1774098238.4572744,audit_epoch,104,1,audit/loss/auc,0.527168
|
| 154 |
+
1774098238.4572744,audit_epoch,104,1,audit/loss/empirical_epsilon/0.05,0.09125839080661535
|
| 155 |
+
1774098238.4572744,audit_epoch,104,1,audit/loss/empirical_epsilon/0.01,0.0
|
| 156 |
+
1774098238.4572744,audit_epoch,104,1,audit/loss/empirical_epsilon_details/0.05/epsilon,0.09125839080661535
|
| 157 |
+
1774098238.4572744,audit_epoch,104,1,audit/loss/empirical_epsilon_details/0.05/num_guesses,100.0
|
| 158 |
+
1774098238.4572744,audit_epoch,104,1,audit/loss/empirical_epsilon_details/0.05/correct_guesses,61.0
|
| 159 |
+
1774098238.4572744,audit_epoch,104,1,audit/loss/empirical_epsilon_details/0.01/epsilon,0.0
|
| 160 |
+
1774098238.4572744,audit_epoch,104,1,audit/loss/empirical_epsilon_details/0.01/num_guesses,0.0
|
| 161 |
+
1774098238.4572744,audit_epoch,104,1,audit/loss/empirical_epsilon_details/0.01/correct_guesses,0.0
|
| 162 |
+
1774098238.4572744,audit_epoch,104,1,audit/embedding/auc,0.514792
|
| 163 |
+
1774098238.4572744,audit_epoch,104,1,audit/embedding/empirical_epsilon/0.05,0.0
|
| 164 |
+
1774098238.4572744,audit_epoch,104,1,audit/embedding/empirical_epsilon/0.01,0.0
|
| 165 |
+
1774098238.4572744,audit_epoch,104,1,audit/embedding/empirical_epsilon_details/0.05/epsilon,0.0
|
| 166 |
+
1774098238.4572744,audit_epoch,104,1,audit/embedding/empirical_epsilon_details/0.05/num_guesses,0.0
|
| 167 |
+
1774098238.4572744,audit_epoch,104,1,audit/embedding/empirical_epsilon_details/0.05/correct_guesses,0.0
|
| 168 |
+
1774098238.4572744,audit_epoch,104,1,audit/embedding/empirical_epsilon_details/0.01/epsilon,0.0
|
| 169 |
+
1774098238.4572744,audit_epoch,104,1,audit/embedding/empirical_epsilon_details/0.01/num_guesses,0.0
|
| 170 |
+
1774098238.4572744,audit_epoch,104,1,audit/embedding/empirical_epsilon_details/0.01/correct_guesses,0.0
|
| 171 |
+
1774098238.4572744,audit_epoch,104,1,perf/audit_duration_sec,5.7471471559256315
|
| 172 |
+
1774098344.630848,train_step,110,2,train/step_loss,1.570878519940732
|
| 173 |
+
1774098344.630848,train_step,110,2,train/step_real_loss,1.2690251618623734
|
| 174 |
+
1774098344.630848,train_step,110,2,train/lr,9.920264990753837e-05
|
| 175 |
+
1774098344.630848,train_step,110,2,train/step_canary_loss,8.010416825612387
|
| 176 |
+
1774098344.630848,train_step,110,2,perf/step_duration_sec,17.30536743812263
|
| 177 |
+
1774098344.630848,train_step,110,2,perf/samples_per_sec,7.685476802243991
|
| 178 |
+
1774098344.630848,train_step,110,2,perf/tokens_per_sec,6186.173184867881
|
| 179 |
+
1774098344.630848,train_step,110,2,perf/logical_batch_size,133.0
|
| 180 |
+
1774098344.630848,train_step,110,2,perf/logical_token_count,107054.0
|
| 181 |
+
1774098344.630848,train_step,110,2,perf/physical_batches,18.0
|
| 182 |
+
1774098344.630848,train_step,110,2,privacy/epsilon,2.155895236118489
|
| 183 |
+
1774098344.630848,train_step,110,2,system/cuda_memory_allocated_gb,16.63217782974243
|
| 184 |
+
1774098344.630848,train_step,110,2,system/cuda_max_memory_allocated_gb,74.69534254074097
|
| 185 |
+
1774098520.7483177,train_step,120,2,train/step_loss,1.733701615437974
|
| 186 |
+
1774098520.7483177,train_step,120,2,train/step_real_loss,1.2730829790234566
|
| 187 |
+
1774098520.7483177,train_step,120,2,train/lr,8.333360798744496e-05
|
| 188 |
+
1774098520.7483177,train_step,120,2,train/step_canary_loss,8.284722222222221
|
| 189 |
+
1774098520.7483177,train_step,120,2,perf/step_duration_sec,18.148367216810584
|
| 190 |
+
1774098520.7483177,train_step,120,2,perf/samples_per_sec,7.493787092539381
|
| 191 |
+
1774098520.7483177,train_step,120,2,perf/tokens_per_sec,5505.233545608104
|
| 192 |
+
1774098520.7483177,train_step,120,2,perf/logical_batch_size,136.0
|
| 193 |
+
1774098520.7483177,train_step,120,2,perf/logical_token_count,99911.0
|
| 194 |
+
1774098520.7483177,train_step,120,2,perf/physical_batches,18.0
|
| 195 |
+
1774098520.7483177,train_step,120,2,privacy/epsilon,2.2538737991591966
|
| 196 |
+
1774098520.7483177,train_step,120,2,system/cuda_memory_allocated_gb,16.916476249694824
|
| 197 |
+
1774098520.7483177,train_step,120,2,system/cuda_max_memory_allocated_gb,74.69534254074097
|
| 198 |
+
1774098695.815589,train_step,130,2,train/step_loss,1.4695805564071194
|
| 199 |
+
1774098695.815589,train_step,130,2,train/step_real_loss,1.2108174487948418
|
| 200 |
+
1774098695.815589,train_step,130,2,train/lr,6.788751536089739e-05
|
| 201 |
+
1774098695.815589,train_step,130,2,train/step_canary_loss,9.75
|
| 202 |
+
1774098695.815589,train_step,130,2,perf/step_duration_sec,16.7934517133981
|
| 203 |
+
1774098695.815589,train_step,130,2,perf/samples_per_sec,7.860206600331495
|
| 204 |
+
1774098695.815589,train_step,130,2,perf/tokens_per_sec,6584.88807942771
|
| 205 |
+
1774098695.815589,train_step,130,2,perf/logical_batch_size,132.0
|
| 206 |
+
1774098695.815589,train_step,130,2,perf/logical_token_count,110583.0
|
| 207 |
+
1774098695.815589,train_step,130,2,perf/physical_batches,17.0
|
| 208 |
+
1774098695.815589,train_step,130,2,privacy/epsilon,2.34829265801524
|
| 209 |
+
1774098695.815589,train_step,130,2,system/cuda_memory_allocated_gb,16.537474155426025
|
| 210 |
+
1774098695.815589,train_step,130,2,system/cuda_max_memory_allocated_gb,74.69534254074097
|
| 211 |
+
1774098871.6479974,train_step,140,2,train/step_loss,1.409767813354958
|
| 212 |
+
1774098871.6479974,train_step,140,2,train/step_real_loss,1.280211590230465
|
| 213 |
+
1774098871.6479974,train_step,140,2,train/lr,5.325635332531864e-05
|
| 214 |
+
1774098871.6479974,train_step,140,2,train/step_canary_loss,6.9375
|
| 215 |
+
1774098871.6479974,train_step,140,2,perf/step_duration_sec,17.296851823106408
|
| 216 |
+
1774098871.6479974,train_step,140,2,perf/samples_per_sec,7.515818562215838
|
| 217 |
+
1774098871.6479974,train_step,140,2,perf/tokens_per_sec,5869.218343212228
|
| 218 |
+
1774098871.6479974,train_step,140,2,perf/logical_batch_size,130.0
|
| 219 |
+
1774098871.6479974,train_step,140,2,perf/logical_token_count,101519.0
|
| 220 |
+
1774098871.6479974,train_step,140,2,perf/physical_batches,18.0
|
| 221 |
+
1774098871.6479974,train_step,140,2,privacy/epsilon,2.4397183333948855
|
| 222 |
+
1774098871.6479974,train_step,140,2,system/cuda_memory_allocated_gb,16.348063945770264
|
| 223 |
+
1774098871.6479974,train_step,140,2,system/cuda_max_memory_allocated_gb,74.69534254074097
|
| 224 |
+
1774099047.329602,train_step,150,2,train/step_loss,1.470641034250041
|
| 225 |
+
1774099047.329602,train_step,150,2,train/step_real_loss,1.2892888560891151
|
| 226 |
+
1774099047.329602,train_step,150,2,train/lr,3.981142237826332e-05
|
| 227 |
+
1774099047.329602,train_step,150,2,train/step_canary_loss,9.208333969116211
|
| 228 |
+
1774099047.329602,train_step,150,2,perf/step_duration_sec,17.535365080460906
|
| 229 |
+
1774099047.329602,train_step,150,2,perf/samples_per_sec,7.4706172012334715
|
| 230 |
+
1774099047.329602,train_step,150,2,perf/tokens_per_sec,5992.347437184814
|
| 231 |
+
1774099047.329602,train_step,150,2,perf/logical_batch_size,131.0
|
| 232 |
+
1774099047.329602,train_step,150,2,perf/logical_token_count,105078.0
|
| 233 |
+
1774099047.329602,train_step,150,2,perf/physical_batches,17.0
|
| 234 |
+
1774099047.329602,train_step,150,2,privacy/epsilon,2.5283910517887054
|
| 235 |
+
1774099047.329602,train_step,150,2,system/cuda_memory_allocated_gb,16.442769050598145
|
| 236 |
+
1774099047.329602,train_step,150,2,system/cuda_max_memory_allocated_gb,74.69534254074097
|
| 237 |
+
1774099074.3730319,eval_step,150,2,eval/loss,0.842570746081999
|
| 238 |
+
1774099074.3730319,eval_step,150,2,eval/duration_sec,27.040043698623776
|
| 239 |
+
1774099249.8653634,train_step,160,2,train/step_loss,1.536541504646415
|
| 240 |
+
1774099249.8653634,train_step,160,2,train/step_real_loss,1.2477270364761353
|
| 241 |
+
1774099249.8653634,train_step,160,2,train/lr,2.789391958515183e-05
|
| 242 |
+
1774099249.8653634,train_step,160,2,train/step_canary_loss,7.697916825612386
|
| 243 |
+
1774099249.8653634,train_step,160,2,perf/step_duration_sec,17.38564824871719
|
| 244 |
+
1774099249.8653634,train_step,160,2,perf/samples_per_sec,7.649987972683935
|
| 245 |
+
1774099249.8653634,train_step,160,2,perf/tokens_per_sec,6056.662282222899
|
| 246 |
+
1774099249.8653634,train_step,160,2,perf/logical_batch_size,133.0
|
| 247 |
+
1774099249.8653634,train_step,160,2,perf/logical_token_count,105299.0
|
| 248 |
+
1774099249.8653634,train_step,160,2,perf/physical_batches,18.0
|
| 249 |
+
1774099249.8653634,train_step,160,2,privacy/epsilon,2.6145698431381854
|
| 250 |
+
1774099249.8653634,train_step,160,2,system/cuda_memory_allocated_gb,16.63217782974243
|
| 251 |
+
1774099249.8653634,train_step,160,2,system/cuda_max_memory_allocated_gb,74.69534540176392
|
| 252 |
+
1774099424.502899,train_step,170,2,train/step_loss,1.4997789441174223
|
| 253 |
+
1774099424.502899,train_step,170,2,train/step_real_loss,1.2844417318701744
|
| 254 |
+
1774099424.502899,train_step,170,2,train/lr,1.7806279893114875e-05
|
| 255 |
+
1774099424.502899,train_step,170,2,train/step_canary_loss,10.6875
|
| 256 |
+
1774099424.502899,train_step,170,2,perf/step_duration_sec,17.294984631240368
|
| 257 |
+
1774099424.502899,train_step,170,2,perf/samples_per_sec,7.574450211616343
|
| 258 |
+
1774099424.502899,train_step,170,2,perf/tokens_per_sec,5730.794337970551
|
| 259 |
+
1774099424.502899,train_step,170,2,perf/logical_batch_size,131.0
|
| 260 |
+
1774099424.502899,train_step,170,2,perf/logical_token_count,99114.0
|
| 261 |
+
1774099424.502899,train_step,170,2,perf/physical_batches,17.0
|
| 262 |
+
1774099424.502899,train_step,170,2,privacy/epsilon,2.6985357679843074
|
| 263 |
+
1774099424.502899,train_step,170,2,system/cuda_memory_allocated_gb,16.442769050598145
|
| 264 |
+
1774099424.502899,train_step,170,2,system/cuda_max_memory_allocated_gb,74.69534540176392
|
| 265 |
+
1774099602.7806635,train_step,180,2,train/step_loss,1.5966231734664351
|
| 266 |
+
1774099602.7806635,train_step,180,2,train/step_real_loss,1.2403349727392197
|
| 267 |
+
1774099602.7806635,train_step,180,2,train/lr,9.804501125681243e-06
|
| 268 |
+
1774099602.7806635,train_step,180,2,train/step_canary_loss,8.111607415335518
|
| 269 |
+
1774099602.7806635,train_step,180,2,perf/step_duration_sec,17.99831911176443
|
| 270 |
+
1774099602.7806635,train_step,180,2,perf/samples_per_sec,7.445139691539982
|
| 271 |
+
1774099602.7806635,train_step,180,2,perf/tokens_per_sec,5983.892125215344
|
| 272 |
+
1774099602.7806635,train_step,180,2,perf/logical_batch_size,134.0
|
| 273 |
+
1774099602.7806635,train_step,180,2,perf/logical_token_count,107700.0
|
| 274 |
+
1774099602.7806635,train_step,180,2,perf/physical_batches,18.0
|
| 275 |
+
1774099602.7806635,train_step,180,2,privacy/epsilon,2.780402267783889
|
| 276 |
+
1774099602.7806635,train_step,180,2,system/cuda_memory_allocated_gb,16.726882934570312
|
| 277 |
+
1774099602.7806635,train_step,180,2,system/cuda_max_memory_allocated_gb,74.69534540176392
|
| 278 |
+
1774099786.8860722,train_step,190,2,train/step_loss,1.301622830904447
|
| 279 |
+
1774099786.8860722,train_step,190,2,train/step_real_loss,1.246277093887329
|
| 280 |
+
1774099786.8860722,train_step,190,2,train/lr,4.091647429802869e-06
|
| 281 |
+
1774099786.8860722,train_step,190,2,train/step_canary_loss,4.84375
|
| 282 |
+
1774099786.8860722,train_step,190,2,perf/step_duration_sec,17.514903953298926
|
| 283 |
+
1774099786.8860722,train_step,190,2,perf/samples_per_sec,7.3651560033649455
|
| 284 |
+
1774099786.8860722,train_step,190,2,perf/tokens_per_sec,6243.596898480441
|
| 285 |
+
1774099786.8860722,train_step,190,2,perf/logical_batch_size,129.0
|
| 286 |
+
1774099786.8860722,train_step,190,2,perf/logical_token_count,109356.0
|
| 287 |
+
1774099786.8860722,train_step,190,2,perf/physical_batches,18.0
|
| 288 |
+
1774099786.8860722,train_step,190,2,privacy/epsilon,2.860377969759561
|
| 289 |
+
1774099786.8860722,train_step,190,2,system/cuda_memory_allocated_gb,16.253358840942383
|
| 290 |
+
1774099786.8860722,train_step,190,2,system/cuda_max_memory_allocated_gb,74.69534540176392
|
| 291 |
+
1774099963.4199688,train_step,200,2,train/step_loss,1.5310632604541201
|
| 292 |
+
1774099963.4199688,train_step,200,2,train/step_real_loss,1.2371121123433113
|
| 293 |
+
1774099963.4199688,train_step,200,2,train/lr,8.126960406835249e-07
|
| 294 |
+
1774099963.4199688,train_step,200,2,train/step_canary_loss,10.9375
|
| 295 |
+
1774099963.4199688,train_step,200,2,perf/step_duration_sec,16.793452875688672
|
| 296 |
+
1774099963.4199688,train_step,200,2,perf/samples_per_sec,7.860206056319249
|
| 297 |
+
1774099963.4199688,train_step,200,2,perf/tokens_per_sec,6343.96039865215
|
| 298 |
+
1774099963.4199688,train_step,200,2,perf/logical_batch_size,132.0
|
| 299 |
+
1774099963.4199688,train_step,200,2,perf/logical_token_count,106537.0
|
| 300 |
+
1774099963.4199688,train_step,200,2,perf/physical_batches,17.0
|
| 301 |
+
1774099963.4199688,train_step,200,2,privacy/epsilon,2.938565800812133
|
| 302 |
+
1774099963.4199688,train_step,200,2,system/cuda_memory_allocated_gb,16.537474155426025
|
| 303 |
+
1774099963.4199688,train_step,200,2,system/cuda_max_memory_allocated_gb,74.69534540176392
|
| 304 |
+
1774099990.6645026,eval_step,200,2,eval/loss,0.8425621900750303
|
| 305 |
+
1774099990.6645026,eval_step,200,2,eval/duration_sec,27.24077794700861
|
| 306 |
+
1774100151.6417675,train_epoch,208,2,train/epoch_loss,1.6631105757675884
|
| 307 |
+
1774100151.6417675,train_epoch,208,2,train/epoch_real_loss,1.2492633356516678
|
| 308 |
+
1774100151.6417675,train_epoch,208,2,train/epoch_canary_loss,9.74998009501868
|
| 309 |
+
1774100151.6417675,train_epoch,208,2,perf/epoch_duration_sec,1886.0955602359027
|
| 310 |
+
1774100151.6417675,train_epoch,208,2,perf/epoch_samples_per_sec,29.488961838732855
|
| 311 |
+
1774100151.6417675,train_epoch,208,2,perf/epoch_tokens_per_sec,23327.959583605018
|
| 312 |
+
1774100151.6417675,train_epoch,208,2,perf/epoch_samples,55619.0
|
| 313 |
+
1774100151.6417675,train_epoch,208,2,perf/epoch_tokens,43998761.0
|
| 314 |
+
1774100151.6417675,train_epoch,208,2,system/cuda_epoch_peak_memory_gb,74.69534540176392
|
| 315 |
+
1774100151.6417675,train_epoch,208,2,eval/loss,0.8425623961221212
|
| 316 |
+
1774100151.6417675,train_epoch,208,2,eval/duration_sec,27.067450830712914
|
| 317 |
+
1774100151.6417675,train_epoch,208,2,privacy/epsilon,2.9999680995370417
|
| 318 |
+
1774100161.8989484,audit_epoch,208,2,audit/delta,1e-05
|
| 319 |
+
1774100161.8989484,audit_epoch,208,2,audit/num_canaries,500.0
|
| 320 |
+
1774100161.8989484,audit_epoch,208,2,audit/num_members,250.0
|
| 321 |
+
1774100161.8989484,audit_epoch,208,2,audit/paper_guess_fraction,0.2
|
| 322 |
+
1774100161.8989484,audit_epoch,208,2,audit/paper_guess_steps,20.0
|
| 323 |
+
1774100161.8989484,audit_epoch,208,2,audit/loss/auc,0.525808
|
| 324 |
+
1774100161.8989484,audit_epoch,208,2,audit/loss/empirical_epsilon/0.05,0.05073561053723097
|
| 325 |
+
1774100161.8989484,audit_epoch,208,2,audit/loss/empirical_epsilon/0.01,0.0
|
| 326 |
+
1774100161.8989484,audit_epoch,208,2,audit/loss/empirical_epsilon_details/0.05/epsilon,0.05073561053723097
|
| 327 |
+
1774100161.8989484,audit_epoch,208,2,audit/loss/empirical_epsilon_details/0.05/num_guesses,100.0
|
| 328 |
+
1774100161.8989484,audit_epoch,208,2,audit/loss/empirical_epsilon_details/0.05/correct_guesses,60.0
|
| 329 |
+
1774100161.8989484,audit_epoch,208,2,audit/loss/empirical_epsilon_details/0.01/epsilon,0.0
|
| 330 |
+
1774100161.8989484,audit_epoch,208,2,audit/loss/empirical_epsilon_details/0.01/num_guesses,0.0
|
| 331 |
+
1774100161.8989484,audit_epoch,208,2,audit/loss/empirical_epsilon_details/0.01/correct_guesses,0.0
|
| 332 |
+
1774100161.8989484,audit_epoch,208,2,audit/embedding/auc,0.520928
|
| 333 |
+
1774100161.8989484,audit_epoch,208,2,audit/embedding/empirical_epsilon/0.05,0.0
|
| 334 |
+
1774100161.8989484,audit_epoch,208,2,audit/embedding/empirical_epsilon/0.01,0.0
|
| 335 |
+
1774100161.8989484,audit_epoch,208,2,audit/embedding/empirical_epsilon_details/0.05/epsilon,0.0
|
| 336 |
+
1774100161.8989484,audit_epoch,208,2,audit/embedding/empirical_epsilon_details/0.05/num_guesses,0.0
|
| 337 |
+
1774100161.8989484,audit_epoch,208,2,audit/embedding/empirical_epsilon_details/0.05/correct_guesses,0.0
|
| 338 |
+
1774100161.8989484,audit_epoch,208,2,audit/embedding/empirical_epsilon_details/0.01/epsilon,0.0
|
| 339 |
+
1774100161.8989484,audit_epoch,208,2,audit/embedding/empirical_epsilon_details/0.01/num_guesses,0.0
|
| 340 |
+
1774100161.8989484,audit_epoch,208,2,audit/embedding/empirical_epsilon_details/0.01/correct_guesses,0.0
|
| 341 |
+
1774100161.8989484,audit_epoch,208,2,perf/audit_duration_sec,5.841585950925946
|
| 342 |
+
1774100172.4001799,audit_final,208,2,audit/delta,1e-05
|
| 343 |
+
1774100172.4001799,audit_final,208,2,audit/num_canaries,500.0
|
| 344 |
+
1774100172.4001799,audit_final,208,2,audit/num_members,250.0
|
| 345 |
+
1774100172.4001799,audit_final,208,2,audit/paper_guess_fraction,0.2
|
| 346 |
+
1774100172.4001799,audit_final,208,2,audit/paper_guess_steps,20.0
|
| 347 |
+
1774100172.4001799,audit_final,208,2,audit/loss/auc,0.525808
|
| 348 |
+
1774100172.4001799,audit_final,208,2,audit/loss/empirical_epsilon/0.05,0.05073561053723097
|
| 349 |
+
1774100172.4001799,audit_final,208,2,audit/loss/empirical_epsilon/0.01,0.0
|
| 350 |
+
1774100172.4001799,audit_final,208,2,audit/loss/empirical_epsilon_details/0.05/epsilon,0.05073561053723097
|
| 351 |
+
1774100172.4001799,audit_final,208,2,audit/loss/empirical_epsilon_details/0.05/num_guesses,100.0
|
| 352 |
+
1774100172.4001799,audit_final,208,2,audit/loss/empirical_epsilon_details/0.05/correct_guesses,60.0
|
| 353 |
+
1774100172.4001799,audit_final,208,2,audit/loss/empirical_epsilon_details/0.01/epsilon,0.0
|
| 354 |
+
1774100172.4001799,audit_final,208,2,audit/loss/empirical_epsilon_details/0.01/num_guesses,0.0
|
| 355 |
+
1774100172.4001799,audit_final,208,2,audit/loss/empirical_epsilon_details/0.01/correct_guesses,0.0
|
| 356 |
+
1774100172.4001799,audit_final,208,2,audit/embedding/auc,0.520928
|
| 357 |
+
1774100172.4001799,audit_final,208,2,audit/embedding/empirical_epsilon/0.05,0.0
|
| 358 |
+
1774100172.4001799,audit_final,208,2,audit/embedding/empirical_epsilon/0.01,0.0
|
| 359 |
+
1774100172.4001799,audit_final,208,2,audit/embedding/empirical_epsilon_details/0.05/epsilon,0.0
|
| 360 |
+
1774100172.4001799,audit_final,208,2,audit/embedding/empirical_epsilon_details/0.05/num_guesses,0.0
|
| 361 |
+
1774100172.4001799,audit_final,208,2,audit/embedding/empirical_epsilon_details/0.05/correct_guesses,0.0
|
| 362 |
+
1774100172.4001799,audit_final,208,2,audit/embedding/empirical_epsilon_details/0.01/epsilon,0.0
|
| 363 |
+
1774100172.4001799,audit_final,208,2,audit/embedding/empirical_epsilon_details/0.01/num_guesses,0.0
|
| 364 |
+
1774100172.4001799,audit_final,208,2,audit/embedding/empirical_epsilon_details/0.01/correct_guesses,0.0
|
| 365 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/duration,3938.0607158355415
|
| 366 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/emissions,0.5702959606126596
|
| 367 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/emissions_rate,0.00014481644691749236
|
| 368 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/cpu_power,179.92904019353898
|
| 369 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/gpu_power,2298.550082608512
|
| 370 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/ram_power,70.0
|
| 371 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/cpu_energy,0.18975838183590477
|
| 372 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/gpu_energy,2.514705143984827
|
| 373 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/ram_energy,0.07382037755078118
|
| 374 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/energy_consumed,2.778283903371513
|
| 375 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/water_consumed,0.0
|
| 376 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/cpu_count,224.0
|
| 377 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/gpu_count,4.0
|
| 378 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/longitude,-121.9552
|
| 379 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/latitude,37.3541
|
| 380 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/ram_total_size,2015.5625190734863
|
| 381 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/cpu_utilization_percent,2.4580389144905275
|
| 382 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/gpu_utilization_percent,95.87682458386683
|
| 383 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/ram_utilization_percent,2.36410650281618
|
| 384 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/ram_used_gb,47.93522234837092
|
| 385 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/pue,1.0
|
| 386 |
+
1774100172.924808,energy_final,208,,energy/codecarbon/wue,0.0
|
starcoder2-7b/dp3/summary.json
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"audit/delta": 1e-05,
|
| 3 |
+
"audit/embedding/auc": 0.520928,
|
| 4 |
+
"audit/embedding/empirical_epsilon/0.01": 0.0,
|
| 5 |
+
"audit/embedding/empirical_epsilon/0.05": 0.0,
|
| 6 |
+
"audit/embedding/empirical_epsilon_details/0.01/correct_guesses": 0.0,
|
| 7 |
+
"audit/embedding/empirical_epsilon_details/0.01/epsilon": 0.0,
|
| 8 |
+
"audit/embedding/empirical_epsilon_details/0.01/num_guesses": 0.0,
|
| 9 |
+
"audit/embedding/empirical_epsilon_details/0.05/correct_guesses": 0.0,
|
| 10 |
+
"audit/embedding/empirical_epsilon_details/0.05/epsilon": 0.0,
|
| 11 |
+
"audit/embedding/empirical_epsilon_details/0.05/num_guesses": 0.0,
|
| 12 |
+
"audit/loss/auc": 0.525808,
|
| 13 |
+
"audit/loss/empirical_epsilon/0.01": 0.0,
|
| 14 |
+
"audit/loss/empirical_epsilon/0.05": 0.05073561053723097,
|
| 15 |
+
"audit/loss/empirical_epsilon_details/0.01/correct_guesses": 0.0,
|
| 16 |
+
"audit/loss/empirical_epsilon_details/0.01/epsilon": 0.0,
|
| 17 |
+
"audit/loss/empirical_epsilon_details/0.01/num_guesses": 0.0,
|
| 18 |
+
"audit/loss/empirical_epsilon_details/0.05/correct_guesses": 60.0,
|
| 19 |
+
"audit/loss/empirical_epsilon_details/0.05/epsilon": 0.05073561053723097,
|
| 20 |
+
"audit/loss/empirical_epsilon_details/0.05/num_guesses": 100.0,
|
| 21 |
+
"audit/num_canaries": 500.0,
|
| 22 |
+
"audit/num_members": 250.0,
|
| 23 |
+
"audit/paper_guess_fraction": 0.2,
|
| 24 |
+
"audit/paper_guess_steps": 20.0,
|
| 25 |
+
"energy/codecarbon/cpu_count": 224.0,
|
| 26 |
+
"energy/codecarbon/cpu_energy": 0.18975838183590477,
|
| 27 |
+
"energy/codecarbon/cpu_power": 179.92904019353898,
|
| 28 |
+
"energy/codecarbon/cpu_utilization_percent": 2.4580389144905275,
|
| 29 |
+
"energy/codecarbon/duration": 3938.0607158355415,
|
| 30 |
+
"energy/codecarbon/emissions": 0.5702959606126596,
|
| 31 |
+
"energy/codecarbon/emissions_rate": 0.00014481644691749236,
|
| 32 |
+
"energy/codecarbon/energy_consumed": 2.778283903371513,
|
| 33 |
+
"energy/codecarbon/gpu_count": 4.0,
|
| 34 |
+
"energy/codecarbon/gpu_energy": 2.514705143984827,
|
| 35 |
+
"energy/codecarbon/gpu_power": 2298.550082608512,
|
| 36 |
+
"energy/codecarbon/gpu_utilization_percent": 95.87682458386683,
|
| 37 |
+
"energy/codecarbon/latitude": 37.3541,
|
| 38 |
+
"energy/codecarbon/longitude": -121.9552,
|
| 39 |
+
"energy/codecarbon/pue": 1.0,
|
| 40 |
+
"energy/codecarbon/ram_energy": 0.07382037755078118,
|
| 41 |
+
"energy/codecarbon/ram_power": 70.0,
|
| 42 |
+
"energy/codecarbon/ram_total_size": 2015.5625190734863,
|
| 43 |
+
"energy/codecarbon/ram_used_gb": 47.93522234837092,
|
| 44 |
+
"energy/codecarbon/ram_utilization_percent": 2.36410650281618,
|
| 45 |
+
"energy/codecarbon/water_consumed": 0.0,
|
| 46 |
+
"energy/codecarbon/wue": 0.0,
|
| 47 |
+
"eval/duration_sec": 27.067450830712914,
|
| 48 |
+
"eval/loss": 0.8425623961221212,
|
| 49 |
+
"perf/audit_duration_sec": 5.841585950925946,
|
| 50 |
+
"perf/epoch_duration_sec": 1886.0955602359027,
|
| 51 |
+
"perf/epoch_samples": 55619.0,
|
| 52 |
+
"perf/epoch_samples_per_sec": 29.488961838732855,
|
| 53 |
+
"perf/epoch_tokens": 43998761.0,
|
| 54 |
+
"perf/epoch_tokens_per_sec": 23327.959583605018,
|
| 55 |
+
"perf/logical_batch_size": 132.0,
|
| 56 |
+
"perf/logical_token_count": 106537.0,
|
| 57 |
+
"perf/physical_batches": 17.0,
|
| 58 |
+
"perf/samples_per_sec": 7.860206056319249,
|
| 59 |
+
"perf/step_duration_sec": 16.793452875688672,
|
| 60 |
+
"perf/tokens_per_sec": 6343.96039865215,
|
| 61 |
+
"privacy/epsilon": 2.9999680995370417,
|
| 62 |
+
"system/cuda_epoch_peak_memory_gb": 74.69534540176392,
|
| 63 |
+
"system/cuda_max_memory_allocated_gb": 74.69534540176392,
|
| 64 |
+
"system/cuda_memory_allocated_gb": 16.537474155426025,
|
| 65 |
+
"train/epoch_canary_loss": 9.74998009501868,
|
| 66 |
+
"train/epoch_loss": 1.6631105757675884,
|
| 67 |
+
"train/epoch_real_loss": 1.2492633356516678,
|
| 68 |
+
"train/lr": 8.126960406835249e-07,
|
| 69 |
+
"train/step_canary_loss": 10.9375,
|
| 70 |
+
"train/step_loss": 1.5310632604541201,
|
| 71 |
+
"train/step_real_loss": 1.2371121123433113
|
| 72 |
+
}
|
starcoder2-7b/dp3/tensorboard/events.out.tfevents.1774096221.364c4f8de9dd.12837.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3b07c9c3ed310b74cc126d10e56bdacca0e2e7e9e9015a0a823e61040e77f436
|
| 3 |
+
size 26671
|
starcoder2-7b/dp3/tokenizer/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|