vijaym commited on
Commit
7acb24f
·
verified ·
1 Parent(s): 5abf953

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ checkpoint-2898/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ checkpoint-3220/tokenizer.json filter=lfs diff=lfs merge=lfs -text
38
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,296 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - ko
5
+ - en
6
+ tags:
7
+ - privacy-filter
8
+ - pii-detection
9
+ - token-classification
10
+ - korean
11
+ - lora
12
+ - openai-privacy-filter
13
+ - bioes
14
+ base_model: openai/privacy-filter
15
+ pipeline_tag: token-classification
16
+ ---
17
+
18
+ # Privacy Filter — Korean
19
+
20
+ Korean fine-tune of [OpenAI Privacy Filter](https://huggingface.co/openai/privacy-filter)
21
+ for span-level PII detection. Adapted via **LoRA** on attention projections only —
22
+ the base's sparse-MoE backbone (1.5B / 50M active params) is kept frozen, with
23
+ just **~614k trainable parameters** (~0.04% of the model).
24
+
25
+ **[Open Test Notebook](https://huggingface.co/FrameByFrame/privacy-filter-korean/blob/main/test_privacy_filter_ko.ipynb)** — load the model and run all examples interactively.
26
+
27
+ ## Capabilities
28
+
29
+ | Category | Description | Example |
30
+ |---|---|---|
31
+ | `private_person` | Personal name (Korean / Western / handles) | 김민수, John Smith |
32
+ | `private_address` | Physical / postal address | 서울특별시 강남구 테헤란로 123 |
33
+ | `private_phone` | Phone number | 010-1234-5678 |
34
+ | `private_email` | Email address | minsu@example.com |
35
+ | `private_date` | Birthday / personally-identifying date | 1985년 3월 12일 |
36
+ | `private_url` | Personal URL | github.com/minsu |
37
+ | `account_number` | Bank, card, RRN, passport, etc. | 110-234-567890 |
38
+ | `personal_handle` | Username / handle | @minsu_dev |
39
+ | `ip_address` | IP address | 192.168.1.5 |
40
+
41
+ ## Benchmark Results
42
+
43
+ Held-out KDPII Korean PII test set, span-level F1:
44
+
45
+ | label | base | fine-tuned | Δ |
46
+ |---|---|---|---|
47
+ | `private_phone` | 0.65 | **1.00** | +0.35 |
48
+ | `private_url` | 0.21 | **1.00** | +0.79 |
49
+ | `private_email` | 0.86 | **1.00** | +0.14 |
50
+ | `account_number` | 0.31 | **0.98** | +0.67 |
51
+ | `private_date` | 0.00 | **0.90** | +0.90 |
52
+ | `private_address` | 0.00 | **0.78** | +0.78 |
53
+ | `private_person` | 0.06 | **0.69** | +0.63 |
54
+ | **Overall** | — | — | **+0.58** |
55
+
56
+ ## Quick Start
57
+
58
+ ### Install
59
+
60
+ ```bash
61
+ pip install transformers peft torch
62
+ ```
63
+
64
+ ### Load Model
65
+
66
+ ```python
67
+ from transformers import AutoTokenizer, AutoModelForTokenClassification
68
+ import torch
69
+
70
+ MODEL_ID = "FrameByFrame/privacy-filter-korean"
71
+
72
+ tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, trust_remote_code=True)
73
+ model = AutoModelForTokenClassification.from_pretrained(
74
+ MODEL_ID, trust_remote_code=True, torch_dtype=torch.bfloat16
75
+ )
76
+ model.eval()
77
+ if torch.cuda.is_available():
78
+ model.cuda()
79
+ ```
80
+
81
+ `trust_remote_code=True` is required because Privacy Filter ships a custom
82
+ `OpenAIPrivacyFilterForTokenClassification` class (gpt-oss-style sparse MoE).
83
+
84
+ ### Inference
85
+
86
+ The model emits per-token BIOES labels. The helper below decodes them into
87
+ character-offset spans with simple constrained logic:
88
+
89
+ ```python
90
+ def extract_pii(text: str, max_length: int = 512):
91
+ enc = tokenizer(
92
+ text,
93
+ truncation=True,
94
+ max_length=max_length,
95
+ return_offsets_mapping=True,
96
+ return_tensors="pt",
97
+ )
98
+ offsets = enc.pop("offset_mapping")[0].tolist()
99
+ enc = {k: v.to(model.device) for k, v in enc.items()}
100
+ with torch.no_grad():
101
+ logits = model(**enc).logits
102
+ pred_ids = logits.argmax(-1)[0].tolist()
103
+ id2label = model.config.id2label
104
+
105
+ spans = []
106
+ active = None # (label, start, end)
107
+ for tok_idx, lid in enumerate(pred_ids):
108
+ label = id2label[int(lid)]
109
+ if label == "O":
110
+ if active is not None:
111
+ spans.append(active); active = None
112
+ continue
113
+ prefix, cat = label.split("-", 1)
114
+ c_start, c_end = offsets[tok_idx]
115
+ if prefix == "S":
116
+ if active is not None: spans.append(active); active = None
117
+ spans.append((cat, c_start, c_end))
118
+ elif prefix == "B":
119
+ if active is not None: spans.append(active)
120
+ active = (cat, c_start, c_end)
121
+ elif prefix in ("I", "E"):
122
+ if active and active[0] == cat:
123
+ active = (active[0], active[1], c_end)
124
+ else:
125
+ if active is not None: spans.append(active); active = None
126
+ if prefix == "E":
127
+ spans.append((cat, c_start, c_end))
128
+ if active is not None:
129
+ spans.append(active)
130
+
131
+ return [
132
+ {"label": cat, "start": s, "end": e, "text": text[s:e].strip()}
133
+ for cat, s, e in spans
134
+ if text[s:e].strip()
135
+ ]
136
+ ```
137
+
138
+ ### Test
139
+
140
+ #### Korean: name + phone + email
141
+ ```python
142
+ >>> extract_pii("김민수의 전화번호는 010-1234-5678이고 이메일은 minsu@example.com입니다.")
143
+ [
144
+ {"label": "private_person", "start": 0, "end": 3, "text": "김민수"},
145
+ {"label": "private_phone", "start": 12, "end": 25, "text": "010-1234-5678"},
146
+ {"label": "private_email", "start": 33, "end": 50, "text": "minsu@example.com"},
147
+ ]
148
+ ```
149
+
150
+ #### Korean: address + name
151
+ ```python
152
+ >>> extract_pii("서울특별시 강남구 테헤란로 123에 사는 박지영씨에게 연락주세요.")
153
+ [
154
+ {"label": "private_address", "start": 0, "end": 5, "text": "서울특별시"},
155
+ {"label": "private_address", "start": 6, "end": 9, "text": "강남구"},
156
+ {"label": "private_address", "start": 10, "end": 17, "text": "테헤란로 123"},
157
+ {"label": "private_person", "start": 22, "end": 25, "text": "박지영"},
158
+ ]
159
+ ```
160
+
161
+ > Note: the model follows KDPII's address convention where each toponym
162
+ > component is its own span. Most downstream redaction systems concatenate
163
+ > adjacent address spans.
164
+
165
+ #### Korean: form-style document
166
+ ```python
167
+ >>> extract_pii('''고객 정보
168
+ ... 이름: 이수진
169
+ ... 생년월일: 1985년 3월 12일
170
+ ... 주소: 부산광역시 해운대구 우동 1457
171
+ ... 연락처: 010-9876-5432''')
172
+ [
173
+ {"label": "private_person", ..., "text": "이수진"},
174
+ {"label": "private_date", ..., "text": "1985년 3월 12일"},
175
+ {"label": "private_address", ..., "text": "부산광역시"},
176
+ {"label": "private_address", ..., "text": "해운대구"},
177
+ {"label": "private_address", ..., "text": "우동 1457"},
178
+ {"label": "private_phone", ..., "text": "010-9876-5432"},
179
+ ]
180
+ ```
181
+
182
+ #### English: account + email
183
+ ```python
184
+ >>> extract_pii("Wire to acct 110-234-567890, contact minsu@example.com")
185
+ [
186
+ {"label": "account_number", "start": 13, "end": 26, "text": "110-234-567890"},
187
+ {"label": "private_email", "start": 36, "end": 53, "text": "minsu@example.com"},
188
+ ]
189
+ ```
190
+
191
+ ### Redaction
192
+
193
+ Wrap the spans into a redactor:
194
+
195
+ ```python
196
+ def redact(text: str, mask: str = "[REDACTED]") -> str:
197
+ spans = extract_pii(text)
198
+ spans.sort(key=lambda s: s["start"], reverse=True)
199
+ out = text
200
+ for s in spans:
201
+ out = out[: s["start"]] + f"[{s['label'].upper()}]" + out[s["end"]:]
202
+ return out
203
+
204
+ >>> redact("김민수님의 번호는 010-1234-5678입니다.")
205
+ "[PRIVATE_PERSON]님의 번호는 [PRIVATE_PHONE]입니다."
206
+ ```
207
+
208
+ ## Output Schema
209
+
210
+ Each detected entity is one dict:
211
+
212
+ | field | description |
213
+ |---|---|
214
+ | `label` | One of the 9 categories above |
215
+ | `start` | Character offset start (inclusive) |
216
+ | `end` | Character offset end (exclusive) |
217
+ | `text` | The matched substring |
218
+
219
+ ## Training Details
220
+
221
+ | | |
222
+ |---|---|
223
+ | **Base model** | `openai/privacy-filter` (sparse MoE, 1.5B total / 50M active params, 128 experts top-4) |
224
+ | **Method** | LoRA r=16, alpha=32, dropout=0.05 on attention projections (`q/k/v/o_proj`); classifier head fully trainable; everything else frozen |
225
+ | **Trainable params** | ~614k (~0.04% of the model) |
226
+ | **Datasets** | KDPII (Korean, ~53k records, deterministic 5/5/90 test/val/train), `korean_rrn_synthetic` (train only) |
227
+ | **Optimizer** | AdamW, lr=5e-4, cosine schedule, warmup 0.1 |
228
+ | **Batch** | 64 per device × 2 GPUs = 128 effective |
229
+ | **Epochs** | 10, early stopping on `eval_span_f1` (patience 3) |
230
+ | **Sequence length** | 512 |
231
+ | **Precision** | bf16 mixed (saved as bf16 safetensors after `merge_and_unload`) |
232
+ | **Hardware** | 2× NVIDIA RTX A5000 (24 GB each) |
233
+ | **Final eval span F1** | 0.848 (validation) |
234
+
235
+ For full reproduction details, see [`TRAINING.md`](./TRAINING.md).
236
+
237
+ ## Why MoE + LoRA
238
+
239
+ Full fine-tuning the privacy-filter base on KDPII consistently *hurt* the
240
+ weakest labels (`private_person` and `private_address` stuck at F1 ≈ 0.13–0.20).
241
+ With 128 experts and top-4 routing, Korean tokens hit a small expert subset;
242
+ across 5–10 epochs each expert receives sparse gradient updates relative to
243
+ its parameter count, and the optimizer drags those experts away from their
244
+ pretrained representations faster than it teaches the new task. Net effect:
245
+ the base's pretrained Korean capability gets corrupted before the new task is
246
+ learned.
247
+
248
+ LoRA on attention only (this model) avoids this entirely — experts, FFN,
249
+ embeddings, and router stay exactly as the base shipped them; only attention
250
+ re-routing and the classifier head adapt. Result: F1 0.69 / 0.78 on the
251
+ previously-stuck labels, with every other label at or above ceiling.
252
+
253
+ ## Known Limitations
254
+
255
+ - **`private_person` residual error** is dominated by KDPII's `PS_NICKNAME`
256
+ policy. ~40% of remaining person errors are online-handle-style strings
257
+ (e.g., `탕비실맥심킹`, `퍼터요정`) that KDPII labels as `PS_NICKNAME →
258
+ private_person`. Downstream redaction is unaffected; classification systems
259
+ may want to post-classify handles separately.
260
+ - **Foreign names** (Western, Japanese, Arabic transliterations) detected at
261
+ lower rates due to limited training exposure.
262
+ - **`private_address` boundaries** follow KDPII's split convention (each
263
+ toponym component is a separate span). Production redactors typically
264
+ concatenate adjacent address spans during post-processing.
265
+ - Raw model output may have leading/trailing whitespace in span offsets;
266
+ the `extract_pii` helper above strips them via `text.strip()` on the slice.
267
+
268
+ ## Serving with vLLM
269
+
270
+ For batched, low-latency inference:
271
+
272
+ ```bash
273
+ vllm serve FrameByFrame/privacy-filter-korean \
274
+ --task token-classification \
275
+ --max-model-len 512 \
276
+ --dtype bfloat16 \
277
+ --trust-remote-code
278
+ ```
279
+
280
+ ## License
281
+
282
+ Apache 2.0 (inherited from base
283
+ [OpenAI Privacy Filter](https://huggingface.co/openai/privacy-filter)).
284
+
285
+ ## Citation
286
+
287
+ If you use this model:
288
+
289
+ ```bibtex
290
+ @misc{framebyframe-privacy-filter-korean-2026,
291
+ title = {Privacy Filter Korean: LoRA fine-tune of OpenAI Privacy Filter for Korean PII},
292
+ author = {FrameByFrame},
293
+ year = {2026},
294
+ url = {https://huggingface.co/FrameByFrame/privacy-filter-korean}
295
+ }
296
+ ```
checkpoint-2898/README.md ADDED
@@ -0,0 +1,206 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: /models/privacy-filter
3
+ library_name: peft
4
+ tags:
5
+ - base_model:adapter:/models/privacy-filter
6
+ - lora
7
+ - transformers
8
+ ---
9
+
10
+ # Model Card for Model ID
11
+
12
+ <!-- Provide a quick summary of what the model is/does. -->
13
+
14
+
15
+
16
+ ## Model Details
17
+
18
+ ### Model Description
19
+
20
+ <!-- Provide a longer summary of what this model is. -->
21
+
22
+
23
+
24
+ - **Developed by:** [More Information Needed]
25
+ - **Funded by [optional]:** [More Information Needed]
26
+ - **Shared by [optional]:** [More Information Needed]
27
+ - **Model type:** [More Information Needed]
28
+ - **Language(s) (NLP):** [More Information Needed]
29
+ - **License:** [More Information Needed]
30
+ - **Finetuned from model [optional]:** [More Information Needed]
31
+
32
+ ### Model Sources [optional]
33
+
34
+ <!-- Provide the basic links for the model. -->
35
+
36
+ - **Repository:** [More Information Needed]
37
+ - **Paper [optional]:** [More Information Needed]
38
+ - **Demo [optional]:** [More Information Needed]
39
+
40
+ ## Uses
41
+
42
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
43
+
44
+ ### Direct Use
45
+
46
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
47
+
48
+ [More Information Needed]
49
+
50
+ ### Downstream Use [optional]
51
+
52
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
53
+
54
+ [More Information Needed]
55
+
56
+ ### Out-of-Scope Use
57
+
58
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
59
+
60
+ [More Information Needed]
61
+
62
+ ## Bias, Risks, and Limitations
63
+
64
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
65
+
66
+ [More Information Needed]
67
+
68
+ ### Recommendations
69
+
70
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
71
+
72
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
73
+
74
+ ## How to Get Started with the Model
75
+
76
+ Use the code below to get started with the model.
77
+
78
+ [More Information Needed]
79
+
80
+ ## Training Details
81
+
82
+ ### Training Data
83
+
84
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
85
+
86
+ [More Information Needed]
87
+
88
+ ### Training Procedure
89
+
90
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
91
+
92
+ #### Preprocessing [optional]
93
+
94
+ [More Information Needed]
95
+
96
+
97
+ #### Training Hyperparameters
98
+
99
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
100
+
101
+ #### Speeds, Sizes, Times [optional]
102
+
103
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
104
+
105
+ [More Information Needed]
106
+
107
+ ## Evaluation
108
+
109
+ <!-- This section describes the evaluation protocols and provides the results. -->
110
+
111
+ ### Testing Data, Factors & Metrics
112
+
113
+ #### Testing Data
114
+
115
+ <!-- This should link to a Dataset Card if possible. -->
116
+
117
+ [More Information Needed]
118
+
119
+ #### Factors
120
+
121
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
122
+
123
+ [More Information Needed]
124
+
125
+ #### Metrics
126
+
127
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
128
+
129
+ [More Information Needed]
130
+
131
+ ### Results
132
+
133
+ [More Information Needed]
134
+
135
+ #### Summary
136
+
137
+
138
+
139
+ ## Model Examination [optional]
140
+
141
+ <!-- Relevant interpretability work for the model goes here -->
142
+
143
+ [More Information Needed]
144
+
145
+ ## Environmental Impact
146
+
147
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
148
+
149
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
150
+
151
+ - **Hardware Type:** [More Information Needed]
152
+ - **Hours used:** [More Information Needed]
153
+ - **Cloud Provider:** [More Information Needed]
154
+ - **Compute Region:** [More Information Needed]
155
+ - **Carbon Emitted:** [More Information Needed]
156
+
157
+ ## Technical Specifications [optional]
158
+
159
+ ### Model Architecture and Objective
160
+
161
+ [More Information Needed]
162
+
163
+ ### Compute Infrastructure
164
+
165
+ [More Information Needed]
166
+
167
+ #### Hardware
168
+
169
+ [More Information Needed]
170
+
171
+ #### Software
172
+
173
+ [More Information Needed]
174
+
175
+ ## Citation [optional]
176
+
177
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
178
+
179
+ **BibTeX:**
180
+
181
+ [More Information Needed]
182
+
183
+ **APA:**
184
+
185
+ [More Information Needed]
186
+
187
+ ## Glossary [optional]
188
+
189
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
190
+
191
+ [More Information Needed]
192
+
193
+ ## More Information [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Authors [optional]
198
+
199
+ [More Information Needed]
200
+
201
+ ## Model Card Contact
202
+
203
+ [More Information Needed]
204
+ ### Framework versions
205
+
206
+ - PEFT 0.19.1
checkpoint-2898/adapter_config.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": null,
6
+ "base_model_name_or_path": "/models/privacy-filter",
7
+ "bias": "none",
8
+ "corda_config": null,
9
+ "ensure_weight_tying": false,
10
+ "eva_config": null,
11
+ "exclude_modules": null,
12
+ "fan_in_fan_out": false,
13
+ "inference_mode": true,
14
+ "init_lora_weights": true,
15
+ "layer_replication": null,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "loftq_config": {},
19
+ "lora_alpha": 32,
20
+ "lora_bias": false,
21
+ "lora_dropout": 0.05,
22
+ "lora_ga_config": null,
23
+ "megatron_config": null,
24
+ "megatron_core": "megatron.core",
25
+ "modules_to_save": [
26
+ "score",
27
+ "classifier",
28
+ "score"
29
+ ],
30
+ "peft_type": "LORA",
31
+ "peft_version": "0.19.1",
32
+ "qalora_group_size": 16,
33
+ "r": 16,
34
+ "rank_pattern": {},
35
+ "revision": null,
36
+ "target_modules": [
37
+ "o_proj",
38
+ "k_proj",
39
+ "v_proj",
40
+ "q_proj"
41
+ ],
42
+ "target_parameters": null,
43
+ "task_type": "TOKEN_CLS",
44
+ "trainable_token_indices": null,
45
+ "use_bdlora": null,
46
+ "use_dora": false,
47
+ "use_qalora": false,
48
+ "use_rslora": false
49
+ }
checkpoint-2898/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9c550f545e2e026d49a13fbf4ef8d15ae33e964d04565a0db49e04ee2000b1e
3
+ size 2415282
checkpoint-2898/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3b960d61f288020cac7e198e4c3e484dfc355f182a03c8a63cc7e9a5a8285e1
3
+ size 4870074
checkpoint-2898/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbbbe718f2a9271a2b86b705ff923bad5924435a994518a8ca80b88cc7e63027
3
+ size 14512
checkpoint-2898/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e454729420519392bd9e7ddb031517430e22c4c8e28d4063d8b44061c052bb7
3
+ size 14512
checkpoint-2898/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f9f8ffecc80842d5ea3bc822eb0af09a785a6d2b9310fa1410758a5de3c0ba3
3
+ size 1064
checkpoint-2898/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e714c627d94fd333b14f9ff32436219a4d7ac969719efe340fdc3385e1c7cd3e
3
+ size 27868272
checkpoint-2898/tokenizer_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "eos_token": "<|endoftext|>",
4
+ "is_local": true,
5
+ "local_files_only": false,
6
+ "model_input_names": [
7
+ "input_ids",
8
+ "attention_mask"
9
+ ],
10
+ "model_max_length": 128000,
11
+ "pad_token": "<|endoftext|>",
12
+ "tokenizer_class": "TokenizersBackend"
13
+ }
checkpoint-2898/trainer_state.json ADDED
@@ -0,0 +1,1386 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 2898,
3
+ "best_metric": 0.847424684159378,
4
+ "best_model_checkpoint": "/workspace/data/checkpoints/ko_pii_hf_ddp_v6_lora/checkpoint-2898",
5
+ "epoch": 9.0,
6
+ "eval_steps": 500,
7
+ "global_step": 2898,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.003105590062111801,
14
+ "grad_norm": 8.771296501159668,
15
+ "learning_rate": 0.0,
16
+ "loss": 1.4503676891326904,
17
+ "step": 1
18
+ },
19
+ {
20
+ "epoch": 0.07763975155279502,
21
+ "grad_norm": 8.387271881103516,
22
+ "learning_rate": 3.7267080745341614e-05,
23
+ "loss": 1.2439465522766113,
24
+ "step": 25
25
+ },
26
+ {
27
+ "epoch": 0.15527950310559005,
28
+ "grad_norm": 4.6111741065979,
29
+ "learning_rate": 7.608695652173914e-05,
30
+ "loss": 0.9780036926269531,
31
+ "step": 50
32
+ },
33
+ {
34
+ "epoch": 0.2329192546583851,
35
+ "grad_norm": 1.846106767654419,
36
+ "learning_rate": 0.00011490683229813665,
37
+ "loss": 0.5338097381591796,
38
+ "step": 75
39
+ },
40
+ {
41
+ "epoch": 0.3105590062111801,
42
+ "grad_norm": 2.225909471511841,
43
+ "learning_rate": 0.00015372670807453417,
44
+ "loss": 0.34235782623291017,
45
+ "step": 100
46
+ },
47
+ {
48
+ "epoch": 0.38819875776397517,
49
+ "grad_norm": 2.14986252784729,
50
+ "learning_rate": 0.0001925465838509317,
51
+ "loss": 0.2744290542602539,
52
+ "step": 125
53
+ },
54
+ {
55
+ "epoch": 0.4658385093167702,
56
+ "grad_norm": 2.1372387409210205,
57
+ "learning_rate": 0.0002313664596273292,
58
+ "loss": 0.2258339309692383,
59
+ "step": 150
60
+ },
61
+ {
62
+ "epoch": 0.5434782608695652,
63
+ "grad_norm": 1.5849782228469849,
64
+ "learning_rate": 0.0002701863354037267,
65
+ "loss": 0.2162160873413086,
66
+ "step": 175
67
+ },
68
+ {
69
+ "epoch": 0.6211180124223602,
70
+ "grad_norm": 1.6991852521896362,
71
+ "learning_rate": 0.00030900621118012424,
72
+ "loss": 0.20915340423583983,
73
+ "step": 200
74
+ },
75
+ {
76
+ "epoch": 0.6987577639751553,
77
+ "grad_norm": 2.723663568496704,
78
+ "learning_rate": 0.00034782608695652176,
79
+ "loss": 0.18431522369384765,
80
+ "step": 225
81
+ },
82
+ {
83
+ "epoch": 0.7763975155279503,
84
+ "grad_norm": 1.2743035554885864,
85
+ "learning_rate": 0.00038664596273291923,
86
+ "loss": 0.17637535095214843,
87
+ "step": 250
88
+ },
89
+ {
90
+ "epoch": 0.8540372670807453,
91
+ "grad_norm": 3.452497720718384,
92
+ "learning_rate": 0.0004254658385093168,
93
+ "loss": 0.16074188232421874,
94
+ "step": 275
95
+ },
96
+ {
97
+ "epoch": 0.9316770186335404,
98
+ "grad_norm": 2.547046661376953,
99
+ "learning_rate": 0.00046428571428571433,
100
+ "loss": 0.1531696319580078,
101
+ "step": 300
102
+ },
103
+ {
104
+ "epoch": 1.0,
105
+ "eval_class_account_number_f1": 0.71900826446281,
106
+ "eval_class_account_number_gold_spans": 113.0,
107
+ "eval_class_account_number_precision": 0.6744186046511628,
108
+ "eval_class_account_number_pred_spans": 129.0,
109
+ "eval_class_account_number_recall": 0.7699115044247787,
110
+ "eval_class_ip_address_f1": 1.0,
111
+ "eval_class_ip_address_gold_spans": 4.0,
112
+ "eval_class_ip_address_precision": 1.0,
113
+ "eval_class_ip_address_pred_spans": 4.0,
114
+ "eval_class_ip_address_recall": 1.0,
115
+ "eval_class_personal_handle_f1": 0.8070175438596492,
116
+ "eval_class_personal_handle_gold_spans": 28.0,
117
+ "eval_class_personal_handle_precision": 0.7931034482758621,
118
+ "eval_class_personal_handle_pred_spans": 29.0,
119
+ "eval_class_personal_handle_recall": 0.8214285714285714,
120
+ "eval_class_private_address_f1": 0.08450704225352113,
121
+ "eval_class_private_address_gold_spans": 48.0,
122
+ "eval_class_private_address_precision": 0.13043478260869565,
123
+ "eval_class_private_address_pred_spans": 23.0,
124
+ "eval_class_private_address_recall": 0.0625,
125
+ "eval_class_private_date_f1": 0.8985507246376813,
126
+ "eval_class_private_date_gold_spans": 33.0,
127
+ "eval_class_private_date_precision": 0.8611111111111112,
128
+ "eval_class_private_date_pred_spans": 36.0,
129
+ "eval_class_private_date_recall": 0.9393939393939394,
130
+ "eval_class_private_email_f1": 0.9620253164556962,
131
+ "eval_class_private_email_gold_spans": 39.0,
132
+ "eval_class_private_email_precision": 0.95,
133
+ "eval_class_private_email_pred_spans": 40.0,
134
+ "eval_class_private_email_recall": 0.9743589743589743,
135
+ "eval_class_private_person_f1": 0.32388663967611336,
136
+ "eval_class_private_person_gold_spans": 163.0,
137
+ "eval_class_private_person_precision": 0.47619047619047616,
138
+ "eval_class_private_person_pred_spans": 84.0,
139
+ "eval_class_private_person_recall": 0.24539877300613497,
140
+ "eval_class_private_phone_f1": 0.9855072463768116,
141
+ "eval_class_private_phone_gold_spans": 69.0,
142
+ "eval_class_private_phone_precision": 0.9855072463768116,
143
+ "eval_class_private_phone_pred_spans": 69.0,
144
+ "eval_class_private_phone_recall": 0.9855072463768116,
145
+ "eval_class_private_url_f1": 1.0,
146
+ "eval_class_private_url_gold_spans": 23.0,
147
+ "eval_class_private_url_precision": 1.0,
148
+ "eval_class_private_url_pred_spans": 23.0,
149
+ "eval_class_private_url_recall": 1.0,
150
+ "eval_gold_spans": 520.0,
151
+ "eval_loss": 0.1431565284729004,
152
+ "eval_pred_spans": 437.0,
153
+ "eval_runtime": 5.911,
154
+ "eval_samples_per_second": 376.753,
155
+ "eval_span_f1": 0.6624869383490073,
156
+ "eval_span_precision": 0.7254004576659039,
157
+ "eval_span_recall": 0.6096153846153847,
158
+ "eval_steps_per_second": 3.045,
159
+ "eval_token_accuracy": 0.9817635387488328,
160
+ "step": 322
161
+ },
162
+ {
163
+ "epoch": 1.0093167701863355,
164
+ "grad_norm": 1.0337207317352295,
165
+ "learning_rate": 0.000499999412412136,
166
+ "loss": 0.1343870735168457,
167
+ "step": 325
168
+ },
169
+ {
170
+ "epoch": 1.0869565217391304,
171
+ "grad_norm": 1.150405764579773,
172
+ "learning_rate": 0.0004998929197148332,
173
+ "loss": 0.1398656463623047,
174
+ "step": 350
175
+ },
176
+ {
177
+ "epoch": 1.1645962732919255,
178
+ "grad_norm": 1.239856243133545,
179
+ "learning_rate": 0.0004996028956208098,
180
+ "loss": 0.11556500434875489,
181
+ "step": 375
182
+ },
183
+ {
184
+ "epoch": 1.2422360248447206,
185
+ "grad_norm": 4.1494951248168945,
186
+ "learning_rate": 0.0004991295531354085,
187
+ "loss": 0.10652177810668945,
188
+ "step": 400
189
+ },
190
+ {
191
+ "epoch": 1.3198757763975155,
192
+ "grad_norm": 1.820974349975586,
193
+ "learning_rate": 0.0004984732399003618,
194
+ "loss": 0.1127833080291748,
195
+ "step": 425
196
+ },
197
+ {
198
+ "epoch": 1.3975155279503104,
199
+ "grad_norm": 0.9540935158729553,
200
+ "learning_rate": 0.0004976344379384687,
201
+ "loss": 0.1028218650817871,
202
+ "step": 450
203
+ },
204
+ {
205
+ "epoch": 1.4751552795031055,
206
+ "grad_norm": 1.7447997331619263,
207
+ "learning_rate": 0.0004966137632995793,
208
+ "loss": 0.11593814849853516,
209
+ "step": 475
210
+ },
211
+ {
212
+ "epoch": 1.5527950310559007,
213
+ "grad_norm": 1.3501027822494507,
214
+ "learning_rate": 0.0004954119656081417,
215
+ "loss": 0.1070643424987793,
216
+ "step": 500
217
+ },
218
+ {
219
+ "epoch": 1.6304347826086958,
220
+ "grad_norm": 0.6298630237579346,
221
+ "learning_rate": 0.0004940299275126486,
222
+ "loss": 0.11064315795898437,
223
+ "step": 525
224
+ },
225
+ {
226
+ "epoch": 1.7080745341614907,
227
+ "grad_norm": 0.8390702605247498,
228
+ "learning_rate": 0.0004924686640373841,
229
+ "loss": 0.09777620315551758,
230
+ "step": 550
231
+ },
232
+ {
233
+ "epoch": 1.7857142857142856,
234
+ "grad_norm": 0.5763865113258362,
235
+ "learning_rate": 0.0004907293218369499,
236
+ "loss": 0.08596687316894532,
237
+ "step": 575
238
+ },
239
+ {
240
+ "epoch": 1.8633540372670807,
241
+ "grad_norm": 0.7198231816291809,
242
+ "learning_rate": 0.0004888131783541152,
243
+ "loss": 0.0955750274658203,
244
+ "step": 600
245
+ },
246
+ {
247
+ "epoch": 1.9409937888198758,
248
+ "grad_norm": 1.0241382122039795,
249
+ "learning_rate": 0.0004867216408816122,
250
+ "loss": 0.0880333137512207,
251
+ "step": 625
252
+ },
253
+ {
254
+ "epoch": 2.0,
255
+ "eval_class_account_number_f1": 0.9184549356223176,
256
+ "eval_class_account_number_gold_spans": 113.0,
257
+ "eval_class_account_number_precision": 0.8916666666666667,
258
+ "eval_class_account_number_pred_spans": 120.0,
259
+ "eval_class_account_number_recall": 0.9469026548672567,
260
+ "eval_class_ip_address_f1": 1.0,
261
+ "eval_class_ip_address_gold_spans": 4.0,
262
+ "eval_class_ip_address_precision": 1.0,
263
+ "eval_class_ip_address_pred_spans": 4.0,
264
+ "eval_class_ip_address_recall": 1.0,
265
+ "eval_class_personal_handle_f1": 0.9310344827586207,
266
+ "eval_class_personal_handle_gold_spans": 28.0,
267
+ "eval_class_personal_handle_precision": 0.9,
268
+ "eval_class_personal_handle_pred_spans": 30.0,
269
+ "eval_class_personal_handle_recall": 0.9642857142857143,
270
+ "eval_class_private_address_f1": 0.5747126436781609,
271
+ "eval_class_private_address_gold_spans": 48.0,
272
+ "eval_class_private_address_precision": 0.6410256410256411,
273
+ "eval_class_private_address_pred_spans": 39.0,
274
+ "eval_class_private_address_recall": 0.5208333333333334,
275
+ "eval_class_private_date_f1": 0.927536231884058,
276
+ "eval_class_private_date_gold_spans": 33.0,
277
+ "eval_class_private_date_precision": 0.8888888888888888,
278
+ "eval_class_private_date_pred_spans": 36.0,
279
+ "eval_class_private_date_recall": 0.9696969696969697,
280
+ "eval_class_private_email_f1": 0.9743589743589743,
281
+ "eval_class_private_email_gold_spans": 39.0,
282
+ "eval_class_private_email_precision": 0.9743589743589743,
283
+ "eval_class_private_email_pred_spans": 39.0,
284
+ "eval_class_private_email_recall": 0.9743589743589743,
285
+ "eval_class_private_person_f1": 0.5182926829268292,
286
+ "eval_class_private_person_gold_spans": 163.0,
287
+ "eval_class_private_person_precision": 0.5151515151515151,
288
+ "eval_class_private_person_pred_spans": 165.0,
289
+ "eval_class_private_person_recall": 0.5214723926380368,
290
+ "eval_class_private_phone_f1": 1.0,
291
+ "eval_class_private_phone_gold_spans": 69.0,
292
+ "eval_class_private_phone_precision": 1.0,
293
+ "eval_class_private_phone_pred_spans": 69.0,
294
+ "eval_class_private_phone_recall": 1.0,
295
+ "eval_class_private_url_f1": 0.9130434782608695,
296
+ "eval_class_private_url_gold_spans": 23.0,
297
+ "eval_class_private_url_precision": 0.9130434782608695,
298
+ "eval_class_private_url_pred_spans": 23.0,
299
+ "eval_class_private_url_recall": 0.9130434782608695,
300
+ "eval_gold_spans": 520.0,
301
+ "eval_loss": 0.08562110364437103,
302
+ "eval_pred_spans": 525.0,
303
+ "eval_runtime": 6.7931,
304
+ "eval_samples_per_second": 327.832,
305
+ "eval_span_f1": 0.7808612440191388,
306
+ "eval_span_precision": 0.7771428571428571,
307
+ "eval_span_recall": 0.7846153846153846,
308
+ "eval_steps_per_second": 2.65,
309
+ "eval_token_accuracy": 0.9886788048552755,
310
+ "step": 644
311
+ },
312
+ {
313
+ "epoch": 2.018633540372671,
314
+ "grad_norm": 1.2737419605255127,
315
+ "learning_rate": 0.00048445624552856343,
316
+ "loss": 0.0896837329864502,
317
+ "step": 650
318
+ },
319
+ {
320
+ "epoch": 2.0962732919254656,
321
+ "grad_norm": 0.5823193788528442,
322
+ "learning_rate": 0.00048201865609230155,
323
+ "loss": 0.07990738868713379,
324
+ "step": 675
325
+ },
326
+ {
327
+ "epoch": 2.1739130434782608,
328
+ "grad_norm": 1.4650110006332397,
329
+ "learning_rate": 0.0004794106628364098,
330
+ "loss": 0.06952488899230957,
331
+ "step": 700
332
+ },
333
+ {
334
+ "epoch": 2.251552795031056,
335
+ "grad_norm": 0.6727014183998108,
336
+ "learning_rate": 0.0004766341811758803,
337
+ "loss": 0.07833251953125,
338
+ "step": 725
339
+ },
340
+ {
341
+ "epoch": 2.329192546583851,
342
+ "grad_norm": 0.6006216406822205,
343
+ "learning_rate": 0.000473691250270357,
344
+ "loss": 0.06939034461975098,
345
+ "step": 750
346
+ },
347
+ {
348
+ "epoch": 2.406832298136646,
349
+ "grad_norm": 0.5175172090530396,
350
+ "learning_rate": 0.00047058403152649405,
351
+ "loss": 0.05780394077301025,
352
+ "step": 775
353
+ },
354
+ {
355
+ "epoch": 2.4844720496894412,
356
+ "grad_norm": 1.1351436376571655,
357
+ "learning_rate": 0.00046731480701053265,
358
+ "loss": 0.06771457672119141,
359
+ "step": 800
360
+ },
361
+ {
362
+ "epoch": 2.562111801242236,
363
+ "grad_norm": 0.8665878772735596,
364
+ "learning_rate": 0.0004638859777722596,
365
+ "loss": 0.0703080177307129,
366
+ "step": 825
367
+ },
368
+ {
369
+ "epoch": 2.639751552795031,
370
+ "grad_norm": 0.8073828220367432,
371
+ "learning_rate": 0.00046030006208158,
372
+ "loss": 0.0754665756225586,
373
+ "step": 850
374
+ },
375
+ {
376
+ "epoch": 2.717391304347826,
377
+ "grad_norm": 0.6520082950592041,
378
+ "learning_rate": 0.00045655969357899875,
379
+ "loss": 0.06543315410614013,
380
+ "step": 875
381
+ },
382
+ {
383
+ "epoch": 2.795031055900621,
384
+ "grad_norm": 1.448807954788208,
385
+ "learning_rate": 0.00045266761934136896,
386
+ "loss": 0.07107664108276367,
387
+ "step": 900
388
+ },
389
+ {
390
+ "epoch": 2.8726708074534164,
391
+ "grad_norm": 0.45655015110969543,
392
+ "learning_rate": 0.00044862669786432865,
393
+ "loss": 0.07224170684814453,
394
+ "step": 925
395
+ },
396
+ {
397
+ "epoch": 2.950310559006211,
398
+ "grad_norm": 0.4825460612773895,
399
+ "learning_rate": 0.0004444398969629067,
400
+ "loss": 0.06763762950897217,
401
+ "step": 950
402
+ },
403
+ {
404
+ "epoch": 3.0,
405
+ "eval_class_account_number_f1": 0.9350649350649352,
406
+ "eval_class_account_number_gold_spans": 113.0,
407
+ "eval_class_account_number_precision": 0.9152542372881356,
408
+ "eval_class_account_number_pred_spans": 118.0,
409
+ "eval_class_account_number_recall": 0.9557522123893806,
410
+ "eval_class_ip_address_f1": 1.0,
411
+ "eval_class_ip_address_gold_spans": 4.0,
412
+ "eval_class_ip_address_precision": 1.0,
413
+ "eval_class_ip_address_pred_spans": 4.0,
414
+ "eval_class_ip_address_recall": 1.0,
415
+ "eval_class_personal_handle_f1": 0.912280701754386,
416
+ "eval_class_personal_handle_gold_spans": 28.0,
417
+ "eval_class_personal_handle_precision": 0.896551724137931,
418
+ "eval_class_personal_handle_pred_spans": 29.0,
419
+ "eval_class_personal_handle_recall": 0.9285714285714286,
420
+ "eval_class_private_address_f1": 0.6041666666666666,
421
+ "eval_class_private_address_gold_spans": 48.0,
422
+ "eval_class_private_address_precision": 0.6041666666666666,
423
+ "eval_class_private_address_pred_spans": 48.0,
424
+ "eval_class_private_address_recall": 0.6041666666666666,
425
+ "eval_class_private_date_f1": 0.9253731343283583,
426
+ "eval_class_private_date_gold_spans": 33.0,
427
+ "eval_class_private_date_precision": 0.9117647058823529,
428
+ "eval_class_private_date_pred_spans": 34.0,
429
+ "eval_class_private_date_recall": 0.9393939393939394,
430
+ "eval_class_private_email_f1": 0.9487179487179487,
431
+ "eval_class_private_email_gold_spans": 39.0,
432
+ "eval_class_private_email_precision": 0.9487179487179487,
433
+ "eval_class_private_email_pred_spans": 39.0,
434
+ "eval_class_private_email_recall": 0.9487179487179487,
435
+ "eval_class_private_person_f1": 0.6013071895424836,
436
+ "eval_class_private_person_gold_spans": 163.0,
437
+ "eval_class_private_person_precision": 0.6433566433566433,
438
+ "eval_class_private_person_pred_spans": 143.0,
439
+ "eval_class_private_person_recall": 0.5644171779141104,
440
+ "eval_class_private_phone_f1": 1.0,
441
+ "eval_class_private_phone_gold_spans": 69.0,
442
+ "eval_class_private_phone_precision": 1.0,
443
+ "eval_class_private_phone_pred_spans": 69.0,
444
+ "eval_class_private_phone_recall": 1.0,
445
+ "eval_class_private_url_f1": 0.9361702127659574,
446
+ "eval_class_private_url_gold_spans": 23.0,
447
+ "eval_class_private_url_precision": 0.9166666666666666,
448
+ "eval_class_private_url_pred_spans": 24.0,
449
+ "eval_class_private_url_recall": 0.9565217391304348,
450
+ "eval_gold_spans": 520.0,
451
+ "eval_loss": 0.07642154395580292,
452
+ "eval_pred_spans": 508.0,
453
+ "eval_runtime": 6.8956,
454
+ "eval_samples_per_second": 322.962,
455
+ "eval_span_f1": 0.8132295719844358,
456
+ "eval_span_precision": 0.8228346456692913,
457
+ "eval_span_recall": 0.8038461538461539,
458
+ "eval_steps_per_second": 2.61,
459
+ "eval_token_accuracy": 0.9904586834733894,
460
+ "step": 966
461
+ },
462
+ {
463
+ "epoch": 3.027950310559006,
464
+ "grad_norm": 0.3684696853160858,
465
+ "learning_rate": 0.00044011029159184015,
466
+ "loss": 0.04707308769226074,
467
+ "step": 975
468
+ },
469
+ {
470
+ "epoch": 3.1055900621118013,
471
+ "grad_norm": 0.47226452827453613,
472
+ "learning_rate": 0.00043564106158720383,
473
+ "loss": 0.04691335678100586,
474
+ "step": 1000
475
+ },
476
+ {
477
+ "epoch": 3.1832298136645965,
478
+ "grad_norm": 0.6678075194358826,
479
+ "learning_rate": 0.00043103548933101155,
480
+ "loss": 0.044693613052368165,
481
+ "step": 1025
482
+ },
483
+ {
484
+ "epoch": 3.260869565217391,
485
+ "grad_norm": 0.675778865814209,
486
+ "learning_rate": 0.00042629695734050255,
487
+ "loss": 0.04987412452697754,
488
+ "step": 1050
489
+ },
490
+ {
491
+ "epoch": 3.3385093167701863,
492
+ "grad_norm": 1.2661679983139038,
493
+ "learning_rate": 0.00042142894578388486,
494
+ "loss": 0.046450929641723634,
495
+ "step": 1075
496
+ },
497
+ {
498
+ "epoch": 3.4161490683229814,
499
+ "grad_norm": 0.7080872058868408,
500
+ "learning_rate": 0.00041643502992436013,
501
+ "loss": 0.050359611511230466,
502
+ "step": 1100
503
+ },
504
+ {
505
+ "epoch": 3.4937888198757765,
506
+ "grad_norm": 0.7029626369476318,
507
+ "learning_rate": 0.0004113188774943071,
508
+ "loss": 0.06056289672851563,
509
+ "step": 1125
510
+ },
511
+ {
512
+ "epoch": 3.571428571428571,
513
+ "grad_norm": 0.558189332485199,
514
+ "learning_rate": 0.000406084246001551,
515
+ "loss": 0.045447168350219724,
516
+ "step": 1150
517
+ },
518
+ {
519
+ "epoch": 3.6490683229813663,
520
+ "grad_norm": 0.5974974036216736,
521
+ "learning_rate": 0.00040073497996969933,
522
+ "loss": 0.045512027740478515,
523
+ "step": 1175
524
+ },
525
+ {
526
+ "epoch": 3.7267080745341614,
527
+ "grad_norm": 0.5707910656929016,
528
+ "learning_rate": 0.000395275008114569,
529
+ "loss": 0.05025949954986572,
530
+ "step": 1200
531
+ },
532
+ {
533
+ "epoch": 3.8043478260869565,
534
+ "grad_norm": 1.0855952501296997,
535
+ "learning_rate": 0.00038970834045878026,
536
+ "loss": 0.06197848796844482,
537
+ "step": 1225
538
+ },
539
+ {
540
+ "epoch": 3.8819875776397517,
541
+ "grad_norm": 0.3965921401977539,
542
+ "learning_rate": 0.0003840390653866341,
543
+ "loss": 0.05249831199645996,
544
+ "step": 1250
545
+ },
546
+ {
547
+ "epoch": 3.9596273291925463,
548
+ "grad_norm": 0.5562620162963867,
549
+ "learning_rate": 0.00037827134664143947,
550
+ "loss": 0.045794315338134765,
551
+ "step": 1275
552
+ },
553
+ {
554
+ "epoch": 4.0,
555
+ "eval_class_account_number_f1": 0.9691629955947135,
556
+ "eval_class_account_number_gold_spans": 113.0,
557
+ "eval_class_account_number_precision": 0.9649122807017544,
558
+ "eval_class_account_number_pred_spans": 114.0,
559
+ "eval_class_account_number_recall": 0.9734513274336283,
560
+ "eval_class_ip_address_f1": 1.0,
561
+ "eval_class_ip_address_gold_spans": 4.0,
562
+ "eval_class_ip_address_precision": 1.0,
563
+ "eval_class_ip_address_pred_spans": 4.0,
564
+ "eval_class_ip_address_recall": 1.0,
565
+ "eval_class_personal_handle_f1": 0.7719298245614034,
566
+ "eval_class_personal_handle_gold_spans": 28.0,
567
+ "eval_class_personal_handle_precision": 0.7586206896551724,
568
+ "eval_class_personal_handle_pred_spans": 29.0,
569
+ "eval_class_personal_handle_recall": 0.7857142857142857,
570
+ "eval_class_private_address_f1": 0.6067415730337078,
571
+ "eval_class_private_address_gold_spans": 48.0,
572
+ "eval_class_private_address_precision": 0.6585365853658537,
573
+ "eval_class_private_address_pred_spans": 41.0,
574
+ "eval_class_private_address_recall": 0.5625,
575
+ "eval_class_private_date_f1": 1.0,
576
+ "eval_class_private_date_gold_spans": 33.0,
577
+ "eval_class_private_date_precision": 1.0,
578
+ "eval_class_private_date_pred_spans": 33.0,
579
+ "eval_class_private_date_recall": 1.0,
580
+ "eval_class_private_email_f1": 0.9873417721518987,
581
+ "eval_class_private_email_gold_spans": 39.0,
582
+ "eval_class_private_email_precision": 0.975,
583
+ "eval_class_private_email_pred_spans": 40.0,
584
+ "eval_class_private_email_recall": 1.0,
585
+ "eval_class_private_person_f1": 0.6257668711656442,
586
+ "eval_class_private_person_gold_spans": 163.0,
587
+ "eval_class_private_person_precision": 0.6257668711656442,
588
+ "eval_class_private_person_pred_spans": 163.0,
589
+ "eval_class_private_person_recall": 0.6257668711656442,
590
+ "eval_class_private_phone_f1": 1.0,
591
+ "eval_class_private_phone_gold_spans": 69.0,
592
+ "eval_class_private_phone_precision": 1.0,
593
+ "eval_class_private_phone_pred_spans": 69.0,
594
+ "eval_class_private_phone_recall": 1.0,
595
+ "eval_class_private_url_f1": 0.9565217391304348,
596
+ "eval_class_private_url_gold_spans": 23.0,
597
+ "eval_class_private_url_precision": 0.9565217391304348,
598
+ "eval_class_private_url_pred_spans": 23.0,
599
+ "eval_class_private_url_recall": 0.9565217391304348,
600
+ "eval_gold_spans": 520.0,
601
+ "eval_loss": 0.07150626927614212,
602
+ "eval_pred_spans": 516.0,
603
+ "eval_runtime": 6.9918,
604
+ "eval_samples_per_second": 318.517,
605
+ "eval_span_f1": 0.8262548262548264,
606
+ "eval_span_precision": 0.8294573643410853,
607
+ "eval_span_recall": 0.823076923076923,
608
+ "eval_steps_per_second": 2.574,
609
+ "eval_token_accuracy": 0.9900501867413632,
610
+ "step": 1288
611
+ },
612
+ {
613
+ "epoch": 4.037267080745342,
614
+ "grad_norm": 0.3308059573173523,
615
+ "learning_rate": 0.00037240942026749233,
616
+ "loss": 0.03683102369308472,
617
+ "step": 1300
618
+ },
619
+ {
620
+ "epoch": 4.114906832298137,
621
+ "grad_norm": 0.9377778768539429,
622
+ "learning_rate": 0.0003664575914989552,
623
+ "loss": 0.033050749301910404,
624
+ "step": 1325
625
+ },
626
+ {
627
+ "epoch": 4.192546583850931,
628
+ "grad_norm": 0.5735832452774048,
629
+ "learning_rate": 0.00036042023159791974,
630
+ "loss": 0.031362690925598145,
631
+ "step": 1350
632
+ },
633
+ {
634
+ "epoch": 4.270186335403727,
635
+ "grad_norm": 0.631287693977356,
636
+ "learning_rate": 0.0003543017746439766,
637
+ "loss": 0.03477231979370117,
638
+ "step": 1375
639
+ },
640
+ {
641
+ "epoch": 4.3478260869565215,
642
+ "grad_norm": 1.0935797691345215,
643
+ "learning_rate": 0.0003481067142776489,
644
+ "loss": 0.03894762754440308,
645
+ "step": 1400
646
+ },
647
+ {
648
+ "epoch": 4.425465838509317,
649
+ "grad_norm": 0.6073715090751648,
650
+ "learning_rate": 0.0003418396004000829,
651
+ "loss": 0.03512312889099121,
652
+ "step": 1425
653
+ },
654
+ {
655
+ "epoch": 4.503105590062112,
656
+ "grad_norm": 0.6249358654022217,
657
+ "learning_rate": 0.0003355050358314172,
658
+ "loss": 0.04299846649169922,
659
+ "step": 1450
660
+ },
661
+ {
662
+ "epoch": 4.580745341614907,
663
+ "grad_norm": 0.7248182892799377,
664
+ "learning_rate": 0.0003291076729302869,
665
+ "loss": 0.03905656576156616,
666
+ "step": 1475
667
+ },
668
+ {
669
+ "epoch": 4.658385093167702,
670
+ "grad_norm": 0.9138495922088623,
671
+ "learning_rate": 0.00032265221017694474,
672
+ "loss": 0.039488759040832516,
673
+ "step": 1500
674
+ },
675
+ {
676
+ "epoch": 4.736024844720497,
677
+ "grad_norm": 0.6601109504699707,
678
+ "learning_rate": 0.00031614338872250835,
679
+ "loss": 0.03172698974609375,
680
+ "step": 1525
681
+ },
682
+ {
683
+ "epoch": 4.813664596273292,
684
+ "grad_norm": 1.141296625137329,
685
+ "learning_rate": 0.0003095859889068687,
686
+ "loss": 0.038228404521942136,
687
+ "step": 1550
688
+ },
689
+ {
690
+ "epoch": 4.891304347826087,
691
+ "grad_norm": 1.3895987272262573,
692
+ "learning_rate": 0.0003029848267478158,
693
+ "loss": 0.03626953125,
694
+ "step": 1575
695
+ },
696
+ {
697
+ "epoch": 4.9689440993788825,
698
+ "grad_norm": 0.5478597283363342,
699
+ "learning_rate": 0.00029634475040396285,
700
+ "loss": 0.03073249340057373,
701
+ "step": 1600
702
+ },
703
+ {
704
+ "epoch": 5.0,
705
+ "eval_class_account_number_f1": 0.9691629955947135,
706
+ "eval_class_account_number_gold_spans": 113.0,
707
+ "eval_class_account_number_precision": 0.9649122807017544,
708
+ "eval_class_account_number_pred_spans": 114.0,
709
+ "eval_class_account_number_recall": 0.9734513274336283,
710
+ "eval_class_ip_address_f1": 1.0,
711
+ "eval_class_ip_address_gold_spans": 4.0,
712
+ "eval_class_ip_address_precision": 1.0,
713
+ "eval_class_ip_address_pred_spans": 4.0,
714
+ "eval_class_ip_address_recall": 1.0,
715
+ "eval_class_personal_handle_f1": 0.8771929824561403,
716
+ "eval_class_personal_handle_gold_spans": 28.0,
717
+ "eval_class_personal_handle_precision": 0.8620689655172413,
718
+ "eval_class_personal_handle_pred_spans": 29.0,
719
+ "eval_class_personal_handle_recall": 0.8928571428571429,
720
+ "eval_class_private_address_f1": 0.6046511627906976,
721
+ "eval_class_private_address_gold_spans": 48.0,
722
+ "eval_class_private_address_precision": 0.6842105263157895,
723
+ "eval_class_private_address_pred_spans": 38.0,
724
+ "eval_class_private_address_recall": 0.5416666666666666,
725
+ "eval_class_private_date_f1": 1.0,
726
+ "eval_class_private_date_gold_spans": 33.0,
727
+ "eval_class_private_date_precision": 1.0,
728
+ "eval_class_private_date_pred_spans": 33.0,
729
+ "eval_class_private_date_recall": 1.0,
730
+ "eval_class_private_email_f1": 0.9873417721518987,
731
+ "eval_class_private_email_gold_spans": 39.0,
732
+ "eval_class_private_email_precision": 0.975,
733
+ "eval_class_private_email_pred_spans": 40.0,
734
+ "eval_class_private_email_recall": 1.0,
735
+ "eval_class_private_person_f1": 0.6027397260273972,
736
+ "eval_class_private_person_gold_spans": 163.0,
737
+ "eval_class_private_person_precision": 0.5445544554455446,
738
+ "eval_class_private_person_pred_spans": 202.0,
739
+ "eval_class_private_person_recall": 0.6748466257668712,
740
+ "eval_class_private_phone_f1": 1.0,
741
+ "eval_class_private_phone_gold_spans": 69.0,
742
+ "eval_class_private_phone_precision": 1.0,
743
+ "eval_class_private_phone_pred_spans": 69.0,
744
+ "eval_class_private_phone_recall": 1.0,
745
+ "eval_class_private_url_f1": 0.9565217391304348,
746
+ "eval_class_private_url_gold_spans": 23.0,
747
+ "eval_class_private_url_precision": 0.9565217391304348,
748
+ "eval_class_private_url_pred_spans": 23.0,
749
+ "eval_class_private_url_recall": 0.9565217391304348,
750
+ "eval_gold_spans": 520.0,
751
+ "eval_loss": 0.07776352018117905,
752
+ "eval_pred_spans": 552.0,
753
+ "eval_runtime": 7.1537,
754
+ "eval_samples_per_second": 311.307,
755
+ "eval_span_f1": 0.8171641791044777,
756
+ "eval_span_precision": 0.7934782608695652,
757
+ "eval_span_recall": 0.8423076923076923,
758
+ "eval_steps_per_second": 2.516,
759
+ "eval_token_accuracy": 0.9894082633053222,
760
+ "step": 1610
761
+ },
762
+ {
763
+ "epoch": 5.046583850931677,
764
+ "grad_norm": 0.8306529521942139,
765
+ "learning_rate": 0.00028967063661406285,
766
+ "loss": 0.02931797981262207,
767
+ "step": 1625
768
+ },
769
+ {
770
+ "epoch": 5.124223602484472,
771
+ "grad_norm": 0.3057140111923218,
772
+ "learning_rate": 0.00028296738711533655,
773
+ "loss": 0.02403106212615967,
774
+ "step": 1650
775
+ },
776
+ {
777
+ "epoch": 5.201863354037267,
778
+ "grad_norm": 0.6968570351600647,
779
+ "learning_rate": 0.0002762399250434392,
780
+ "loss": 0.024544227123260497,
781
+ "step": 1675
782
+ },
783
+ {
784
+ "epoch": 5.279503105590062,
785
+ "grad_norm": 0.8571075797080994,
786
+ "learning_rate": 0.0002694931913167124,
787
+ "loss": 0.021933317184448242,
788
+ "step": 1700
789
+ },
790
+ {
791
+ "epoch": 5.357142857142857,
792
+ "grad_norm": 0.5366156101226807,
793
+ "learning_rate": 0.00026273214100737516,
794
+ "loss": 0.031152501106262206,
795
+ "step": 1725
796
+ },
797
+ {
798
+ "epoch": 5.434782608695652,
799
+ "grad_norm": 0.5963289141654968,
800
+ "learning_rate": 0.0002559617397023193,
801
+ "loss": 0.02120798110961914,
802
+ "step": 1750
803
+ },
804
+ {
805
+ "epoch": 5.512422360248447,
806
+ "grad_norm": 0.36561551690101624,
807
+ "learning_rate": 0.00024918695985618394,
808
+ "loss": 0.023069217205047607,
809
+ "step": 1775
810
+ },
811
+ {
812
+ "epoch": 5.590062111801243,
813
+ "grad_norm": 0.834469199180603,
814
+ "learning_rate": 0.00024241277713938443,
815
+ "loss": 0.018797587156295776,
816
+ "step": 1800
817
+ },
818
+ {
819
+ "epoch": 5.667701863354037,
820
+ "grad_norm": 0.3487168848514557,
821
+ "learning_rate": 0.0002356441667837804,
822
+ "loss": 0.027416987419128416,
823
+ "step": 1825
824
+ },
825
+ {
826
+ "epoch": 5.745341614906832,
827
+ "grad_norm": 0.49782446026802063,
828
+ "learning_rate": 0.0002288860999286656,
829
+ "loss": 0.029009580612182617,
830
+ "step": 1850
831
+ },
832
+ {
833
+ "epoch": 5.8229813664596275,
834
+ "grad_norm": 0.5541417598724365,
835
+ "learning_rate": 0.00022214353996976349,
836
+ "loss": 0.02573819875717163,
837
+ "step": 1875
838
+ },
839
+ {
840
+ "epoch": 5.900621118012422,
841
+ "grad_norm": 0.9394005537033081,
842
+ "learning_rate": 0.00021542143891391025,
843
+ "loss": 0.024653725624084473,
844
+ "step": 1900
845
+ },
846
+ {
847
+ "epoch": 5.978260869565218,
848
+ "grad_norm": 0.9744410514831543,
849
+ "learning_rate": 0.0002087247337421019,
850
+ "loss": 0.028091182708740236,
851
+ "step": 1925
852
+ },
853
+ {
854
+ "epoch": 6.0,
855
+ "eval_class_account_number_f1": 0.9603524229074888,
856
+ "eval_class_account_number_gold_spans": 113.0,
857
+ "eval_class_account_number_precision": 0.956140350877193,
858
+ "eval_class_account_number_pred_spans": 114.0,
859
+ "eval_class_account_number_recall": 0.9646017699115044,
860
+ "eval_class_ip_address_f1": 1.0,
861
+ "eval_class_ip_address_gold_spans": 4.0,
862
+ "eval_class_ip_address_precision": 1.0,
863
+ "eval_class_ip_address_pred_spans": 4.0,
864
+ "eval_class_ip_address_recall": 1.0,
865
+ "eval_class_personal_handle_f1": 0.9090909090909091,
866
+ "eval_class_personal_handle_gold_spans": 28.0,
867
+ "eval_class_personal_handle_precision": 0.9259259259259259,
868
+ "eval_class_personal_handle_pred_spans": 27.0,
869
+ "eval_class_personal_handle_recall": 0.8928571428571429,
870
+ "eval_class_private_address_f1": 0.6813186813186812,
871
+ "eval_class_private_address_gold_spans": 48.0,
872
+ "eval_class_private_address_precision": 0.7209302325581395,
873
+ "eval_class_private_address_pred_spans": 43.0,
874
+ "eval_class_private_address_recall": 0.6458333333333334,
875
+ "eval_class_private_date_f1": 0.9393939393939394,
876
+ "eval_class_private_date_gold_spans": 33.0,
877
+ "eval_class_private_date_precision": 0.9393939393939394,
878
+ "eval_class_private_date_pred_spans": 33.0,
879
+ "eval_class_private_date_recall": 0.9393939393939394,
880
+ "eval_class_private_email_f1": 0.9620253164556962,
881
+ "eval_class_private_email_gold_spans": 39.0,
882
+ "eval_class_private_email_precision": 0.95,
883
+ "eval_class_private_email_pred_spans": 40.0,
884
+ "eval_class_private_email_recall": 0.9743589743589743,
885
+ "eval_class_private_person_f1": 0.6514657980456026,
886
+ "eval_class_private_person_gold_spans": 163.0,
887
+ "eval_class_private_person_precision": 0.6944444444444444,
888
+ "eval_class_private_person_pred_spans": 144.0,
889
+ "eval_class_private_person_recall": 0.6134969325153374,
890
+ "eval_class_private_phone_f1": 1.0,
891
+ "eval_class_private_phone_gold_spans": 69.0,
892
+ "eval_class_private_phone_precision": 1.0,
893
+ "eval_class_private_phone_pred_spans": 69.0,
894
+ "eval_class_private_phone_recall": 1.0,
895
+ "eval_class_private_url_f1": 0.9565217391304348,
896
+ "eval_class_private_url_gold_spans": 23.0,
897
+ "eval_class_private_url_precision": 0.9565217391304348,
898
+ "eval_class_private_url_pred_spans": 23.0,
899
+ "eval_class_private_url_recall": 0.9565217391304348,
900
+ "eval_gold_spans": 520.0,
901
+ "eval_loss": 0.07920802384614944,
902
+ "eval_pred_spans": 497.0,
903
+ "eval_runtime": 6.6716,
904
+ "eval_samples_per_second": 333.804,
905
+ "eval_span_f1": 0.8436578171091444,
906
+ "eval_span_precision": 0.8631790744466801,
907
+ "eval_span_recall": 0.825,
908
+ "eval_steps_per_second": 2.698,
909
+ "eval_token_accuracy": 0.9906921101774043,
910
+ "step": 1932
911
+ },
912
+ {
913
+ "epoch": 6.055900621118012,
914
+ "grad_norm": 0.5681946277618408,
915
+ "learning_rate": 0.00020205834278357742,
916
+ "loss": 0.01689502239227295,
917
+ "step": 1950
918
+ },
919
+ {
920
+ "epoch": 6.133540372670807,
921
+ "grad_norm": 0.3505108058452606,
922
+ "learning_rate": 0.00019542716210359964,
923
+ "loss": 0.015278645753860474,
924
+ "step": 1975
925
+ },
926
+ {
927
+ "epoch": 6.211180124223603,
928
+ "grad_norm": 0.3547140955924988,
929
+ "learning_rate": 0.00018883606190758903,
930
+ "loss": 0.015209052562713623,
931
+ "step": 2000
932
+ },
933
+ {
934
+ "epoch": 6.288819875776397,
935
+ "grad_norm": 0.3973138928413391,
936
+ "learning_rate": 0.00018228988296424876,
937
+ "loss": 0.01439091444015503,
938
+ "step": 2025
939
+ },
940
+ {
941
+ "epoch": 6.366459627329193,
942
+ "grad_norm": 0.5407560467720032,
943
+ "learning_rate": 0.00017579343305030985,
944
+ "loss": 0.014540103673934936,
945
+ "step": 2050
946
+ },
947
+ {
948
+ "epoch": 6.444099378881988,
949
+ "grad_norm": 0.3447468876838684,
950
+ "learning_rate": 0.00016935148341950694,
951
+ "loss": 0.017405078411102295,
952
+ "step": 2075
953
+ },
954
+ {
955
+ "epoch": 6.521739130434782,
956
+ "grad_norm": 0.30423322319984436,
957
+ "learning_rate": 0.0001629687652983776,
958
+ "loss": 0.015654172897338867,
959
+ "step": 2100
960
+ },
961
+ {
962
+ "epoch": 6.599378881987578,
963
+ "grad_norm": 0.47501835227012634,
964
+ "learning_rate": 0.0001566499664114601,
965
+ "loss": 0.017501153945922852,
966
+ "step": 2125
967
+ },
968
+ {
969
+ "epoch": 6.6770186335403725,
970
+ "grad_norm": 0.5283868312835693,
971
+ "learning_rate": 0.00015039972753843963,
972
+ "loss": 0.014991183280944825,
973
+ "step": 2150
974
+ },
975
+ {
976
+ "epoch": 6.754658385093168,
977
+ "grad_norm": 0.9889101982116699,
978
+ "learning_rate": 0.00014422263910577376,
979
+ "loss": 0.021359992027282716,
980
+ "step": 2175
981
+ },
982
+ {
983
+ "epoch": 6.832298136645963,
984
+ "grad_norm": 0.6929994225502014,
985
+ "learning_rate": 0.0001381232378152987,
986
+ "loss": 0.01291258454322815,
987
+ "step": 2200
988
+ },
989
+ {
990
+ "epoch": 6.909937888198757,
991
+ "grad_norm": 0.7863985896110535,
992
+ "learning_rate": 0.00013210600331229317,
993
+ "loss": 0.018135499954223634,
994
+ "step": 2225
995
+ },
996
+ {
997
+ "epoch": 6.987577639751553,
998
+ "grad_norm": 0.45059671998023987,
999
+ "learning_rate": 0.000126175354895447,
1000
+ "loss": 0.022398817539215087,
1001
+ "step": 2250
1002
+ },
1003
+ {
1004
+ "epoch": 7.0,
1005
+ "eval_class_account_number_f1": 0.9779735682819383,
1006
+ "eval_class_account_number_gold_spans": 113.0,
1007
+ "eval_class_account_number_precision": 0.9736842105263158,
1008
+ "eval_class_account_number_pred_spans": 114.0,
1009
+ "eval_class_account_number_recall": 0.9823008849557522,
1010
+ "eval_class_ip_address_f1": 1.0,
1011
+ "eval_class_ip_address_gold_spans": 4.0,
1012
+ "eval_class_ip_address_precision": 1.0,
1013
+ "eval_class_ip_address_pred_spans": 4.0,
1014
+ "eval_class_ip_address_recall": 1.0,
1015
+ "eval_class_personal_handle_f1": 0.9090909090909091,
1016
+ "eval_class_personal_handle_gold_spans": 28.0,
1017
+ "eval_class_personal_handle_precision": 0.9259259259259259,
1018
+ "eval_class_personal_handle_pred_spans": 27.0,
1019
+ "eval_class_personal_handle_recall": 0.8928571428571429,
1020
+ "eval_class_private_address_f1": 0.7333333333333334,
1021
+ "eval_class_private_address_gold_spans": 48.0,
1022
+ "eval_class_private_address_precision": 0.7857142857142857,
1023
+ "eval_class_private_address_pred_spans": 42.0,
1024
+ "eval_class_private_address_recall": 0.6875,
1025
+ "eval_class_private_date_f1": 1.0,
1026
+ "eval_class_private_date_gold_spans": 33.0,
1027
+ "eval_class_private_date_precision": 1.0,
1028
+ "eval_class_private_date_pred_spans": 33.0,
1029
+ "eval_class_private_date_recall": 1.0,
1030
+ "eval_class_private_email_f1": 0.9500000000000001,
1031
+ "eval_class_private_email_gold_spans": 39.0,
1032
+ "eval_class_private_email_precision": 0.926829268292683,
1033
+ "eval_class_private_email_pred_spans": 41.0,
1034
+ "eval_class_private_email_recall": 0.9743589743589743,
1035
+ "eval_class_private_person_f1": 0.6198083067092651,
1036
+ "eval_class_private_person_gold_spans": 163.0,
1037
+ "eval_class_private_person_precision": 0.6466666666666666,
1038
+ "eval_class_private_person_pred_spans": 150.0,
1039
+ "eval_class_private_person_recall": 0.5950920245398773,
1040
+ "eval_class_private_phone_f1": 1.0,
1041
+ "eval_class_private_phone_gold_spans": 69.0,
1042
+ "eval_class_private_phone_precision": 1.0,
1043
+ "eval_class_private_phone_pred_spans": 69.0,
1044
+ "eval_class_private_phone_recall": 1.0,
1045
+ "eval_class_private_url_f1": 0.9583333333333334,
1046
+ "eval_class_private_url_gold_spans": 23.0,
1047
+ "eval_class_private_url_precision": 0.92,
1048
+ "eval_class_private_url_pred_spans": 25.0,
1049
+ "eval_class_private_url_recall": 1.0,
1050
+ "eval_gold_spans": 520.0,
1051
+ "eval_loss": 0.08392658829689026,
1052
+ "eval_pred_spans": 505.0,
1053
+ "eval_runtime": 6.6789,
1054
+ "eval_samples_per_second": 333.44,
1055
+ "eval_span_f1": 0.8448780487804879,
1056
+ "eval_span_precision": 0.8574257425742574,
1057
+ "eval_span_recall": 0.8326923076923077,
1058
+ "eval_steps_per_second": 2.695,
1059
+ "eval_token_accuracy": 0.9905753968253969,
1060
+ "step": 2254
1061
+ },
1062
+ {
1063
+ "epoch": 7.065217391304348,
1064
+ "grad_norm": 0.24804139137268066,
1065
+ "learning_rate": 0.00012033564827115123,
1066
+ "loss": 0.013271838426589966,
1067
+ "step": 2275
1068
+ },
1069
+ {
1070
+ "epoch": 7.142857142857143,
1071
+ "grad_norm": 0.34166398644447327,
1072
+ "learning_rate": 0.0001145911723544914,
1073
+ "loss": 0.009375990629196166,
1074
+ "step": 2300
1075
+ },
1076
+ {
1077
+ "epoch": 7.220496894409938,
1078
+ "grad_norm": 0.8205583691596985,
1079
+ "learning_rate": 0.00010894614611929701,
1080
+ "loss": 0.014302024841308594,
1081
+ "step": 2325
1082
+ },
1083
+ {
1084
+ "epoch": 7.298136645962733,
1085
+ "grad_norm": 0.10090164840221405,
1086
+ "learning_rate": 0.00010340471549955708,
1087
+ "loss": 0.013667027950286865,
1088
+ "step": 2350
1089
+ },
1090
+ {
1091
+ "epoch": 7.375776397515528,
1092
+ "grad_norm": 0.38398560881614685,
1093
+ "learning_rate": 9.797095034447984e-05,
1094
+ "loss": 0.008717143535614013,
1095
+ "step": 2375
1096
+ },
1097
+ {
1098
+ "epoch": 7.453416149068323,
1099
+ "grad_norm": 0.4515543580055237,
1100
+ "learning_rate": 9.264884142943237e-05,
1101
+ "loss": 0.009974555373191833,
1102
+ "step": 2400
1103
+ },
1104
+ {
1105
+ "epoch": 7.5310559006211175,
1106
+ "grad_norm": 0.18586605787277222,
1107
+ "learning_rate": 8.744229752495425e-05,
1108
+ "loss": 0.011132798194885253,
1109
+ "step": 2425
1110
+ },
1111
+ {
1112
+ "epoch": 7.608695652173913,
1113
+ "grad_norm": 0.8527100682258606,
1114
+ "learning_rate": 8.235514252600023e-05,
1115
+ "loss": 0.013178889751434325,
1116
+ "step": 2450
1117
+ },
1118
+ {
1119
+ "epoch": 7.686335403726708,
1120
+ "grad_norm": 0.08932485431432724,
1121
+ "learning_rate": 7.739111264351752e-05,
1122
+ "loss": 0.008569929599761963,
1123
+ "step": 2475
1124
+ },
1125
+ {
1126
+ "epoch": 7.763975155279503,
1127
+ "grad_norm": 0.7609089612960815,
1128
+ "learning_rate": 7.255385366042274e-05,
1129
+ "loss": 0.01057699203491211,
1130
+ "step": 2500
1131
+ },
1132
+ {
1133
+ "epoch": 7.841614906832298,
1134
+ "grad_norm": 0.30278390645980835,
1135
+ "learning_rate": 6.784691825399267e-05,
1136
+ "loss": 0.007927331924438476,
1137
+ "step": 2525
1138
+ },
1139
+ {
1140
+ "epoch": 7.919254658385093,
1141
+ "grad_norm": 0.3382692039012909,
1142
+ "learning_rate": 6.327376338663493e-05,
1143
+ "loss": 0.012843213081359862,
1144
+ "step": 2550
1145
+ },
1146
+ {
1147
+ "epoch": 7.996894409937888,
1148
+ "grad_norm": 1.0424972772598267,
1149
+ "learning_rate": 5.883774776695652e-05,
1150
+ "loss": 0.010109683275222778,
1151
+ "step": 2575
1152
+ },
1153
+ {
1154
+ "epoch": 8.0,
1155
+ "eval_class_account_number_f1": 0.9779735682819383,
1156
+ "eval_class_account_number_gold_spans": 113.0,
1157
+ "eval_class_account_number_precision": 0.9736842105263158,
1158
+ "eval_class_account_number_pred_spans": 114.0,
1159
+ "eval_class_account_number_recall": 0.9823008849557522,
1160
+ "eval_class_ip_address_f1": 1.0,
1161
+ "eval_class_ip_address_gold_spans": 4.0,
1162
+ "eval_class_ip_address_precision": 1.0,
1163
+ "eval_class_ip_address_pred_spans": 4.0,
1164
+ "eval_class_ip_address_recall": 1.0,
1165
+ "eval_class_personal_handle_f1": 0.8214285714285714,
1166
+ "eval_class_personal_handle_gold_spans": 28.0,
1167
+ "eval_class_personal_handle_precision": 0.8214285714285714,
1168
+ "eval_class_personal_handle_pred_spans": 28.0,
1169
+ "eval_class_personal_handle_recall": 0.8214285714285714,
1170
+ "eval_class_private_address_f1": 0.6888888888888889,
1171
+ "eval_class_private_address_gold_spans": 48.0,
1172
+ "eval_class_private_address_precision": 0.7380952380952381,
1173
+ "eval_class_private_address_pred_spans": 42.0,
1174
+ "eval_class_private_address_recall": 0.6458333333333334,
1175
+ "eval_class_private_date_f1": 1.0,
1176
+ "eval_class_private_date_gold_spans": 33.0,
1177
+ "eval_class_private_date_precision": 1.0,
1178
+ "eval_class_private_date_pred_spans": 33.0,
1179
+ "eval_class_private_date_recall": 1.0,
1180
+ "eval_class_private_email_f1": 0.9500000000000001,
1181
+ "eval_class_private_email_gold_spans": 39.0,
1182
+ "eval_class_private_email_precision": 0.926829268292683,
1183
+ "eval_class_private_email_pred_spans": 41.0,
1184
+ "eval_class_private_email_recall": 0.9743589743589743,
1185
+ "eval_class_private_person_f1": 0.6230529595015576,
1186
+ "eval_class_private_person_gold_spans": 163.0,
1187
+ "eval_class_private_person_precision": 0.6329113924050633,
1188
+ "eval_class_private_person_pred_spans": 158.0,
1189
+ "eval_class_private_person_recall": 0.6134969325153374,
1190
+ "eval_class_private_phone_f1": 1.0,
1191
+ "eval_class_private_phone_gold_spans": 69.0,
1192
+ "eval_class_private_phone_precision": 1.0,
1193
+ "eval_class_private_phone_pred_spans": 69.0,
1194
+ "eval_class_private_phone_recall": 1.0,
1195
+ "eval_class_private_url_f1": 0.9583333333333334,
1196
+ "eval_class_private_url_gold_spans": 23.0,
1197
+ "eval_class_private_url_precision": 0.92,
1198
+ "eval_class_private_url_pred_spans": 25.0,
1199
+ "eval_class_private_url_recall": 1.0,
1200
+ "eval_gold_spans": 520.0,
1201
+ "eval_loss": 0.09201034158468246,
1202
+ "eval_pred_spans": 514.0,
1203
+ "eval_runtime": 6.9909,
1204
+ "eval_samples_per_second": 318.559,
1205
+ "eval_span_f1": 0.8355899419729208,
1206
+ "eval_span_precision": 0.8404669260700389,
1207
+ "eval_span_recall": 0.8307692307692308,
1208
+ "eval_steps_per_second": 2.575,
1209
+ "eval_token_accuracy": 0.9901960784313726,
1210
+ "step": 2576
1211
+ },
1212
+ {
1213
+ "epoch": 8.074534161490684,
1214
+ "grad_norm": 0.1411687433719635,
1215
+ "learning_rate": 5.454212938299255e-05,
1216
+ "loss": 0.009849590063095093,
1217
+ "step": 2600
1218
+ },
1219
+ {
1220
+ "epoch": 8.152173913043478,
1221
+ "grad_norm": 0.3560449182987213,
1222
+ "learning_rate": 5.0390063109409914e-05,
1223
+ "loss": 0.007954022884368896,
1224
+ "step": 2625
1225
+ },
1226
+ {
1227
+ "epoch": 8.229813664596273,
1228
+ "grad_norm": 0.36226364970207214,
1229
+ "learning_rate": 4.6384598390439706e-05,
1230
+ "loss": 0.005342819690704346,
1231
+ "step": 2650
1232
+ },
1233
+ {
1234
+ "epoch": 8.307453416149068,
1235
+ "grad_norm": 0.5297871232032776,
1236
+ "learning_rate": 4.252867700024374e-05,
1237
+ "loss": 0.00969659686088562,
1238
+ "step": 2675
1239
+ },
1240
+ {
1241
+ "epoch": 8.385093167701863,
1242
+ "grad_norm": 0.5881589651107788,
1243
+ "learning_rate": 3.882513088235676e-05,
1244
+ "loss": 0.006519293785095215,
1245
+ "step": 2700
1246
+ },
1247
+ {
1248
+ "epoch": 8.462732919254659,
1249
+ "grad_norm": 0.50627201795578,
1250
+ "learning_rate": 3.5276680069793705e-05,
1251
+ "loss": 0.00985595166683197,
1252
+ "step": 2725
1253
+ },
1254
+ {
1255
+ "epoch": 8.540372670807454,
1256
+ "grad_norm": 0.17046724259853363,
1257
+ "learning_rate": 3.18859306873483e-05,
1258
+ "loss": 0.004720044732093811,
1259
+ "step": 2750
1260
+ },
1261
+ {
1262
+ "epoch": 8.618012422360248,
1263
+ "grad_norm": 0.294168084859848,
1264
+ "learning_rate": 2.8655373037550058e-05,
1265
+ "loss": 0.011722609996795655,
1266
+ "step": 2775
1267
+ },
1268
+ {
1269
+ "epoch": 8.695652173913043,
1270
+ "grad_norm": 0.7373964190483093,
1271
+ "learning_rate": 2.5587379771686403e-05,
1272
+ "loss": 0.006650465726852417,
1273
+ "step": 2800
1274
+ },
1275
+ {
1276
+ "epoch": 8.773291925465838,
1277
+ "grad_norm": 0.4546540081501007,
1278
+ "learning_rate": 2.2684204147231934e-05,
1279
+ "loss": 0.01016199827194214,
1280
+ "step": 2825
1281
+ },
1282
+ {
1283
+ "epoch": 8.850931677018634,
1284
+ "grad_norm": 0.9172267317771912,
1285
+ "learning_rate": 1.9947978372965804e-05,
1286
+ "loss": 0.009012922048568725,
1287
+ "step": 2850
1288
+ },
1289
+ {
1290
+ "epoch": 8.928571428571429,
1291
+ "grad_norm": 0.29897835850715637,
1292
+ "learning_rate": 1.7380712042992184e-05,
1293
+ "loss": 0.005773799419403076,
1294
+ "step": 2875
1295
+ },
1296
+ {
1297
+ "epoch": 9.0,
1298
+ "eval_class_account_number_f1": 0.9779735682819383,
1299
+ "eval_class_account_number_gold_spans": 113.0,
1300
+ "eval_class_account_number_precision": 0.9736842105263158,
1301
+ "eval_class_account_number_pred_spans": 114.0,
1302
+ "eval_class_account_number_recall": 0.9823008849557522,
1303
+ "eval_class_ip_address_f1": 1.0,
1304
+ "eval_class_ip_address_gold_spans": 4.0,
1305
+ "eval_class_ip_address_precision": 1.0,
1306
+ "eval_class_ip_address_pred_spans": 4.0,
1307
+ "eval_class_ip_address_recall": 1.0,
1308
+ "eval_class_personal_handle_f1": 0.8928571428571429,
1309
+ "eval_class_personal_handle_gold_spans": 28.0,
1310
+ "eval_class_personal_handle_precision": 0.8928571428571429,
1311
+ "eval_class_personal_handle_pred_spans": 28.0,
1312
+ "eval_class_personal_handle_recall": 0.8928571428571429,
1313
+ "eval_class_private_address_f1": 0.7032967032967034,
1314
+ "eval_class_private_address_gold_spans": 48.0,
1315
+ "eval_class_private_address_precision": 0.7441860465116279,
1316
+ "eval_class_private_address_pred_spans": 43.0,
1317
+ "eval_class_private_address_recall": 0.6666666666666666,
1318
+ "eval_class_private_date_f1": 1.0,
1319
+ "eval_class_private_date_gold_spans": 33.0,
1320
+ "eval_class_private_date_precision": 1.0,
1321
+ "eval_class_private_date_pred_spans": 33.0,
1322
+ "eval_class_private_date_recall": 1.0,
1323
+ "eval_class_private_email_f1": 0.9500000000000001,
1324
+ "eval_class_private_email_gold_spans": 39.0,
1325
+ "eval_class_private_email_precision": 0.926829268292683,
1326
+ "eval_class_private_email_pred_spans": 41.0,
1327
+ "eval_class_private_email_recall": 0.9743589743589743,
1328
+ "eval_class_private_person_f1": 0.6412698412698412,
1329
+ "eval_class_private_person_gold_spans": 163.0,
1330
+ "eval_class_private_person_precision": 0.6644736842105263,
1331
+ "eval_class_private_person_pred_spans": 152.0,
1332
+ "eval_class_private_person_recall": 0.6196319018404908,
1333
+ "eval_class_private_phone_f1": 1.0,
1334
+ "eval_class_private_phone_gold_spans": 69.0,
1335
+ "eval_class_private_phone_precision": 1.0,
1336
+ "eval_class_private_phone_pred_spans": 69.0,
1337
+ "eval_class_private_phone_recall": 1.0,
1338
+ "eval_class_private_url_f1": 0.9583333333333334,
1339
+ "eval_class_private_url_gold_spans": 23.0,
1340
+ "eval_class_private_url_precision": 0.92,
1341
+ "eval_class_private_url_pred_spans": 25.0,
1342
+ "eval_class_private_url_recall": 1.0,
1343
+ "eval_gold_spans": 520.0,
1344
+ "eval_loss": 0.09796658158302307,
1345
+ "eval_pred_spans": 509.0,
1346
+ "eval_runtime": 7.2699,
1347
+ "eval_samples_per_second": 306.33,
1348
+ "eval_span_f1": 0.847424684159378,
1349
+ "eval_span_precision": 0.8565815324165029,
1350
+ "eval_span_recall": 0.8384615384615385,
1351
+ "eval_steps_per_second": 2.476,
1352
+ "eval_token_accuracy": 0.9907504668534081,
1353
+ "step": 2898
1354
+ }
1355
+ ],
1356
+ "logging_steps": 25,
1357
+ "max_steps": 3220,
1358
+ "num_input_tokens_seen": 0,
1359
+ "num_train_epochs": 10,
1360
+ "save_steps": 500,
1361
+ "stateful_callbacks": {
1362
+ "EarlyStoppingCallback": {
1363
+ "args": {
1364
+ "early_stopping_patience": 3,
1365
+ "early_stopping_threshold": 0.0
1366
+ },
1367
+ "attributes": {
1368
+ "early_stopping_patience_counter": 0
1369
+ }
1370
+ },
1371
+ "TrainerControl": {
1372
+ "args": {
1373
+ "should_epoch_stop": false,
1374
+ "should_evaluate": false,
1375
+ "should_log": false,
1376
+ "should_save": true,
1377
+ "should_training_stop": false
1378
+ },
1379
+ "attributes": {}
1380
+ }
1381
+ },
1382
+ "total_flos": 1.1622112111991194e+17,
1383
+ "train_batch_size": 64,
1384
+ "trial_name": null,
1385
+ "trial_params": null
1386
+ }
checkpoint-2898/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:195eed197af2b70c2fd47223db6e41715897a271a1bcbac65be2af77ec79752c
3
+ size 4920
checkpoint-3220/README.md ADDED
@@ -0,0 +1,206 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: /models/privacy-filter
3
+ library_name: peft
4
+ tags:
5
+ - base_model:adapter:/models/privacy-filter
6
+ - lora
7
+ - transformers
8
+ ---
9
+
10
+ # Model Card for Model ID
11
+
12
+ <!-- Provide a quick summary of what the model is/does. -->
13
+
14
+
15
+
16
+ ## Model Details
17
+
18
+ ### Model Description
19
+
20
+ <!-- Provide a longer summary of what this model is. -->
21
+
22
+
23
+
24
+ - **Developed by:** [More Information Needed]
25
+ - **Funded by [optional]:** [More Information Needed]
26
+ - **Shared by [optional]:** [More Information Needed]
27
+ - **Model type:** [More Information Needed]
28
+ - **Language(s) (NLP):** [More Information Needed]
29
+ - **License:** [More Information Needed]
30
+ - **Finetuned from model [optional]:** [More Information Needed]
31
+
32
+ ### Model Sources [optional]
33
+
34
+ <!-- Provide the basic links for the model. -->
35
+
36
+ - **Repository:** [More Information Needed]
37
+ - **Paper [optional]:** [More Information Needed]
38
+ - **Demo [optional]:** [More Information Needed]
39
+
40
+ ## Uses
41
+
42
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
43
+
44
+ ### Direct Use
45
+
46
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
47
+
48
+ [More Information Needed]
49
+
50
+ ### Downstream Use [optional]
51
+
52
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
53
+
54
+ [More Information Needed]
55
+
56
+ ### Out-of-Scope Use
57
+
58
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
59
+
60
+ [More Information Needed]
61
+
62
+ ## Bias, Risks, and Limitations
63
+
64
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
65
+
66
+ [More Information Needed]
67
+
68
+ ### Recommendations
69
+
70
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
71
+
72
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
73
+
74
+ ## How to Get Started with the Model
75
+
76
+ Use the code below to get started with the model.
77
+
78
+ [More Information Needed]
79
+
80
+ ## Training Details
81
+
82
+ ### Training Data
83
+
84
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
85
+
86
+ [More Information Needed]
87
+
88
+ ### Training Procedure
89
+
90
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
91
+
92
+ #### Preprocessing [optional]
93
+
94
+ [More Information Needed]
95
+
96
+
97
+ #### Training Hyperparameters
98
+
99
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
100
+
101
+ #### Speeds, Sizes, Times [optional]
102
+
103
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
104
+
105
+ [More Information Needed]
106
+
107
+ ## Evaluation
108
+
109
+ <!-- This section describes the evaluation protocols and provides the results. -->
110
+
111
+ ### Testing Data, Factors & Metrics
112
+
113
+ #### Testing Data
114
+
115
+ <!-- This should link to a Dataset Card if possible. -->
116
+
117
+ [More Information Needed]
118
+
119
+ #### Factors
120
+
121
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
122
+
123
+ [More Information Needed]
124
+
125
+ #### Metrics
126
+
127
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
128
+
129
+ [More Information Needed]
130
+
131
+ ### Results
132
+
133
+ [More Information Needed]
134
+
135
+ #### Summary
136
+
137
+
138
+
139
+ ## Model Examination [optional]
140
+
141
+ <!-- Relevant interpretability work for the model goes here -->
142
+
143
+ [More Information Needed]
144
+
145
+ ## Environmental Impact
146
+
147
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
148
+
149
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
150
+
151
+ - **Hardware Type:** [More Information Needed]
152
+ - **Hours used:** [More Information Needed]
153
+ - **Cloud Provider:** [More Information Needed]
154
+ - **Compute Region:** [More Information Needed]
155
+ - **Carbon Emitted:** [More Information Needed]
156
+
157
+ ## Technical Specifications [optional]
158
+
159
+ ### Model Architecture and Objective
160
+
161
+ [More Information Needed]
162
+
163
+ ### Compute Infrastructure
164
+
165
+ [More Information Needed]
166
+
167
+ #### Hardware
168
+
169
+ [More Information Needed]
170
+
171
+ #### Software
172
+
173
+ [More Information Needed]
174
+
175
+ ## Citation [optional]
176
+
177
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
178
+
179
+ **BibTeX:**
180
+
181
+ [More Information Needed]
182
+
183
+ **APA:**
184
+
185
+ [More Information Needed]
186
+
187
+ ## Glossary [optional]
188
+
189
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
190
+
191
+ [More Information Needed]
192
+
193
+ ## More Information [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Authors [optional]
198
+
199
+ [More Information Needed]
200
+
201
+ ## Model Card Contact
202
+
203
+ [More Information Needed]
204
+ ### Framework versions
205
+
206
+ - PEFT 0.19.1
checkpoint-3220/adapter_config.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": null,
6
+ "base_model_name_or_path": "/models/privacy-filter",
7
+ "bias": "none",
8
+ "corda_config": null,
9
+ "ensure_weight_tying": false,
10
+ "eva_config": null,
11
+ "exclude_modules": null,
12
+ "fan_in_fan_out": false,
13
+ "inference_mode": true,
14
+ "init_lora_weights": true,
15
+ "layer_replication": null,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "loftq_config": {},
19
+ "lora_alpha": 32,
20
+ "lora_bias": false,
21
+ "lora_dropout": 0.05,
22
+ "lora_ga_config": null,
23
+ "megatron_config": null,
24
+ "megatron_core": "megatron.core",
25
+ "modules_to_save": [
26
+ "score",
27
+ "classifier",
28
+ "score"
29
+ ],
30
+ "peft_type": "LORA",
31
+ "peft_version": "0.19.1",
32
+ "qalora_group_size": 16,
33
+ "r": 16,
34
+ "rank_pattern": {},
35
+ "revision": null,
36
+ "target_modules": [
37
+ "o_proj",
38
+ "k_proj",
39
+ "v_proj",
40
+ "q_proj"
41
+ ],
42
+ "target_parameters": null,
43
+ "task_type": "TOKEN_CLS",
44
+ "trainable_token_indices": null,
45
+ "use_bdlora": null,
46
+ "use_dora": false,
47
+ "use_qalora": false,
48
+ "use_rslora": false
49
+ }
checkpoint-3220/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d74f50cb3cf9daf901621a6d3c261536df2696ad33a7948d017f3c739d6fcc79
3
+ size 2415282
checkpoint-3220/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c42b314fd5effda59213150b36df174d95e11e36e3a5d057a9b1c27bc534d33
3
+ size 4870074
checkpoint-3220/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6228fc6b2df3522794015282d9d6ec31c5d64a50d73c37e13a7098d41fb3e4f1
3
+ size 14512
checkpoint-3220/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fcf4f46cfe440c3d4b8b4742b0f1543f19b637bf35f17acdcd47cd790f627e2
3
+ size 14512
checkpoint-3220/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8e680af0a1e89e60865481bc7e212dab8ba6721c1e3cc66fb1e617e0b05d8da
3
+ size 1064
checkpoint-3220/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e714c627d94fd333b14f9ff32436219a4d7ac969719efe340fdc3385e1c7cd3e
3
+ size 27868272
checkpoint-3220/tokenizer_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "eos_token": "<|endoftext|>",
4
+ "is_local": true,
5
+ "local_files_only": false,
6
+ "model_input_names": [
7
+ "input_ids",
8
+ "attention_mask"
9
+ ],
10
+ "model_max_length": 128000,
11
+ "pad_token": "<|endoftext|>",
12
+ "tokenizer_class": "TokenizersBackend"
13
+ }
checkpoint-3220/trainer_state.json ADDED
@@ -0,0 +1,1536 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 3220,
3
+ "best_metric": 0.8477206595538312,
4
+ "best_model_checkpoint": "/workspace/data/checkpoints/ko_pii_hf_ddp_v6_lora/checkpoint-3220",
5
+ "epoch": 10.0,
6
+ "eval_steps": 500,
7
+ "global_step": 3220,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.003105590062111801,
14
+ "grad_norm": 8.771296501159668,
15
+ "learning_rate": 0.0,
16
+ "loss": 1.4503676891326904,
17
+ "step": 1
18
+ },
19
+ {
20
+ "epoch": 0.07763975155279502,
21
+ "grad_norm": 8.387271881103516,
22
+ "learning_rate": 3.7267080745341614e-05,
23
+ "loss": 1.2439465522766113,
24
+ "step": 25
25
+ },
26
+ {
27
+ "epoch": 0.15527950310559005,
28
+ "grad_norm": 4.6111741065979,
29
+ "learning_rate": 7.608695652173914e-05,
30
+ "loss": 0.9780036926269531,
31
+ "step": 50
32
+ },
33
+ {
34
+ "epoch": 0.2329192546583851,
35
+ "grad_norm": 1.846106767654419,
36
+ "learning_rate": 0.00011490683229813665,
37
+ "loss": 0.5338097381591796,
38
+ "step": 75
39
+ },
40
+ {
41
+ "epoch": 0.3105590062111801,
42
+ "grad_norm": 2.225909471511841,
43
+ "learning_rate": 0.00015372670807453417,
44
+ "loss": 0.34235782623291017,
45
+ "step": 100
46
+ },
47
+ {
48
+ "epoch": 0.38819875776397517,
49
+ "grad_norm": 2.14986252784729,
50
+ "learning_rate": 0.0001925465838509317,
51
+ "loss": 0.2744290542602539,
52
+ "step": 125
53
+ },
54
+ {
55
+ "epoch": 0.4658385093167702,
56
+ "grad_norm": 2.1372387409210205,
57
+ "learning_rate": 0.0002313664596273292,
58
+ "loss": 0.2258339309692383,
59
+ "step": 150
60
+ },
61
+ {
62
+ "epoch": 0.5434782608695652,
63
+ "grad_norm": 1.5849782228469849,
64
+ "learning_rate": 0.0002701863354037267,
65
+ "loss": 0.2162160873413086,
66
+ "step": 175
67
+ },
68
+ {
69
+ "epoch": 0.6211180124223602,
70
+ "grad_norm": 1.6991852521896362,
71
+ "learning_rate": 0.00030900621118012424,
72
+ "loss": 0.20915340423583983,
73
+ "step": 200
74
+ },
75
+ {
76
+ "epoch": 0.6987577639751553,
77
+ "grad_norm": 2.723663568496704,
78
+ "learning_rate": 0.00034782608695652176,
79
+ "loss": 0.18431522369384765,
80
+ "step": 225
81
+ },
82
+ {
83
+ "epoch": 0.7763975155279503,
84
+ "grad_norm": 1.2743035554885864,
85
+ "learning_rate": 0.00038664596273291923,
86
+ "loss": 0.17637535095214843,
87
+ "step": 250
88
+ },
89
+ {
90
+ "epoch": 0.8540372670807453,
91
+ "grad_norm": 3.452497720718384,
92
+ "learning_rate": 0.0004254658385093168,
93
+ "loss": 0.16074188232421874,
94
+ "step": 275
95
+ },
96
+ {
97
+ "epoch": 0.9316770186335404,
98
+ "grad_norm": 2.547046661376953,
99
+ "learning_rate": 0.00046428571428571433,
100
+ "loss": 0.1531696319580078,
101
+ "step": 300
102
+ },
103
+ {
104
+ "epoch": 1.0,
105
+ "eval_class_account_number_f1": 0.71900826446281,
106
+ "eval_class_account_number_gold_spans": 113.0,
107
+ "eval_class_account_number_precision": 0.6744186046511628,
108
+ "eval_class_account_number_pred_spans": 129.0,
109
+ "eval_class_account_number_recall": 0.7699115044247787,
110
+ "eval_class_ip_address_f1": 1.0,
111
+ "eval_class_ip_address_gold_spans": 4.0,
112
+ "eval_class_ip_address_precision": 1.0,
113
+ "eval_class_ip_address_pred_spans": 4.0,
114
+ "eval_class_ip_address_recall": 1.0,
115
+ "eval_class_personal_handle_f1": 0.8070175438596492,
116
+ "eval_class_personal_handle_gold_spans": 28.0,
117
+ "eval_class_personal_handle_precision": 0.7931034482758621,
118
+ "eval_class_personal_handle_pred_spans": 29.0,
119
+ "eval_class_personal_handle_recall": 0.8214285714285714,
120
+ "eval_class_private_address_f1": 0.08450704225352113,
121
+ "eval_class_private_address_gold_spans": 48.0,
122
+ "eval_class_private_address_precision": 0.13043478260869565,
123
+ "eval_class_private_address_pred_spans": 23.0,
124
+ "eval_class_private_address_recall": 0.0625,
125
+ "eval_class_private_date_f1": 0.8985507246376813,
126
+ "eval_class_private_date_gold_spans": 33.0,
127
+ "eval_class_private_date_precision": 0.8611111111111112,
128
+ "eval_class_private_date_pred_spans": 36.0,
129
+ "eval_class_private_date_recall": 0.9393939393939394,
130
+ "eval_class_private_email_f1": 0.9620253164556962,
131
+ "eval_class_private_email_gold_spans": 39.0,
132
+ "eval_class_private_email_precision": 0.95,
133
+ "eval_class_private_email_pred_spans": 40.0,
134
+ "eval_class_private_email_recall": 0.9743589743589743,
135
+ "eval_class_private_person_f1": 0.32388663967611336,
136
+ "eval_class_private_person_gold_spans": 163.0,
137
+ "eval_class_private_person_precision": 0.47619047619047616,
138
+ "eval_class_private_person_pred_spans": 84.0,
139
+ "eval_class_private_person_recall": 0.24539877300613497,
140
+ "eval_class_private_phone_f1": 0.9855072463768116,
141
+ "eval_class_private_phone_gold_spans": 69.0,
142
+ "eval_class_private_phone_precision": 0.9855072463768116,
143
+ "eval_class_private_phone_pred_spans": 69.0,
144
+ "eval_class_private_phone_recall": 0.9855072463768116,
145
+ "eval_class_private_url_f1": 1.0,
146
+ "eval_class_private_url_gold_spans": 23.0,
147
+ "eval_class_private_url_precision": 1.0,
148
+ "eval_class_private_url_pred_spans": 23.0,
149
+ "eval_class_private_url_recall": 1.0,
150
+ "eval_gold_spans": 520.0,
151
+ "eval_loss": 0.1431565284729004,
152
+ "eval_pred_spans": 437.0,
153
+ "eval_runtime": 5.911,
154
+ "eval_samples_per_second": 376.753,
155
+ "eval_span_f1": 0.6624869383490073,
156
+ "eval_span_precision": 0.7254004576659039,
157
+ "eval_span_recall": 0.6096153846153847,
158
+ "eval_steps_per_second": 3.045,
159
+ "eval_token_accuracy": 0.9817635387488328,
160
+ "step": 322
161
+ },
162
+ {
163
+ "epoch": 1.0093167701863355,
164
+ "grad_norm": 1.0337207317352295,
165
+ "learning_rate": 0.000499999412412136,
166
+ "loss": 0.1343870735168457,
167
+ "step": 325
168
+ },
169
+ {
170
+ "epoch": 1.0869565217391304,
171
+ "grad_norm": 1.150405764579773,
172
+ "learning_rate": 0.0004998929197148332,
173
+ "loss": 0.1398656463623047,
174
+ "step": 350
175
+ },
176
+ {
177
+ "epoch": 1.1645962732919255,
178
+ "grad_norm": 1.239856243133545,
179
+ "learning_rate": 0.0004996028956208098,
180
+ "loss": 0.11556500434875489,
181
+ "step": 375
182
+ },
183
+ {
184
+ "epoch": 1.2422360248447206,
185
+ "grad_norm": 4.1494951248168945,
186
+ "learning_rate": 0.0004991295531354085,
187
+ "loss": 0.10652177810668945,
188
+ "step": 400
189
+ },
190
+ {
191
+ "epoch": 1.3198757763975155,
192
+ "grad_norm": 1.820974349975586,
193
+ "learning_rate": 0.0004984732399003618,
194
+ "loss": 0.1127833080291748,
195
+ "step": 425
196
+ },
197
+ {
198
+ "epoch": 1.3975155279503104,
199
+ "grad_norm": 0.9540935158729553,
200
+ "learning_rate": 0.0004976344379384687,
201
+ "loss": 0.1028218650817871,
202
+ "step": 450
203
+ },
204
+ {
205
+ "epoch": 1.4751552795031055,
206
+ "grad_norm": 1.7447997331619263,
207
+ "learning_rate": 0.0004966137632995793,
208
+ "loss": 0.11593814849853516,
209
+ "step": 475
210
+ },
211
+ {
212
+ "epoch": 1.5527950310559007,
213
+ "grad_norm": 1.3501027822494507,
214
+ "learning_rate": 0.0004954119656081417,
215
+ "loss": 0.1070643424987793,
216
+ "step": 500
217
+ },
218
+ {
219
+ "epoch": 1.6304347826086958,
220
+ "grad_norm": 0.6298630237579346,
221
+ "learning_rate": 0.0004940299275126486,
222
+ "loss": 0.11064315795898437,
223
+ "step": 525
224
+ },
225
+ {
226
+ "epoch": 1.7080745341614907,
227
+ "grad_norm": 0.8390702605247498,
228
+ "learning_rate": 0.0004924686640373841,
229
+ "loss": 0.09777620315551758,
230
+ "step": 550
231
+ },
232
+ {
233
+ "epoch": 1.7857142857142856,
234
+ "grad_norm": 0.5763865113258362,
235
+ "learning_rate": 0.0004907293218369499,
236
+ "loss": 0.08596687316894532,
237
+ "step": 575
238
+ },
239
+ {
240
+ "epoch": 1.8633540372670807,
241
+ "grad_norm": 0.7198231816291809,
242
+ "learning_rate": 0.0004888131783541152,
243
+ "loss": 0.0955750274658203,
244
+ "step": 600
245
+ },
246
+ {
247
+ "epoch": 1.9409937888198758,
248
+ "grad_norm": 1.0241382122039795,
249
+ "learning_rate": 0.0004867216408816122,
250
+ "loss": 0.0880333137512207,
251
+ "step": 625
252
+ },
253
+ {
254
+ "epoch": 2.0,
255
+ "eval_class_account_number_f1": 0.9184549356223176,
256
+ "eval_class_account_number_gold_spans": 113.0,
257
+ "eval_class_account_number_precision": 0.8916666666666667,
258
+ "eval_class_account_number_pred_spans": 120.0,
259
+ "eval_class_account_number_recall": 0.9469026548672567,
260
+ "eval_class_ip_address_f1": 1.0,
261
+ "eval_class_ip_address_gold_spans": 4.0,
262
+ "eval_class_ip_address_precision": 1.0,
263
+ "eval_class_ip_address_pred_spans": 4.0,
264
+ "eval_class_ip_address_recall": 1.0,
265
+ "eval_class_personal_handle_f1": 0.9310344827586207,
266
+ "eval_class_personal_handle_gold_spans": 28.0,
267
+ "eval_class_personal_handle_precision": 0.9,
268
+ "eval_class_personal_handle_pred_spans": 30.0,
269
+ "eval_class_personal_handle_recall": 0.9642857142857143,
270
+ "eval_class_private_address_f1": 0.5747126436781609,
271
+ "eval_class_private_address_gold_spans": 48.0,
272
+ "eval_class_private_address_precision": 0.6410256410256411,
273
+ "eval_class_private_address_pred_spans": 39.0,
274
+ "eval_class_private_address_recall": 0.5208333333333334,
275
+ "eval_class_private_date_f1": 0.927536231884058,
276
+ "eval_class_private_date_gold_spans": 33.0,
277
+ "eval_class_private_date_precision": 0.8888888888888888,
278
+ "eval_class_private_date_pred_spans": 36.0,
279
+ "eval_class_private_date_recall": 0.9696969696969697,
280
+ "eval_class_private_email_f1": 0.9743589743589743,
281
+ "eval_class_private_email_gold_spans": 39.0,
282
+ "eval_class_private_email_precision": 0.9743589743589743,
283
+ "eval_class_private_email_pred_spans": 39.0,
284
+ "eval_class_private_email_recall": 0.9743589743589743,
285
+ "eval_class_private_person_f1": 0.5182926829268292,
286
+ "eval_class_private_person_gold_spans": 163.0,
287
+ "eval_class_private_person_precision": 0.5151515151515151,
288
+ "eval_class_private_person_pred_spans": 165.0,
289
+ "eval_class_private_person_recall": 0.5214723926380368,
290
+ "eval_class_private_phone_f1": 1.0,
291
+ "eval_class_private_phone_gold_spans": 69.0,
292
+ "eval_class_private_phone_precision": 1.0,
293
+ "eval_class_private_phone_pred_spans": 69.0,
294
+ "eval_class_private_phone_recall": 1.0,
295
+ "eval_class_private_url_f1": 0.9130434782608695,
296
+ "eval_class_private_url_gold_spans": 23.0,
297
+ "eval_class_private_url_precision": 0.9130434782608695,
298
+ "eval_class_private_url_pred_spans": 23.0,
299
+ "eval_class_private_url_recall": 0.9130434782608695,
300
+ "eval_gold_spans": 520.0,
301
+ "eval_loss": 0.08562110364437103,
302
+ "eval_pred_spans": 525.0,
303
+ "eval_runtime": 6.7931,
304
+ "eval_samples_per_second": 327.832,
305
+ "eval_span_f1": 0.7808612440191388,
306
+ "eval_span_precision": 0.7771428571428571,
307
+ "eval_span_recall": 0.7846153846153846,
308
+ "eval_steps_per_second": 2.65,
309
+ "eval_token_accuracy": 0.9886788048552755,
310
+ "step": 644
311
+ },
312
+ {
313
+ "epoch": 2.018633540372671,
314
+ "grad_norm": 1.2737419605255127,
315
+ "learning_rate": 0.00048445624552856343,
316
+ "loss": 0.0896837329864502,
317
+ "step": 650
318
+ },
319
+ {
320
+ "epoch": 2.0962732919254656,
321
+ "grad_norm": 0.5823193788528442,
322
+ "learning_rate": 0.00048201865609230155,
323
+ "loss": 0.07990738868713379,
324
+ "step": 675
325
+ },
326
+ {
327
+ "epoch": 2.1739130434782608,
328
+ "grad_norm": 1.4650110006332397,
329
+ "learning_rate": 0.0004794106628364098,
330
+ "loss": 0.06952488899230957,
331
+ "step": 700
332
+ },
333
+ {
334
+ "epoch": 2.251552795031056,
335
+ "grad_norm": 0.6727014183998108,
336
+ "learning_rate": 0.0004766341811758803,
337
+ "loss": 0.07833251953125,
338
+ "step": 725
339
+ },
340
+ {
341
+ "epoch": 2.329192546583851,
342
+ "grad_norm": 0.6006216406822205,
343
+ "learning_rate": 0.000473691250270357,
344
+ "loss": 0.06939034461975098,
345
+ "step": 750
346
+ },
347
+ {
348
+ "epoch": 2.406832298136646,
349
+ "grad_norm": 0.5175172090530396,
350
+ "learning_rate": 0.00047058403152649405,
351
+ "loss": 0.05780394077301025,
352
+ "step": 775
353
+ },
354
+ {
355
+ "epoch": 2.4844720496894412,
356
+ "grad_norm": 1.1351436376571655,
357
+ "learning_rate": 0.00046731480701053265,
358
+ "loss": 0.06771457672119141,
359
+ "step": 800
360
+ },
361
+ {
362
+ "epoch": 2.562111801242236,
363
+ "grad_norm": 0.8665878772735596,
364
+ "learning_rate": 0.0004638859777722596,
365
+ "loss": 0.0703080177307129,
366
+ "step": 825
367
+ },
368
+ {
369
+ "epoch": 2.639751552795031,
370
+ "grad_norm": 0.8073828220367432,
371
+ "learning_rate": 0.00046030006208158,
372
+ "loss": 0.0754665756225586,
373
+ "step": 850
374
+ },
375
+ {
376
+ "epoch": 2.717391304347826,
377
+ "grad_norm": 0.6520082950592041,
378
+ "learning_rate": 0.00045655969357899875,
379
+ "loss": 0.06543315410614013,
380
+ "step": 875
381
+ },
382
+ {
383
+ "epoch": 2.795031055900621,
384
+ "grad_norm": 1.448807954788208,
385
+ "learning_rate": 0.00045266761934136896,
386
+ "loss": 0.07107664108276367,
387
+ "step": 900
388
+ },
389
+ {
390
+ "epoch": 2.8726708074534164,
391
+ "grad_norm": 0.45655015110969543,
392
+ "learning_rate": 0.00044862669786432865,
393
+ "loss": 0.07224170684814453,
394
+ "step": 925
395
+ },
396
+ {
397
+ "epoch": 2.950310559006211,
398
+ "grad_norm": 0.4825460612773895,
399
+ "learning_rate": 0.0004444398969629067,
400
+ "loss": 0.06763762950897217,
401
+ "step": 950
402
+ },
403
+ {
404
+ "epoch": 3.0,
405
+ "eval_class_account_number_f1": 0.9350649350649352,
406
+ "eval_class_account_number_gold_spans": 113.0,
407
+ "eval_class_account_number_precision": 0.9152542372881356,
408
+ "eval_class_account_number_pred_spans": 118.0,
409
+ "eval_class_account_number_recall": 0.9557522123893806,
410
+ "eval_class_ip_address_f1": 1.0,
411
+ "eval_class_ip_address_gold_spans": 4.0,
412
+ "eval_class_ip_address_precision": 1.0,
413
+ "eval_class_ip_address_pred_spans": 4.0,
414
+ "eval_class_ip_address_recall": 1.0,
415
+ "eval_class_personal_handle_f1": 0.912280701754386,
416
+ "eval_class_personal_handle_gold_spans": 28.0,
417
+ "eval_class_personal_handle_precision": 0.896551724137931,
418
+ "eval_class_personal_handle_pred_spans": 29.0,
419
+ "eval_class_personal_handle_recall": 0.9285714285714286,
420
+ "eval_class_private_address_f1": 0.6041666666666666,
421
+ "eval_class_private_address_gold_spans": 48.0,
422
+ "eval_class_private_address_precision": 0.6041666666666666,
423
+ "eval_class_private_address_pred_spans": 48.0,
424
+ "eval_class_private_address_recall": 0.6041666666666666,
425
+ "eval_class_private_date_f1": 0.9253731343283583,
426
+ "eval_class_private_date_gold_spans": 33.0,
427
+ "eval_class_private_date_precision": 0.9117647058823529,
428
+ "eval_class_private_date_pred_spans": 34.0,
429
+ "eval_class_private_date_recall": 0.9393939393939394,
430
+ "eval_class_private_email_f1": 0.9487179487179487,
431
+ "eval_class_private_email_gold_spans": 39.0,
432
+ "eval_class_private_email_precision": 0.9487179487179487,
433
+ "eval_class_private_email_pred_spans": 39.0,
434
+ "eval_class_private_email_recall": 0.9487179487179487,
435
+ "eval_class_private_person_f1": 0.6013071895424836,
436
+ "eval_class_private_person_gold_spans": 163.0,
437
+ "eval_class_private_person_precision": 0.6433566433566433,
438
+ "eval_class_private_person_pred_spans": 143.0,
439
+ "eval_class_private_person_recall": 0.5644171779141104,
440
+ "eval_class_private_phone_f1": 1.0,
441
+ "eval_class_private_phone_gold_spans": 69.0,
442
+ "eval_class_private_phone_precision": 1.0,
443
+ "eval_class_private_phone_pred_spans": 69.0,
444
+ "eval_class_private_phone_recall": 1.0,
445
+ "eval_class_private_url_f1": 0.9361702127659574,
446
+ "eval_class_private_url_gold_spans": 23.0,
447
+ "eval_class_private_url_precision": 0.9166666666666666,
448
+ "eval_class_private_url_pred_spans": 24.0,
449
+ "eval_class_private_url_recall": 0.9565217391304348,
450
+ "eval_gold_spans": 520.0,
451
+ "eval_loss": 0.07642154395580292,
452
+ "eval_pred_spans": 508.0,
453
+ "eval_runtime": 6.8956,
454
+ "eval_samples_per_second": 322.962,
455
+ "eval_span_f1": 0.8132295719844358,
456
+ "eval_span_precision": 0.8228346456692913,
457
+ "eval_span_recall": 0.8038461538461539,
458
+ "eval_steps_per_second": 2.61,
459
+ "eval_token_accuracy": 0.9904586834733894,
460
+ "step": 966
461
+ },
462
+ {
463
+ "epoch": 3.027950310559006,
464
+ "grad_norm": 0.3684696853160858,
465
+ "learning_rate": 0.00044011029159184015,
466
+ "loss": 0.04707308769226074,
467
+ "step": 975
468
+ },
469
+ {
470
+ "epoch": 3.1055900621118013,
471
+ "grad_norm": 0.47226452827453613,
472
+ "learning_rate": 0.00043564106158720383,
473
+ "loss": 0.04691335678100586,
474
+ "step": 1000
475
+ },
476
+ {
477
+ "epoch": 3.1832298136645965,
478
+ "grad_norm": 0.6678075194358826,
479
+ "learning_rate": 0.00043103548933101155,
480
+ "loss": 0.044693613052368165,
481
+ "step": 1025
482
+ },
483
+ {
484
+ "epoch": 3.260869565217391,
485
+ "grad_norm": 0.675778865814209,
486
+ "learning_rate": 0.00042629695734050255,
487
+ "loss": 0.04987412452697754,
488
+ "step": 1050
489
+ },
490
+ {
491
+ "epoch": 3.3385093167701863,
492
+ "grad_norm": 1.2661679983139038,
493
+ "learning_rate": 0.00042142894578388486,
494
+ "loss": 0.046450929641723634,
495
+ "step": 1075
496
+ },
497
+ {
498
+ "epoch": 3.4161490683229814,
499
+ "grad_norm": 0.7080872058868408,
500
+ "learning_rate": 0.00041643502992436013,
501
+ "loss": 0.050359611511230466,
502
+ "step": 1100
503
+ },
504
+ {
505
+ "epoch": 3.4937888198757765,
506
+ "grad_norm": 0.7029626369476318,
507
+ "learning_rate": 0.0004113188774943071,
508
+ "loss": 0.06056289672851563,
509
+ "step": 1125
510
+ },
511
+ {
512
+ "epoch": 3.571428571428571,
513
+ "grad_norm": 0.558189332485199,
514
+ "learning_rate": 0.000406084246001551,
515
+ "loss": 0.045447168350219724,
516
+ "step": 1150
517
+ },
518
+ {
519
+ "epoch": 3.6490683229813663,
520
+ "grad_norm": 0.5974974036216736,
521
+ "learning_rate": 0.00040073497996969933,
522
+ "loss": 0.045512027740478515,
523
+ "step": 1175
524
+ },
525
+ {
526
+ "epoch": 3.7267080745341614,
527
+ "grad_norm": 0.5707910656929016,
528
+ "learning_rate": 0.000395275008114569,
529
+ "loss": 0.05025949954986572,
530
+ "step": 1200
531
+ },
532
+ {
533
+ "epoch": 3.8043478260869565,
534
+ "grad_norm": 1.0855952501296997,
535
+ "learning_rate": 0.00038970834045878026,
536
+ "loss": 0.06197848796844482,
537
+ "step": 1225
538
+ },
539
+ {
540
+ "epoch": 3.8819875776397517,
541
+ "grad_norm": 0.3965921401977539,
542
+ "learning_rate": 0.0003840390653866341,
543
+ "loss": 0.05249831199645996,
544
+ "step": 1250
545
+ },
546
+ {
547
+ "epoch": 3.9596273291925463,
548
+ "grad_norm": 0.5562620162963867,
549
+ "learning_rate": 0.00037827134664143947,
550
+ "loss": 0.045794315338134765,
551
+ "step": 1275
552
+ },
553
+ {
554
+ "epoch": 4.0,
555
+ "eval_class_account_number_f1": 0.9691629955947135,
556
+ "eval_class_account_number_gold_spans": 113.0,
557
+ "eval_class_account_number_precision": 0.9649122807017544,
558
+ "eval_class_account_number_pred_spans": 114.0,
559
+ "eval_class_account_number_recall": 0.9734513274336283,
560
+ "eval_class_ip_address_f1": 1.0,
561
+ "eval_class_ip_address_gold_spans": 4.0,
562
+ "eval_class_ip_address_precision": 1.0,
563
+ "eval_class_ip_address_pred_spans": 4.0,
564
+ "eval_class_ip_address_recall": 1.0,
565
+ "eval_class_personal_handle_f1": 0.7719298245614034,
566
+ "eval_class_personal_handle_gold_spans": 28.0,
567
+ "eval_class_personal_handle_precision": 0.7586206896551724,
568
+ "eval_class_personal_handle_pred_spans": 29.0,
569
+ "eval_class_personal_handle_recall": 0.7857142857142857,
570
+ "eval_class_private_address_f1": 0.6067415730337078,
571
+ "eval_class_private_address_gold_spans": 48.0,
572
+ "eval_class_private_address_precision": 0.6585365853658537,
573
+ "eval_class_private_address_pred_spans": 41.0,
574
+ "eval_class_private_address_recall": 0.5625,
575
+ "eval_class_private_date_f1": 1.0,
576
+ "eval_class_private_date_gold_spans": 33.0,
577
+ "eval_class_private_date_precision": 1.0,
578
+ "eval_class_private_date_pred_spans": 33.0,
579
+ "eval_class_private_date_recall": 1.0,
580
+ "eval_class_private_email_f1": 0.9873417721518987,
581
+ "eval_class_private_email_gold_spans": 39.0,
582
+ "eval_class_private_email_precision": 0.975,
583
+ "eval_class_private_email_pred_spans": 40.0,
584
+ "eval_class_private_email_recall": 1.0,
585
+ "eval_class_private_person_f1": 0.6257668711656442,
586
+ "eval_class_private_person_gold_spans": 163.0,
587
+ "eval_class_private_person_precision": 0.6257668711656442,
588
+ "eval_class_private_person_pred_spans": 163.0,
589
+ "eval_class_private_person_recall": 0.6257668711656442,
590
+ "eval_class_private_phone_f1": 1.0,
591
+ "eval_class_private_phone_gold_spans": 69.0,
592
+ "eval_class_private_phone_precision": 1.0,
593
+ "eval_class_private_phone_pred_spans": 69.0,
594
+ "eval_class_private_phone_recall": 1.0,
595
+ "eval_class_private_url_f1": 0.9565217391304348,
596
+ "eval_class_private_url_gold_spans": 23.0,
597
+ "eval_class_private_url_precision": 0.9565217391304348,
598
+ "eval_class_private_url_pred_spans": 23.0,
599
+ "eval_class_private_url_recall": 0.9565217391304348,
600
+ "eval_gold_spans": 520.0,
601
+ "eval_loss": 0.07150626927614212,
602
+ "eval_pred_spans": 516.0,
603
+ "eval_runtime": 6.9918,
604
+ "eval_samples_per_second": 318.517,
605
+ "eval_span_f1": 0.8262548262548264,
606
+ "eval_span_precision": 0.8294573643410853,
607
+ "eval_span_recall": 0.823076923076923,
608
+ "eval_steps_per_second": 2.574,
609
+ "eval_token_accuracy": 0.9900501867413632,
610
+ "step": 1288
611
+ },
612
+ {
613
+ "epoch": 4.037267080745342,
614
+ "grad_norm": 0.3308059573173523,
615
+ "learning_rate": 0.00037240942026749233,
616
+ "loss": 0.03683102369308472,
617
+ "step": 1300
618
+ },
619
+ {
620
+ "epoch": 4.114906832298137,
621
+ "grad_norm": 0.9377778768539429,
622
+ "learning_rate": 0.0003664575914989552,
623
+ "loss": 0.033050749301910404,
624
+ "step": 1325
625
+ },
626
+ {
627
+ "epoch": 4.192546583850931,
628
+ "grad_norm": 0.5735832452774048,
629
+ "learning_rate": 0.00036042023159791974,
630
+ "loss": 0.031362690925598145,
631
+ "step": 1350
632
+ },
633
+ {
634
+ "epoch": 4.270186335403727,
635
+ "grad_norm": 0.631287693977356,
636
+ "learning_rate": 0.0003543017746439766,
637
+ "loss": 0.03477231979370117,
638
+ "step": 1375
639
+ },
640
+ {
641
+ "epoch": 4.3478260869565215,
642
+ "grad_norm": 1.0935797691345215,
643
+ "learning_rate": 0.0003481067142776489,
644
+ "loss": 0.03894762754440308,
645
+ "step": 1400
646
+ },
647
+ {
648
+ "epoch": 4.425465838509317,
649
+ "grad_norm": 0.6073715090751648,
650
+ "learning_rate": 0.0003418396004000829,
651
+ "loss": 0.03512312889099121,
652
+ "step": 1425
653
+ },
654
+ {
655
+ "epoch": 4.503105590062112,
656
+ "grad_norm": 0.6249358654022217,
657
+ "learning_rate": 0.0003355050358314172,
658
+ "loss": 0.04299846649169922,
659
+ "step": 1450
660
+ },
661
+ {
662
+ "epoch": 4.580745341614907,
663
+ "grad_norm": 0.7248182892799377,
664
+ "learning_rate": 0.0003291076729302869,
665
+ "loss": 0.03905656576156616,
666
+ "step": 1475
667
+ },
668
+ {
669
+ "epoch": 4.658385093167702,
670
+ "grad_norm": 0.9138495922088623,
671
+ "learning_rate": 0.00032265221017694474,
672
+ "loss": 0.039488759040832516,
673
+ "step": 1500
674
+ },
675
+ {
676
+ "epoch": 4.736024844720497,
677
+ "grad_norm": 0.6601109504699707,
678
+ "learning_rate": 0.00031614338872250835,
679
+ "loss": 0.03172698974609375,
680
+ "step": 1525
681
+ },
682
+ {
683
+ "epoch": 4.813664596273292,
684
+ "grad_norm": 1.141296625137329,
685
+ "learning_rate": 0.0003095859889068687,
686
+ "loss": 0.038228404521942136,
687
+ "step": 1550
688
+ },
689
+ {
690
+ "epoch": 4.891304347826087,
691
+ "grad_norm": 1.3895987272262573,
692
+ "learning_rate": 0.0003029848267478158,
693
+ "loss": 0.03626953125,
694
+ "step": 1575
695
+ },
696
+ {
697
+ "epoch": 4.9689440993788825,
698
+ "grad_norm": 0.5478597283363342,
699
+ "learning_rate": 0.00029634475040396285,
700
+ "loss": 0.03073249340057373,
701
+ "step": 1600
702
+ },
703
+ {
704
+ "epoch": 5.0,
705
+ "eval_class_account_number_f1": 0.9691629955947135,
706
+ "eval_class_account_number_gold_spans": 113.0,
707
+ "eval_class_account_number_precision": 0.9649122807017544,
708
+ "eval_class_account_number_pred_spans": 114.0,
709
+ "eval_class_account_number_recall": 0.9734513274336283,
710
+ "eval_class_ip_address_f1": 1.0,
711
+ "eval_class_ip_address_gold_spans": 4.0,
712
+ "eval_class_ip_address_precision": 1.0,
713
+ "eval_class_ip_address_pred_spans": 4.0,
714
+ "eval_class_ip_address_recall": 1.0,
715
+ "eval_class_personal_handle_f1": 0.8771929824561403,
716
+ "eval_class_personal_handle_gold_spans": 28.0,
717
+ "eval_class_personal_handle_precision": 0.8620689655172413,
718
+ "eval_class_personal_handle_pred_spans": 29.0,
719
+ "eval_class_personal_handle_recall": 0.8928571428571429,
720
+ "eval_class_private_address_f1": 0.6046511627906976,
721
+ "eval_class_private_address_gold_spans": 48.0,
722
+ "eval_class_private_address_precision": 0.6842105263157895,
723
+ "eval_class_private_address_pred_spans": 38.0,
724
+ "eval_class_private_address_recall": 0.5416666666666666,
725
+ "eval_class_private_date_f1": 1.0,
726
+ "eval_class_private_date_gold_spans": 33.0,
727
+ "eval_class_private_date_precision": 1.0,
728
+ "eval_class_private_date_pred_spans": 33.0,
729
+ "eval_class_private_date_recall": 1.0,
730
+ "eval_class_private_email_f1": 0.9873417721518987,
731
+ "eval_class_private_email_gold_spans": 39.0,
732
+ "eval_class_private_email_precision": 0.975,
733
+ "eval_class_private_email_pred_spans": 40.0,
734
+ "eval_class_private_email_recall": 1.0,
735
+ "eval_class_private_person_f1": 0.6027397260273972,
736
+ "eval_class_private_person_gold_spans": 163.0,
737
+ "eval_class_private_person_precision": 0.5445544554455446,
738
+ "eval_class_private_person_pred_spans": 202.0,
739
+ "eval_class_private_person_recall": 0.6748466257668712,
740
+ "eval_class_private_phone_f1": 1.0,
741
+ "eval_class_private_phone_gold_spans": 69.0,
742
+ "eval_class_private_phone_precision": 1.0,
743
+ "eval_class_private_phone_pred_spans": 69.0,
744
+ "eval_class_private_phone_recall": 1.0,
745
+ "eval_class_private_url_f1": 0.9565217391304348,
746
+ "eval_class_private_url_gold_spans": 23.0,
747
+ "eval_class_private_url_precision": 0.9565217391304348,
748
+ "eval_class_private_url_pred_spans": 23.0,
749
+ "eval_class_private_url_recall": 0.9565217391304348,
750
+ "eval_gold_spans": 520.0,
751
+ "eval_loss": 0.07776352018117905,
752
+ "eval_pred_spans": 552.0,
753
+ "eval_runtime": 7.1537,
754
+ "eval_samples_per_second": 311.307,
755
+ "eval_span_f1": 0.8171641791044777,
756
+ "eval_span_precision": 0.7934782608695652,
757
+ "eval_span_recall": 0.8423076923076923,
758
+ "eval_steps_per_second": 2.516,
759
+ "eval_token_accuracy": 0.9894082633053222,
760
+ "step": 1610
761
+ },
762
+ {
763
+ "epoch": 5.046583850931677,
764
+ "grad_norm": 0.8306529521942139,
765
+ "learning_rate": 0.00028967063661406285,
766
+ "loss": 0.02931797981262207,
767
+ "step": 1625
768
+ },
769
+ {
770
+ "epoch": 5.124223602484472,
771
+ "grad_norm": 0.3057140111923218,
772
+ "learning_rate": 0.00028296738711533655,
773
+ "loss": 0.02403106212615967,
774
+ "step": 1650
775
+ },
776
+ {
777
+ "epoch": 5.201863354037267,
778
+ "grad_norm": 0.6968570351600647,
779
+ "learning_rate": 0.0002762399250434392,
780
+ "loss": 0.024544227123260497,
781
+ "step": 1675
782
+ },
783
+ {
784
+ "epoch": 5.279503105590062,
785
+ "grad_norm": 0.8571075797080994,
786
+ "learning_rate": 0.0002694931913167124,
787
+ "loss": 0.021933317184448242,
788
+ "step": 1700
789
+ },
790
+ {
791
+ "epoch": 5.357142857142857,
792
+ "grad_norm": 0.5366156101226807,
793
+ "learning_rate": 0.00026273214100737516,
794
+ "loss": 0.031152501106262206,
795
+ "step": 1725
796
+ },
797
+ {
798
+ "epoch": 5.434782608695652,
799
+ "grad_norm": 0.5963289141654968,
800
+ "learning_rate": 0.0002559617397023193,
801
+ "loss": 0.02120798110961914,
802
+ "step": 1750
803
+ },
804
+ {
805
+ "epoch": 5.512422360248447,
806
+ "grad_norm": 0.36561551690101624,
807
+ "learning_rate": 0.00024918695985618394,
808
+ "loss": 0.023069217205047607,
809
+ "step": 1775
810
+ },
811
+ {
812
+ "epoch": 5.590062111801243,
813
+ "grad_norm": 0.834469199180603,
814
+ "learning_rate": 0.00024241277713938443,
815
+ "loss": 0.018797587156295776,
816
+ "step": 1800
817
+ },
818
+ {
819
+ "epoch": 5.667701863354037,
820
+ "grad_norm": 0.3487168848514557,
821
+ "learning_rate": 0.0002356441667837804,
822
+ "loss": 0.027416987419128416,
823
+ "step": 1825
824
+ },
825
+ {
826
+ "epoch": 5.745341614906832,
827
+ "grad_norm": 0.49782446026802063,
828
+ "learning_rate": 0.0002288860999286656,
829
+ "loss": 0.029009580612182617,
830
+ "step": 1850
831
+ },
832
+ {
833
+ "epoch": 5.8229813664596275,
834
+ "grad_norm": 0.5541417598724365,
835
+ "learning_rate": 0.00022214353996976349,
836
+ "loss": 0.02573819875717163,
837
+ "step": 1875
838
+ },
839
+ {
840
+ "epoch": 5.900621118012422,
841
+ "grad_norm": 0.9394005537033081,
842
+ "learning_rate": 0.00021542143891391025,
843
+ "loss": 0.024653725624084473,
844
+ "step": 1900
845
+ },
846
+ {
847
+ "epoch": 5.978260869565218,
848
+ "grad_norm": 0.9744410514831543,
849
+ "learning_rate": 0.0002087247337421019,
850
+ "loss": 0.028091182708740236,
851
+ "step": 1925
852
+ },
853
+ {
854
+ "epoch": 6.0,
855
+ "eval_class_account_number_f1": 0.9603524229074888,
856
+ "eval_class_account_number_gold_spans": 113.0,
857
+ "eval_class_account_number_precision": 0.956140350877193,
858
+ "eval_class_account_number_pred_spans": 114.0,
859
+ "eval_class_account_number_recall": 0.9646017699115044,
860
+ "eval_class_ip_address_f1": 1.0,
861
+ "eval_class_ip_address_gold_spans": 4.0,
862
+ "eval_class_ip_address_precision": 1.0,
863
+ "eval_class_ip_address_pred_spans": 4.0,
864
+ "eval_class_ip_address_recall": 1.0,
865
+ "eval_class_personal_handle_f1": 0.9090909090909091,
866
+ "eval_class_personal_handle_gold_spans": 28.0,
867
+ "eval_class_personal_handle_precision": 0.9259259259259259,
868
+ "eval_class_personal_handle_pred_spans": 27.0,
869
+ "eval_class_personal_handle_recall": 0.8928571428571429,
870
+ "eval_class_private_address_f1": 0.6813186813186812,
871
+ "eval_class_private_address_gold_spans": 48.0,
872
+ "eval_class_private_address_precision": 0.7209302325581395,
873
+ "eval_class_private_address_pred_spans": 43.0,
874
+ "eval_class_private_address_recall": 0.6458333333333334,
875
+ "eval_class_private_date_f1": 0.9393939393939394,
876
+ "eval_class_private_date_gold_spans": 33.0,
877
+ "eval_class_private_date_precision": 0.9393939393939394,
878
+ "eval_class_private_date_pred_spans": 33.0,
879
+ "eval_class_private_date_recall": 0.9393939393939394,
880
+ "eval_class_private_email_f1": 0.9620253164556962,
881
+ "eval_class_private_email_gold_spans": 39.0,
882
+ "eval_class_private_email_precision": 0.95,
883
+ "eval_class_private_email_pred_spans": 40.0,
884
+ "eval_class_private_email_recall": 0.9743589743589743,
885
+ "eval_class_private_person_f1": 0.6514657980456026,
886
+ "eval_class_private_person_gold_spans": 163.0,
887
+ "eval_class_private_person_precision": 0.6944444444444444,
888
+ "eval_class_private_person_pred_spans": 144.0,
889
+ "eval_class_private_person_recall": 0.6134969325153374,
890
+ "eval_class_private_phone_f1": 1.0,
891
+ "eval_class_private_phone_gold_spans": 69.0,
892
+ "eval_class_private_phone_precision": 1.0,
893
+ "eval_class_private_phone_pred_spans": 69.0,
894
+ "eval_class_private_phone_recall": 1.0,
895
+ "eval_class_private_url_f1": 0.9565217391304348,
896
+ "eval_class_private_url_gold_spans": 23.0,
897
+ "eval_class_private_url_precision": 0.9565217391304348,
898
+ "eval_class_private_url_pred_spans": 23.0,
899
+ "eval_class_private_url_recall": 0.9565217391304348,
900
+ "eval_gold_spans": 520.0,
901
+ "eval_loss": 0.07920802384614944,
902
+ "eval_pred_spans": 497.0,
903
+ "eval_runtime": 6.6716,
904
+ "eval_samples_per_second": 333.804,
905
+ "eval_span_f1": 0.8436578171091444,
906
+ "eval_span_precision": 0.8631790744466801,
907
+ "eval_span_recall": 0.825,
908
+ "eval_steps_per_second": 2.698,
909
+ "eval_token_accuracy": 0.9906921101774043,
910
+ "step": 1932
911
+ },
912
+ {
913
+ "epoch": 6.055900621118012,
914
+ "grad_norm": 0.5681946277618408,
915
+ "learning_rate": 0.00020205834278357742,
916
+ "loss": 0.01689502239227295,
917
+ "step": 1950
918
+ },
919
+ {
920
+ "epoch": 6.133540372670807,
921
+ "grad_norm": 0.3505108058452606,
922
+ "learning_rate": 0.00019542716210359964,
923
+ "loss": 0.015278645753860474,
924
+ "step": 1975
925
+ },
926
+ {
927
+ "epoch": 6.211180124223603,
928
+ "grad_norm": 0.3547140955924988,
929
+ "learning_rate": 0.00018883606190758903,
930
+ "loss": 0.015209052562713623,
931
+ "step": 2000
932
+ },
933
+ {
934
+ "epoch": 6.288819875776397,
935
+ "grad_norm": 0.3973138928413391,
936
+ "learning_rate": 0.00018228988296424876,
937
+ "loss": 0.01439091444015503,
938
+ "step": 2025
939
+ },
940
+ {
941
+ "epoch": 6.366459627329193,
942
+ "grad_norm": 0.5407560467720032,
943
+ "learning_rate": 0.00017579343305030985,
944
+ "loss": 0.014540103673934936,
945
+ "step": 2050
946
+ },
947
+ {
948
+ "epoch": 6.444099378881988,
949
+ "grad_norm": 0.3447468876838684,
950
+ "learning_rate": 0.00016935148341950694,
951
+ "loss": 0.017405078411102295,
952
+ "step": 2075
953
+ },
954
+ {
955
+ "epoch": 6.521739130434782,
956
+ "grad_norm": 0.30423322319984436,
957
+ "learning_rate": 0.0001629687652983776,
958
+ "loss": 0.015654172897338867,
959
+ "step": 2100
960
+ },
961
+ {
962
+ "epoch": 6.599378881987578,
963
+ "grad_norm": 0.47501835227012634,
964
+ "learning_rate": 0.0001566499664114601,
965
+ "loss": 0.017501153945922852,
966
+ "step": 2125
967
+ },
968
+ {
969
+ "epoch": 6.6770186335403725,
970
+ "grad_norm": 0.5283868312835693,
971
+ "learning_rate": 0.00015039972753843963,
972
+ "loss": 0.014991183280944825,
973
+ "step": 2150
974
+ },
975
+ {
976
+ "epoch": 6.754658385093168,
977
+ "grad_norm": 0.9889101982116699,
978
+ "learning_rate": 0.00014422263910577376,
979
+ "loss": 0.021359992027282716,
980
+ "step": 2175
981
+ },
982
+ {
983
+ "epoch": 6.832298136645963,
984
+ "grad_norm": 0.6929994225502014,
985
+ "learning_rate": 0.0001381232378152987,
986
+ "loss": 0.01291258454322815,
987
+ "step": 2200
988
+ },
989
+ {
990
+ "epoch": 6.909937888198757,
991
+ "grad_norm": 0.7863985896110535,
992
+ "learning_rate": 0.00013210600331229317,
993
+ "loss": 0.018135499954223634,
994
+ "step": 2225
995
+ },
996
+ {
997
+ "epoch": 6.987577639751553,
998
+ "grad_norm": 0.45059671998023987,
999
+ "learning_rate": 0.000126175354895447,
1000
+ "loss": 0.022398817539215087,
1001
+ "step": 2250
1002
+ },
1003
+ {
1004
+ "epoch": 7.0,
1005
+ "eval_class_account_number_f1": 0.9779735682819383,
1006
+ "eval_class_account_number_gold_spans": 113.0,
1007
+ "eval_class_account_number_precision": 0.9736842105263158,
1008
+ "eval_class_account_number_pred_spans": 114.0,
1009
+ "eval_class_account_number_recall": 0.9823008849557522,
1010
+ "eval_class_ip_address_f1": 1.0,
1011
+ "eval_class_ip_address_gold_spans": 4.0,
1012
+ "eval_class_ip_address_precision": 1.0,
1013
+ "eval_class_ip_address_pred_spans": 4.0,
1014
+ "eval_class_ip_address_recall": 1.0,
1015
+ "eval_class_personal_handle_f1": 0.9090909090909091,
1016
+ "eval_class_personal_handle_gold_spans": 28.0,
1017
+ "eval_class_personal_handle_precision": 0.9259259259259259,
1018
+ "eval_class_personal_handle_pred_spans": 27.0,
1019
+ "eval_class_personal_handle_recall": 0.8928571428571429,
1020
+ "eval_class_private_address_f1": 0.7333333333333334,
1021
+ "eval_class_private_address_gold_spans": 48.0,
1022
+ "eval_class_private_address_precision": 0.7857142857142857,
1023
+ "eval_class_private_address_pred_spans": 42.0,
1024
+ "eval_class_private_address_recall": 0.6875,
1025
+ "eval_class_private_date_f1": 1.0,
1026
+ "eval_class_private_date_gold_spans": 33.0,
1027
+ "eval_class_private_date_precision": 1.0,
1028
+ "eval_class_private_date_pred_spans": 33.0,
1029
+ "eval_class_private_date_recall": 1.0,
1030
+ "eval_class_private_email_f1": 0.9500000000000001,
1031
+ "eval_class_private_email_gold_spans": 39.0,
1032
+ "eval_class_private_email_precision": 0.926829268292683,
1033
+ "eval_class_private_email_pred_spans": 41.0,
1034
+ "eval_class_private_email_recall": 0.9743589743589743,
1035
+ "eval_class_private_person_f1": 0.6198083067092651,
1036
+ "eval_class_private_person_gold_spans": 163.0,
1037
+ "eval_class_private_person_precision": 0.6466666666666666,
1038
+ "eval_class_private_person_pred_spans": 150.0,
1039
+ "eval_class_private_person_recall": 0.5950920245398773,
1040
+ "eval_class_private_phone_f1": 1.0,
1041
+ "eval_class_private_phone_gold_spans": 69.0,
1042
+ "eval_class_private_phone_precision": 1.0,
1043
+ "eval_class_private_phone_pred_spans": 69.0,
1044
+ "eval_class_private_phone_recall": 1.0,
1045
+ "eval_class_private_url_f1": 0.9583333333333334,
1046
+ "eval_class_private_url_gold_spans": 23.0,
1047
+ "eval_class_private_url_precision": 0.92,
1048
+ "eval_class_private_url_pred_spans": 25.0,
1049
+ "eval_class_private_url_recall": 1.0,
1050
+ "eval_gold_spans": 520.0,
1051
+ "eval_loss": 0.08392658829689026,
1052
+ "eval_pred_spans": 505.0,
1053
+ "eval_runtime": 6.6789,
1054
+ "eval_samples_per_second": 333.44,
1055
+ "eval_span_f1": 0.8448780487804879,
1056
+ "eval_span_precision": 0.8574257425742574,
1057
+ "eval_span_recall": 0.8326923076923077,
1058
+ "eval_steps_per_second": 2.695,
1059
+ "eval_token_accuracy": 0.9905753968253969,
1060
+ "step": 2254
1061
+ },
1062
+ {
1063
+ "epoch": 7.065217391304348,
1064
+ "grad_norm": 0.24804139137268066,
1065
+ "learning_rate": 0.00012033564827115123,
1066
+ "loss": 0.013271838426589966,
1067
+ "step": 2275
1068
+ },
1069
+ {
1070
+ "epoch": 7.142857142857143,
1071
+ "grad_norm": 0.34166398644447327,
1072
+ "learning_rate": 0.0001145911723544914,
1073
+ "loss": 0.009375990629196166,
1074
+ "step": 2300
1075
+ },
1076
+ {
1077
+ "epoch": 7.220496894409938,
1078
+ "grad_norm": 0.8205583691596985,
1079
+ "learning_rate": 0.00010894614611929701,
1080
+ "loss": 0.014302024841308594,
1081
+ "step": 2325
1082
+ },
1083
+ {
1084
+ "epoch": 7.298136645962733,
1085
+ "grad_norm": 0.10090164840221405,
1086
+ "learning_rate": 0.00010340471549955708,
1087
+ "loss": 0.013667027950286865,
1088
+ "step": 2350
1089
+ },
1090
+ {
1091
+ "epoch": 7.375776397515528,
1092
+ "grad_norm": 0.38398560881614685,
1093
+ "learning_rate": 9.797095034447984e-05,
1094
+ "loss": 0.008717143535614013,
1095
+ "step": 2375
1096
+ },
1097
+ {
1098
+ "epoch": 7.453416149068323,
1099
+ "grad_norm": 0.4515543580055237,
1100
+ "learning_rate": 9.264884142943237e-05,
1101
+ "loss": 0.009974555373191833,
1102
+ "step": 2400
1103
+ },
1104
+ {
1105
+ "epoch": 7.5310559006211175,
1106
+ "grad_norm": 0.18586605787277222,
1107
+ "learning_rate": 8.744229752495425e-05,
1108
+ "loss": 0.011132798194885253,
1109
+ "step": 2425
1110
+ },
1111
+ {
1112
+ "epoch": 7.608695652173913,
1113
+ "grad_norm": 0.8527100682258606,
1114
+ "learning_rate": 8.235514252600023e-05,
1115
+ "loss": 0.013178889751434325,
1116
+ "step": 2450
1117
+ },
1118
+ {
1119
+ "epoch": 7.686335403726708,
1120
+ "grad_norm": 0.08932485431432724,
1121
+ "learning_rate": 7.739111264351752e-05,
1122
+ "loss": 0.008569929599761963,
1123
+ "step": 2475
1124
+ },
1125
+ {
1126
+ "epoch": 7.763975155279503,
1127
+ "grad_norm": 0.7609089612960815,
1128
+ "learning_rate": 7.255385366042274e-05,
1129
+ "loss": 0.01057699203491211,
1130
+ "step": 2500
1131
+ },
1132
+ {
1133
+ "epoch": 7.841614906832298,
1134
+ "grad_norm": 0.30278390645980835,
1135
+ "learning_rate": 6.784691825399267e-05,
1136
+ "loss": 0.007927331924438476,
1137
+ "step": 2525
1138
+ },
1139
+ {
1140
+ "epoch": 7.919254658385093,
1141
+ "grad_norm": 0.3382692039012909,
1142
+ "learning_rate": 6.327376338663493e-05,
1143
+ "loss": 0.012843213081359862,
1144
+ "step": 2550
1145
+ },
1146
+ {
1147
+ "epoch": 7.996894409937888,
1148
+ "grad_norm": 1.0424972772598267,
1149
+ "learning_rate": 5.883774776695652e-05,
1150
+ "loss": 0.010109683275222778,
1151
+ "step": 2575
1152
+ },
1153
+ {
1154
+ "epoch": 8.0,
1155
+ "eval_class_account_number_f1": 0.9779735682819383,
1156
+ "eval_class_account_number_gold_spans": 113.0,
1157
+ "eval_class_account_number_precision": 0.9736842105263158,
1158
+ "eval_class_account_number_pred_spans": 114.0,
1159
+ "eval_class_account_number_recall": 0.9823008849557522,
1160
+ "eval_class_ip_address_f1": 1.0,
1161
+ "eval_class_ip_address_gold_spans": 4.0,
1162
+ "eval_class_ip_address_precision": 1.0,
1163
+ "eval_class_ip_address_pred_spans": 4.0,
1164
+ "eval_class_ip_address_recall": 1.0,
1165
+ "eval_class_personal_handle_f1": 0.8214285714285714,
1166
+ "eval_class_personal_handle_gold_spans": 28.0,
1167
+ "eval_class_personal_handle_precision": 0.8214285714285714,
1168
+ "eval_class_personal_handle_pred_spans": 28.0,
1169
+ "eval_class_personal_handle_recall": 0.8214285714285714,
1170
+ "eval_class_private_address_f1": 0.6888888888888889,
1171
+ "eval_class_private_address_gold_spans": 48.0,
1172
+ "eval_class_private_address_precision": 0.7380952380952381,
1173
+ "eval_class_private_address_pred_spans": 42.0,
1174
+ "eval_class_private_address_recall": 0.6458333333333334,
1175
+ "eval_class_private_date_f1": 1.0,
1176
+ "eval_class_private_date_gold_spans": 33.0,
1177
+ "eval_class_private_date_precision": 1.0,
1178
+ "eval_class_private_date_pred_spans": 33.0,
1179
+ "eval_class_private_date_recall": 1.0,
1180
+ "eval_class_private_email_f1": 0.9500000000000001,
1181
+ "eval_class_private_email_gold_spans": 39.0,
1182
+ "eval_class_private_email_precision": 0.926829268292683,
1183
+ "eval_class_private_email_pred_spans": 41.0,
1184
+ "eval_class_private_email_recall": 0.9743589743589743,
1185
+ "eval_class_private_person_f1": 0.6230529595015576,
1186
+ "eval_class_private_person_gold_spans": 163.0,
1187
+ "eval_class_private_person_precision": 0.6329113924050633,
1188
+ "eval_class_private_person_pred_spans": 158.0,
1189
+ "eval_class_private_person_recall": 0.6134969325153374,
1190
+ "eval_class_private_phone_f1": 1.0,
1191
+ "eval_class_private_phone_gold_spans": 69.0,
1192
+ "eval_class_private_phone_precision": 1.0,
1193
+ "eval_class_private_phone_pred_spans": 69.0,
1194
+ "eval_class_private_phone_recall": 1.0,
1195
+ "eval_class_private_url_f1": 0.9583333333333334,
1196
+ "eval_class_private_url_gold_spans": 23.0,
1197
+ "eval_class_private_url_precision": 0.92,
1198
+ "eval_class_private_url_pred_spans": 25.0,
1199
+ "eval_class_private_url_recall": 1.0,
1200
+ "eval_gold_spans": 520.0,
1201
+ "eval_loss": 0.09201034158468246,
1202
+ "eval_pred_spans": 514.0,
1203
+ "eval_runtime": 6.9909,
1204
+ "eval_samples_per_second": 318.559,
1205
+ "eval_span_f1": 0.8355899419729208,
1206
+ "eval_span_precision": 0.8404669260700389,
1207
+ "eval_span_recall": 0.8307692307692308,
1208
+ "eval_steps_per_second": 2.575,
1209
+ "eval_token_accuracy": 0.9901960784313726,
1210
+ "step": 2576
1211
+ },
1212
+ {
1213
+ "epoch": 8.074534161490684,
1214
+ "grad_norm": 0.1411687433719635,
1215
+ "learning_rate": 5.454212938299255e-05,
1216
+ "loss": 0.009849590063095093,
1217
+ "step": 2600
1218
+ },
1219
+ {
1220
+ "epoch": 8.152173913043478,
1221
+ "grad_norm": 0.3560449182987213,
1222
+ "learning_rate": 5.0390063109409914e-05,
1223
+ "loss": 0.007954022884368896,
1224
+ "step": 2625
1225
+ },
1226
+ {
1227
+ "epoch": 8.229813664596273,
1228
+ "grad_norm": 0.36226364970207214,
1229
+ "learning_rate": 4.6384598390439706e-05,
1230
+ "loss": 0.005342819690704346,
1231
+ "step": 2650
1232
+ },
1233
+ {
1234
+ "epoch": 8.307453416149068,
1235
+ "grad_norm": 0.5297871232032776,
1236
+ "learning_rate": 4.252867700024374e-05,
1237
+ "loss": 0.00969659686088562,
1238
+ "step": 2675
1239
+ },
1240
+ {
1241
+ "epoch": 8.385093167701863,
1242
+ "grad_norm": 0.5881589651107788,
1243
+ "learning_rate": 3.882513088235676e-05,
1244
+ "loss": 0.006519293785095215,
1245
+ "step": 2700
1246
+ },
1247
+ {
1248
+ "epoch": 8.462732919254659,
1249
+ "grad_norm": 0.50627201795578,
1250
+ "learning_rate": 3.5276680069793705e-05,
1251
+ "loss": 0.00985595166683197,
1252
+ "step": 2725
1253
+ },
1254
+ {
1255
+ "epoch": 8.540372670807454,
1256
+ "grad_norm": 0.17046724259853363,
1257
+ "learning_rate": 3.18859306873483e-05,
1258
+ "loss": 0.004720044732093811,
1259
+ "step": 2750
1260
+ },
1261
+ {
1262
+ "epoch": 8.618012422360248,
1263
+ "grad_norm": 0.294168084859848,
1264
+ "learning_rate": 2.8655373037550058e-05,
1265
+ "loss": 0.011722609996795655,
1266
+ "step": 2775
1267
+ },
1268
+ {
1269
+ "epoch": 8.695652173913043,
1270
+ "grad_norm": 0.7373964190483093,
1271
+ "learning_rate": 2.5587379771686403e-05,
1272
+ "loss": 0.006650465726852417,
1273
+ "step": 2800
1274
+ },
1275
+ {
1276
+ "epoch": 8.773291925465838,
1277
+ "grad_norm": 0.4546540081501007,
1278
+ "learning_rate": 2.2684204147231934e-05,
1279
+ "loss": 0.01016199827194214,
1280
+ "step": 2825
1281
+ },
1282
+ {
1283
+ "epoch": 8.850931677018634,
1284
+ "grad_norm": 0.9172267317771912,
1285
+ "learning_rate": 1.9947978372965804e-05,
1286
+ "loss": 0.009012922048568725,
1287
+ "step": 2850
1288
+ },
1289
+ {
1290
+ "epoch": 8.928571428571429,
1291
+ "grad_norm": 0.29897835850715637,
1292
+ "learning_rate": 1.7380712042992184e-05,
1293
+ "loss": 0.005773799419403076,
1294
+ "step": 2875
1295
+ },
1296
+ {
1297
+ "epoch": 9.0,
1298
+ "eval_class_account_number_f1": 0.9779735682819383,
1299
+ "eval_class_account_number_gold_spans": 113.0,
1300
+ "eval_class_account_number_precision": 0.9736842105263158,
1301
+ "eval_class_account_number_pred_spans": 114.0,
1302
+ "eval_class_account_number_recall": 0.9823008849557522,
1303
+ "eval_class_ip_address_f1": 1.0,
1304
+ "eval_class_ip_address_gold_spans": 4.0,
1305
+ "eval_class_ip_address_precision": 1.0,
1306
+ "eval_class_ip_address_pred_spans": 4.0,
1307
+ "eval_class_ip_address_recall": 1.0,
1308
+ "eval_class_personal_handle_f1": 0.8928571428571429,
1309
+ "eval_class_personal_handle_gold_spans": 28.0,
1310
+ "eval_class_personal_handle_precision": 0.8928571428571429,
1311
+ "eval_class_personal_handle_pred_spans": 28.0,
1312
+ "eval_class_personal_handle_recall": 0.8928571428571429,
1313
+ "eval_class_private_address_f1": 0.7032967032967034,
1314
+ "eval_class_private_address_gold_spans": 48.0,
1315
+ "eval_class_private_address_precision": 0.7441860465116279,
1316
+ "eval_class_private_address_pred_spans": 43.0,
1317
+ "eval_class_private_address_recall": 0.6666666666666666,
1318
+ "eval_class_private_date_f1": 1.0,
1319
+ "eval_class_private_date_gold_spans": 33.0,
1320
+ "eval_class_private_date_precision": 1.0,
1321
+ "eval_class_private_date_pred_spans": 33.0,
1322
+ "eval_class_private_date_recall": 1.0,
1323
+ "eval_class_private_email_f1": 0.9500000000000001,
1324
+ "eval_class_private_email_gold_spans": 39.0,
1325
+ "eval_class_private_email_precision": 0.926829268292683,
1326
+ "eval_class_private_email_pred_spans": 41.0,
1327
+ "eval_class_private_email_recall": 0.9743589743589743,
1328
+ "eval_class_private_person_f1": 0.6412698412698412,
1329
+ "eval_class_private_person_gold_spans": 163.0,
1330
+ "eval_class_private_person_precision": 0.6644736842105263,
1331
+ "eval_class_private_person_pred_spans": 152.0,
1332
+ "eval_class_private_person_recall": 0.6196319018404908,
1333
+ "eval_class_private_phone_f1": 1.0,
1334
+ "eval_class_private_phone_gold_spans": 69.0,
1335
+ "eval_class_private_phone_precision": 1.0,
1336
+ "eval_class_private_phone_pred_spans": 69.0,
1337
+ "eval_class_private_phone_recall": 1.0,
1338
+ "eval_class_private_url_f1": 0.9583333333333334,
1339
+ "eval_class_private_url_gold_spans": 23.0,
1340
+ "eval_class_private_url_precision": 0.92,
1341
+ "eval_class_private_url_pred_spans": 25.0,
1342
+ "eval_class_private_url_recall": 1.0,
1343
+ "eval_gold_spans": 520.0,
1344
+ "eval_loss": 0.09796658158302307,
1345
+ "eval_pred_spans": 509.0,
1346
+ "eval_runtime": 7.2699,
1347
+ "eval_samples_per_second": 306.33,
1348
+ "eval_span_f1": 0.847424684159378,
1349
+ "eval_span_precision": 0.8565815324165029,
1350
+ "eval_span_recall": 0.8384615384615385,
1351
+ "eval_steps_per_second": 2.476,
1352
+ "eval_token_accuracy": 0.9907504668534081,
1353
+ "step": 2898
1354
+ },
1355
+ {
1356
+ "epoch": 9.006211180124224,
1357
+ "grad_norm": 0.1876034289598465,
1358
+ "learning_rate": 1.498429066081311e-05,
1359
+ "loss": 0.008111558556556701,
1360
+ "step": 2900
1361
+ },
1362
+ {
1363
+ "epoch": 9.083850931677018,
1364
+ "grad_norm": 0.23115915060043335,
1365
+ "learning_rate": 1.276047425453955e-05,
1366
+ "loss": 0.007400209903717041,
1367
+ "step": 2925
1368
+ },
1369
+ {
1370
+ "epoch": 9.161490683229813,
1371
+ "grad_norm": 0.30637550354003906,
1372
+ "learning_rate": 1.0710896084255684e-05,
1373
+ "loss": 0.007166728377342224,
1374
+ "step": 2950
1375
+ },
1376
+ {
1377
+ "epoch": 9.23913043478261,
1378
+ "grad_norm": 0.2221984565258026,
1379
+ "learning_rate": 8.837061442487193e-06,
1380
+ "loss": 0.006242750883102417,
1381
+ "step": 2975
1382
+ },
1383
+ {
1384
+ "epoch": 9.316770186335404,
1385
+ "grad_norm": 0.38621774315834045,
1386
+ "learning_rate": 7.1403465486540584e-06,
1387
+ "loss": 0.008013453483581543,
1388
+ "step": 3000
1389
+ },
1390
+ {
1391
+ "epoch": 9.394409937888199,
1392
+ "grad_norm": 0.2457902729511261,
1393
+ "learning_rate": 5.6219975383197675e-06,
1394
+ "loss": 0.008517981171607972,
1395
+ "step": 3025
1396
+ },
1397
+ {
1398
+ "epoch": 9.472049689440993,
1399
+ "grad_norm": 0.5641500353813171,
1400
+ "learning_rate": 4.2831295479791034e-06,
1401
+ "loss": 0.006109429001808167,
1402
+ "step": 3050
1403
+ },
1404
+ {
1405
+ "epoch": 9.549689440993788,
1406
+ "grad_norm": 0.18688508868217468,
1407
+ "learning_rate": 3.1247258960573067e-06,
1408
+ "loss": 0.005958260297775269,
1409
+ "step": 3075
1410
+ },
1411
+ {
1412
+ "epoch": 9.627329192546584,
1413
+ "grad_norm": 0.2731750011444092,
1414
+ "learning_rate": 2.1476373607214882e-06,
1415
+ "loss": 0.009225964546203613,
1416
+ "step": 3100
1417
+ },
1418
+ {
1419
+ "epoch": 9.70496894409938,
1420
+ "grad_norm": 0.12572576105594635,
1421
+ "learning_rate": 1.3525815550349262e-06,
1422
+ "loss": 0.0062656128406524654,
1423
+ "step": 3125
1424
+ },
1425
+ {
1426
+ "epoch": 9.782608695652174,
1427
+ "grad_norm": 0.28346896171569824,
1428
+ "learning_rate": 7.401423999132406e-07,
1429
+ "loss": 0.0059302222728729245,
1430
+ "step": 3150
1431
+ },
1432
+ {
1433
+ "epoch": 9.860248447204969,
1434
+ "grad_norm": 0.4496349096298218,
1435
+ "learning_rate": 3.107696952694139e-07,
1436
+ "loss": 0.005810813903808594,
1437
+ "step": 3175
1438
+ },
1439
+ {
1440
+ "epoch": 9.937888198757763,
1441
+ "grad_norm": 0.7214242815971375,
1442
+ "learning_rate": 6.477878966271167e-08,
1443
+ "loss": 0.008135276436805726,
1444
+ "step": 3200
1445
+ },
1446
+ {
1447
+ "epoch": 10.0,
1448
+ "eval_class_account_number_f1": 0.9779735682819383,
1449
+ "eval_class_account_number_gold_spans": 113.0,
1450
+ "eval_class_account_number_precision": 0.9736842105263158,
1451
+ "eval_class_account_number_pred_spans": 114.0,
1452
+ "eval_class_account_number_recall": 0.9823008849557522,
1453
+ "eval_class_ip_address_f1": 1.0,
1454
+ "eval_class_ip_address_gold_spans": 4.0,
1455
+ "eval_class_ip_address_precision": 1.0,
1456
+ "eval_class_ip_address_pred_spans": 4.0,
1457
+ "eval_class_ip_address_recall": 1.0,
1458
+ "eval_class_personal_handle_f1": 0.8928571428571429,
1459
+ "eval_class_personal_handle_gold_spans": 28.0,
1460
+ "eval_class_personal_handle_precision": 0.8928571428571429,
1461
+ "eval_class_personal_handle_pred_spans": 28.0,
1462
+ "eval_class_personal_handle_recall": 0.8928571428571429,
1463
+ "eval_class_private_address_f1": 0.7032967032967034,
1464
+ "eval_class_private_address_gold_spans": 48.0,
1465
+ "eval_class_private_address_precision": 0.7441860465116279,
1466
+ "eval_class_private_address_pred_spans": 43.0,
1467
+ "eval_class_private_address_recall": 0.6666666666666666,
1468
+ "eval_class_private_date_f1": 1.0,
1469
+ "eval_class_private_date_gold_spans": 33.0,
1470
+ "eval_class_private_date_precision": 1.0,
1471
+ "eval_class_private_date_pred_spans": 33.0,
1472
+ "eval_class_private_date_recall": 1.0,
1473
+ "eval_class_private_email_f1": 0.9500000000000001,
1474
+ "eval_class_private_email_gold_spans": 39.0,
1475
+ "eval_class_private_email_precision": 0.926829268292683,
1476
+ "eval_class_private_email_pred_spans": 41.0,
1477
+ "eval_class_private_email_recall": 0.9743589743589743,
1478
+ "eval_class_private_person_f1": 0.6435331230283912,
1479
+ "eval_class_private_person_gold_spans": 163.0,
1480
+ "eval_class_private_person_precision": 0.6623376623376623,
1481
+ "eval_class_private_person_pred_spans": 154.0,
1482
+ "eval_class_private_person_recall": 0.6257668711656442,
1483
+ "eval_class_private_phone_f1": 1.0,
1484
+ "eval_class_private_phone_gold_spans": 69.0,
1485
+ "eval_class_private_phone_precision": 1.0,
1486
+ "eval_class_private_phone_pred_spans": 69.0,
1487
+ "eval_class_private_phone_recall": 1.0,
1488
+ "eval_class_private_url_f1": 0.9583333333333334,
1489
+ "eval_class_private_url_gold_spans": 23.0,
1490
+ "eval_class_private_url_precision": 0.92,
1491
+ "eval_class_private_url_pred_spans": 25.0,
1492
+ "eval_class_private_url_recall": 1.0,
1493
+ "eval_gold_spans": 520.0,
1494
+ "eval_loss": 0.09789847582578659,
1495
+ "eval_pred_spans": 511.0,
1496
+ "eval_runtime": 7.0696,
1497
+ "eval_samples_per_second": 315.012,
1498
+ "eval_span_f1": 0.8477206595538312,
1499
+ "eval_span_precision": 0.8551859099804305,
1500
+ "eval_span_recall": 0.8403846153846154,
1501
+ "eval_steps_per_second": 2.546,
1502
+ "eval_token_accuracy": 0.9908088235294118,
1503
+ "step": 3220
1504
+ }
1505
+ ],
1506
+ "logging_steps": 25,
1507
+ "max_steps": 3220,
1508
+ "num_input_tokens_seen": 0,
1509
+ "num_train_epochs": 10,
1510
+ "save_steps": 500,
1511
+ "stateful_callbacks": {
1512
+ "EarlyStoppingCallback": {
1513
+ "args": {
1514
+ "early_stopping_patience": 3,
1515
+ "early_stopping_threshold": 0.0
1516
+ },
1517
+ "attributes": {
1518
+ "early_stopping_patience_counter": 0
1519
+ }
1520
+ },
1521
+ "TrainerControl": {
1522
+ "args": {
1523
+ "should_epoch_stop": false,
1524
+ "should_evaluate": false,
1525
+ "should_log": false,
1526
+ "should_save": true,
1527
+ "should_training_stop": true
1528
+ },
1529
+ "attributes": {}
1530
+ }
1531
+ },
1532
+ "total_flos": 1.291714856788951e+17,
1533
+ "train_batch_size": 64,
1534
+ "trial_name": null,
1535
+ "trial_params": null
1536
+ }
checkpoint-3220/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:195eed197af2b70c2fd47223db6e41715897a271a1bcbac65be2af77ec79752c
3
+ size 4920
config.json ADDED
@@ -0,0 +1,128 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "OpenAIPrivacyFilterForTokenClassification"
4
+ ],
5
+ "attention_bias": true,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": null,
8
+ "classifier_dropout": 0.0,
9
+ "default_n_ctx": 128000,
10
+ "dtype": "bfloat16",
11
+ "eos_token_id": 199999,
12
+ "head_dim": 64,
13
+ "hidden_act": "silu",
14
+ "hidden_size": 640,
15
+ "id2label": {
16
+ "0": "O",
17
+ "1": "B-private_person",
18
+ "2": "I-private_person",
19
+ "3": "E-private_person",
20
+ "4": "S-private_person",
21
+ "5": "B-personal_handle",
22
+ "6": "I-personal_handle",
23
+ "7": "E-personal_handle",
24
+ "8": "S-personal_handle",
25
+ "9": "B-private_phone",
26
+ "10": "I-private_phone",
27
+ "11": "E-private_phone",
28
+ "12": "S-private_phone",
29
+ "13": "B-private_email",
30
+ "14": "I-private_email",
31
+ "15": "E-private_email",
32
+ "16": "S-private_email",
33
+ "17": "B-private_address",
34
+ "18": "I-private_address",
35
+ "19": "E-private_address",
36
+ "20": "S-private_address",
37
+ "21": "B-private_date",
38
+ "22": "I-private_date",
39
+ "23": "E-private_date",
40
+ "24": "S-private_date",
41
+ "25": "B-private_url",
42
+ "26": "I-private_url",
43
+ "27": "E-private_url",
44
+ "28": "S-private_url",
45
+ "29": "B-account_number",
46
+ "30": "I-account_number",
47
+ "31": "E-account_number",
48
+ "32": "S-account_number",
49
+ "33": "B-ip_address",
50
+ "34": "I-ip_address",
51
+ "35": "E-ip_address",
52
+ "36": "S-ip_address"
53
+ },
54
+ "initial_context_length": 4096,
55
+ "initializer_range": 0.02,
56
+ "intermediate_size": 640,
57
+ "label2id": {
58
+ "B-account_number": 29,
59
+ "B-ip_address": 33,
60
+ "B-personal_handle": 5,
61
+ "B-private_address": 17,
62
+ "B-private_date": 21,
63
+ "B-private_email": 13,
64
+ "B-private_person": 1,
65
+ "B-private_phone": 9,
66
+ "B-private_url": 25,
67
+ "E-account_number": 31,
68
+ "E-ip_address": 35,
69
+ "E-personal_handle": 7,
70
+ "E-private_address": 19,
71
+ "E-private_date": 23,
72
+ "E-private_email": 15,
73
+ "E-private_person": 3,
74
+ "E-private_phone": 11,
75
+ "E-private_url": 27,
76
+ "I-account_number": 30,
77
+ "I-ip_address": 34,
78
+ "I-personal_handle": 6,
79
+ "I-private_address": 18,
80
+ "I-private_date": 22,
81
+ "I-private_email": 14,
82
+ "I-private_person": 2,
83
+ "I-private_phone": 10,
84
+ "I-private_url": 26,
85
+ "O": 0,
86
+ "S-account_number": 32,
87
+ "S-ip_address": 36,
88
+ "S-personal_handle": 8,
89
+ "S-private_address": 20,
90
+ "S-private_date": 24,
91
+ "S-private_email": 16,
92
+ "S-private_person": 4,
93
+ "S-private_phone": 12,
94
+ "S-private_url": 28
95
+ },
96
+ "max_position_embeddings": 131072,
97
+ "model_type": "openai_privacy_filter",
98
+ "num_attention_heads": 14,
99
+ "num_experts_per_tok": 4,
100
+ "num_hidden_layers": 8,
101
+ "num_key_value_heads": 2,
102
+ "num_local_experts": 128,
103
+ "output_router_logits": false,
104
+ "pad_token_id": 199999,
105
+ "rms_norm_eps": 1e-05,
106
+ "rope_parameters": {
107
+ "beta_fast": 32.0,
108
+ "beta_slow": 1.0,
109
+ "factor": 32.0,
110
+ "original_max_position_embeddings": 4096,
111
+ "rope_theta": 150000.0,
112
+ "rope_type": "yarn",
113
+ "truncate": false
114
+ },
115
+ "router_aux_loss_coef": 0.001,
116
+ "sliding_window": 128,
117
+ "tie_word_embeddings": false,
118
+ "transformers.js_config": {
119
+ "use_external_data_format": {
120
+ "model": 1,
121
+ "model.onnx": 3,
122
+ "model_fp16.onnx": 2
123
+ }
124
+ },
125
+ "transformers_version": "5.7.0.dev0",
126
+ "use_cache": false,
127
+ "vocab_size": 200064
128
+ }
label_space.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "category_version": "ko_pii_v4",
3
+ "span_class_names": [
4
+ "O",
5
+ "private_person",
6
+ "personal_handle",
7
+ "private_phone",
8
+ "private_email",
9
+ "private_address",
10
+ "private_date",
11
+ "private_url",
12
+ "account_number",
13
+ "ip_address"
14
+ ]
15
+ }
log_history.json ADDED
@@ -0,0 +1,1622 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "loss": 1.4503676891326904,
4
+ "grad_norm": 8.771296501159668,
5
+ "learning_rate": 0.0,
6
+ "epoch": 0.003105590062111801,
7
+ "step": 1
8
+ },
9
+ {
10
+ "loss": 1.2439465522766113,
11
+ "grad_norm": 8.387271881103516,
12
+ "learning_rate": 3.7267080745341614e-05,
13
+ "epoch": 0.07763975155279502,
14
+ "step": 25
15
+ },
16
+ {
17
+ "loss": 0.9780036926269531,
18
+ "grad_norm": 4.6111741065979,
19
+ "learning_rate": 7.608695652173914e-05,
20
+ "epoch": 0.15527950310559005,
21
+ "step": 50
22
+ },
23
+ {
24
+ "loss": 0.5338097381591796,
25
+ "grad_norm": 1.846106767654419,
26
+ "learning_rate": 0.00011490683229813665,
27
+ "epoch": 0.2329192546583851,
28
+ "step": 75
29
+ },
30
+ {
31
+ "loss": 0.34235782623291017,
32
+ "grad_norm": 2.225909471511841,
33
+ "learning_rate": 0.00015372670807453417,
34
+ "epoch": 0.3105590062111801,
35
+ "step": 100
36
+ },
37
+ {
38
+ "loss": 0.2744290542602539,
39
+ "grad_norm": 2.14986252784729,
40
+ "learning_rate": 0.0001925465838509317,
41
+ "epoch": 0.38819875776397517,
42
+ "step": 125
43
+ },
44
+ {
45
+ "loss": 0.2258339309692383,
46
+ "grad_norm": 2.1372387409210205,
47
+ "learning_rate": 0.0002313664596273292,
48
+ "epoch": 0.4658385093167702,
49
+ "step": 150
50
+ },
51
+ {
52
+ "loss": 0.2162160873413086,
53
+ "grad_norm": 1.5849782228469849,
54
+ "learning_rate": 0.0002701863354037267,
55
+ "epoch": 0.5434782608695652,
56
+ "step": 175
57
+ },
58
+ {
59
+ "loss": 0.20915340423583983,
60
+ "grad_norm": 1.6991852521896362,
61
+ "learning_rate": 0.00030900621118012424,
62
+ "epoch": 0.6211180124223602,
63
+ "step": 200
64
+ },
65
+ {
66
+ "loss": 0.18431522369384765,
67
+ "grad_norm": 2.723663568496704,
68
+ "learning_rate": 0.00034782608695652176,
69
+ "epoch": 0.6987577639751553,
70
+ "step": 225
71
+ },
72
+ {
73
+ "loss": 0.17637535095214843,
74
+ "grad_norm": 1.2743035554885864,
75
+ "learning_rate": 0.00038664596273291923,
76
+ "epoch": 0.7763975155279503,
77
+ "step": 250
78
+ },
79
+ {
80
+ "loss": 0.16074188232421874,
81
+ "grad_norm": 3.452497720718384,
82
+ "learning_rate": 0.0004254658385093168,
83
+ "epoch": 0.8540372670807453,
84
+ "step": 275
85
+ },
86
+ {
87
+ "loss": 0.1531696319580078,
88
+ "grad_norm": 2.547046661376953,
89
+ "learning_rate": 0.00046428571428571433,
90
+ "epoch": 0.9316770186335404,
91
+ "step": 300
92
+ },
93
+ {
94
+ "eval_loss": 0.1431565284729004,
95
+ "eval_token_accuracy": 0.9817635387488328,
96
+ "eval_span_precision": 0.7254004576659039,
97
+ "eval_span_recall": 0.6096153846153847,
98
+ "eval_span_f1": 0.6624869383490073,
99
+ "eval_gold_spans": 520.0,
100
+ "eval_pred_spans": 437.0,
101
+ "eval_class_account_number_precision": 0.6744186046511628,
102
+ "eval_class_account_number_recall": 0.7699115044247787,
103
+ "eval_class_account_number_f1": 0.71900826446281,
104
+ "eval_class_account_number_gold_spans": 113.0,
105
+ "eval_class_account_number_pred_spans": 129.0,
106
+ "eval_class_ip_address_precision": 1.0,
107
+ "eval_class_ip_address_recall": 1.0,
108
+ "eval_class_ip_address_f1": 1.0,
109
+ "eval_class_ip_address_gold_spans": 4.0,
110
+ "eval_class_ip_address_pred_spans": 4.0,
111
+ "eval_class_personal_handle_precision": 0.7931034482758621,
112
+ "eval_class_personal_handle_recall": 0.8214285714285714,
113
+ "eval_class_personal_handle_f1": 0.8070175438596492,
114
+ "eval_class_personal_handle_gold_spans": 28.0,
115
+ "eval_class_personal_handle_pred_spans": 29.0,
116
+ "eval_class_private_address_precision": 0.13043478260869565,
117
+ "eval_class_private_address_recall": 0.0625,
118
+ "eval_class_private_address_f1": 0.08450704225352113,
119
+ "eval_class_private_address_gold_spans": 48.0,
120
+ "eval_class_private_address_pred_spans": 23.0,
121
+ "eval_class_private_date_precision": 0.8611111111111112,
122
+ "eval_class_private_date_recall": 0.9393939393939394,
123
+ "eval_class_private_date_f1": 0.8985507246376813,
124
+ "eval_class_private_date_gold_spans": 33.0,
125
+ "eval_class_private_date_pred_spans": 36.0,
126
+ "eval_class_private_email_precision": 0.95,
127
+ "eval_class_private_email_recall": 0.9743589743589743,
128
+ "eval_class_private_email_f1": 0.9620253164556962,
129
+ "eval_class_private_email_gold_spans": 39.0,
130
+ "eval_class_private_email_pred_spans": 40.0,
131
+ "eval_class_private_person_precision": 0.47619047619047616,
132
+ "eval_class_private_person_recall": 0.24539877300613497,
133
+ "eval_class_private_person_f1": 0.32388663967611336,
134
+ "eval_class_private_person_gold_spans": 163.0,
135
+ "eval_class_private_person_pred_spans": 84.0,
136
+ "eval_class_private_phone_precision": 0.9855072463768116,
137
+ "eval_class_private_phone_recall": 0.9855072463768116,
138
+ "eval_class_private_phone_f1": 0.9855072463768116,
139
+ "eval_class_private_phone_gold_spans": 69.0,
140
+ "eval_class_private_phone_pred_spans": 69.0,
141
+ "eval_class_private_url_precision": 1.0,
142
+ "eval_class_private_url_recall": 1.0,
143
+ "eval_class_private_url_f1": 1.0,
144
+ "eval_class_private_url_gold_spans": 23.0,
145
+ "eval_class_private_url_pred_spans": 23.0,
146
+ "eval_runtime": 5.911,
147
+ "eval_samples_per_second": 376.753,
148
+ "eval_steps_per_second": 3.045,
149
+ "epoch": 1.0,
150
+ "step": 322
151
+ },
152
+ {
153
+ "loss": 0.1343870735168457,
154
+ "grad_norm": 1.0337207317352295,
155
+ "learning_rate": 0.000499999412412136,
156
+ "epoch": 1.0093167701863355,
157
+ "step": 325
158
+ },
159
+ {
160
+ "loss": 0.1398656463623047,
161
+ "grad_norm": 1.150405764579773,
162
+ "learning_rate": 0.0004998929197148332,
163
+ "epoch": 1.0869565217391304,
164
+ "step": 350
165
+ },
166
+ {
167
+ "loss": 0.11556500434875489,
168
+ "grad_norm": 1.239856243133545,
169
+ "learning_rate": 0.0004996028956208098,
170
+ "epoch": 1.1645962732919255,
171
+ "step": 375
172
+ },
173
+ {
174
+ "loss": 0.10652177810668945,
175
+ "grad_norm": 4.1494951248168945,
176
+ "learning_rate": 0.0004991295531354085,
177
+ "epoch": 1.2422360248447206,
178
+ "step": 400
179
+ },
180
+ {
181
+ "loss": 0.1127833080291748,
182
+ "grad_norm": 1.820974349975586,
183
+ "learning_rate": 0.0004984732399003618,
184
+ "epoch": 1.3198757763975155,
185
+ "step": 425
186
+ },
187
+ {
188
+ "loss": 0.1028218650817871,
189
+ "grad_norm": 0.9540935158729553,
190
+ "learning_rate": 0.0004976344379384687,
191
+ "epoch": 1.3975155279503104,
192
+ "step": 450
193
+ },
194
+ {
195
+ "loss": 0.11593814849853516,
196
+ "grad_norm": 1.7447997331619263,
197
+ "learning_rate": 0.0004966137632995793,
198
+ "epoch": 1.4751552795031055,
199
+ "step": 475
200
+ },
201
+ {
202
+ "loss": 0.1070643424987793,
203
+ "grad_norm": 1.3501027822494507,
204
+ "learning_rate": 0.0004954119656081417,
205
+ "epoch": 1.5527950310559007,
206
+ "step": 500
207
+ },
208
+ {
209
+ "loss": 0.11064315795898437,
210
+ "grad_norm": 0.6298630237579346,
211
+ "learning_rate": 0.0004940299275126486,
212
+ "epoch": 1.6304347826086958,
213
+ "step": 525
214
+ },
215
+ {
216
+ "loss": 0.09777620315551758,
217
+ "grad_norm": 0.8390702605247498,
218
+ "learning_rate": 0.0004924686640373841,
219
+ "epoch": 1.7080745341614907,
220
+ "step": 550
221
+ },
222
+ {
223
+ "loss": 0.08596687316894532,
224
+ "grad_norm": 0.5763865113258362,
225
+ "learning_rate": 0.0004907293218369499,
226
+ "epoch": 1.7857142857142856,
227
+ "step": 575
228
+ },
229
+ {
230
+ "loss": 0.0955750274658203,
231
+ "grad_norm": 0.7198231816291809,
232
+ "learning_rate": 0.0004888131783541152,
233
+ "epoch": 1.8633540372670807,
234
+ "step": 600
235
+ },
236
+ {
237
+ "loss": 0.0880333137512207,
238
+ "grad_norm": 1.0241382122039795,
239
+ "learning_rate": 0.0004867216408816122,
240
+ "epoch": 1.9409937888198758,
241
+ "step": 625
242
+ },
243
+ {
244
+ "eval_loss": 0.08562110364437103,
245
+ "eval_token_accuracy": 0.9886788048552755,
246
+ "eval_span_precision": 0.7771428571428571,
247
+ "eval_span_recall": 0.7846153846153846,
248
+ "eval_span_f1": 0.7808612440191388,
249
+ "eval_gold_spans": 520.0,
250
+ "eval_pred_spans": 525.0,
251
+ "eval_class_account_number_precision": 0.8916666666666667,
252
+ "eval_class_account_number_recall": 0.9469026548672567,
253
+ "eval_class_account_number_f1": 0.9184549356223176,
254
+ "eval_class_account_number_gold_spans": 113.0,
255
+ "eval_class_account_number_pred_spans": 120.0,
256
+ "eval_class_ip_address_precision": 1.0,
257
+ "eval_class_ip_address_recall": 1.0,
258
+ "eval_class_ip_address_f1": 1.0,
259
+ "eval_class_ip_address_gold_spans": 4.0,
260
+ "eval_class_ip_address_pred_spans": 4.0,
261
+ "eval_class_personal_handle_precision": 0.9,
262
+ "eval_class_personal_handle_recall": 0.9642857142857143,
263
+ "eval_class_personal_handle_f1": 0.9310344827586207,
264
+ "eval_class_personal_handle_gold_spans": 28.0,
265
+ "eval_class_personal_handle_pred_spans": 30.0,
266
+ "eval_class_private_address_precision": 0.6410256410256411,
267
+ "eval_class_private_address_recall": 0.5208333333333334,
268
+ "eval_class_private_address_f1": 0.5747126436781609,
269
+ "eval_class_private_address_gold_spans": 48.0,
270
+ "eval_class_private_address_pred_spans": 39.0,
271
+ "eval_class_private_date_precision": 0.8888888888888888,
272
+ "eval_class_private_date_recall": 0.9696969696969697,
273
+ "eval_class_private_date_f1": 0.927536231884058,
274
+ "eval_class_private_date_gold_spans": 33.0,
275
+ "eval_class_private_date_pred_spans": 36.0,
276
+ "eval_class_private_email_precision": 0.9743589743589743,
277
+ "eval_class_private_email_recall": 0.9743589743589743,
278
+ "eval_class_private_email_f1": 0.9743589743589743,
279
+ "eval_class_private_email_gold_spans": 39.0,
280
+ "eval_class_private_email_pred_spans": 39.0,
281
+ "eval_class_private_person_precision": 0.5151515151515151,
282
+ "eval_class_private_person_recall": 0.5214723926380368,
283
+ "eval_class_private_person_f1": 0.5182926829268292,
284
+ "eval_class_private_person_gold_spans": 163.0,
285
+ "eval_class_private_person_pred_spans": 165.0,
286
+ "eval_class_private_phone_precision": 1.0,
287
+ "eval_class_private_phone_recall": 1.0,
288
+ "eval_class_private_phone_f1": 1.0,
289
+ "eval_class_private_phone_gold_spans": 69.0,
290
+ "eval_class_private_phone_pred_spans": 69.0,
291
+ "eval_class_private_url_precision": 0.9130434782608695,
292
+ "eval_class_private_url_recall": 0.9130434782608695,
293
+ "eval_class_private_url_f1": 0.9130434782608695,
294
+ "eval_class_private_url_gold_spans": 23.0,
295
+ "eval_class_private_url_pred_spans": 23.0,
296
+ "eval_runtime": 6.7931,
297
+ "eval_samples_per_second": 327.832,
298
+ "eval_steps_per_second": 2.65,
299
+ "epoch": 2.0,
300
+ "step": 644
301
+ },
302
+ {
303
+ "loss": 0.0896837329864502,
304
+ "grad_norm": 1.2737419605255127,
305
+ "learning_rate": 0.00048445624552856343,
306
+ "epoch": 2.018633540372671,
307
+ "step": 650
308
+ },
309
+ {
310
+ "loss": 0.07990738868713379,
311
+ "grad_norm": 0.5823193788528442,
312
+ "learning_rate": 0.00048201865609230155,
313
+ "epoch": 2.0962732919254656,
314
+ "step": 675
315
+ },
316
+ {
317
+ "loss": 0.06952488899230957,
318
+ "grad_norm": 1.4650110006332397,
319
+ "learning_rate": 0.0004794106628364098,
320
+ "epoch": 2.1739130434782608,
321
+ "step": 700
322
+ },
323
+ {
324
+ "loss": 0.07833251953125,
325
+ "grad_norm": 0.6727014183998108,
326
+ "learning_rate": 0.0004766341811758803,
327
+ "epoch": 2.251552795031056,
328
+ "step": 725
329
+ },
330
+ {
331
+ "loss": 0.06939034461975098,
332
+ "grad_norm": 0.6006216406822205,
333
+ "learning_rate": 0.000473691250270357,
334
+ "epoch": 2.329192546583851,
335
+ "step": 750
336
+ },
337
+ {
338
+ "loss": 0.05780394077301025,
339
+ "grad_norm": 0.5175172090530396,
340
+ "learning_rate": 0.00047058403152649405,
341
+ "epoch": 2.406832298136646,
342
+ "step": 775
343
+ },
344
+ {
345
+ "loss": 0.06771457672119141,
346
+ "grad_norm": 1.1351436376571655,
347
+ "learning_rate": 0.00046731480701053265,
348
+ "epoch": 2.4844720496894412,
349
+ "step": 800
350
+ },
351
+ {
352
+ "loss": 0.0703080177307129,
353
+ "grad_norm": 0.8665878772735596,
354
+ "learning_rate": 0.0004638859777722596,
355
+ "epoch": 2.562111801242236,
356
+ "step": 825
357
+ },
358
+ {
359
+ "loss": 0.0754665756225586,
360
+ "grad_norm": 0.8073828220367432,
361
+ "learning_rate": 0.00046030006208158,
362
+ "epoch": 2.639751552795031,
363
+ "step": 850
364
+ },
365
+ {
366
+ "loss": 0.06543315410614013,
367
+ "grad_norm": 0.6520082950592041,
368
+ "learning_rate": 0.00045655969357899875,
369
+ "epoch": 2.717391304347826,
370
+ "step": 875
371
+ },
372
+ {
373
+ "loss": 0.07107664108276367,
374
+ "grad_norm": 1.448807954788208,
375
+ "learning_rate": 0.00045266761934136896,
376
+ "epoch": 2.795031055900621,
377
+ "step": 900
378
+ },
379
+ {
380
+ "loss": 0.07224170684814453,
381
+ "grad_norm": 0.45655015110969543,
382
+ "learning_rate": 0.00044862669786432865,
383
+ "epoch": 2.8726708074534164,
384
+ "step": 925
385
+ },
386
+ {
387
+ "loss": 0.06763762950897217,
388
+ "grad_norm": 0.4825460612773895,
389
+ "learning_rate": 0.0004444398969629067,
390
+ "epoch": 2.950310559006211,
391
+ "step": 950
392
+ },
393
+ {
394
+ "eval_loss": 0.07642154395580292,
395
+ "eval_token_accuracy": 0.9904586834733894,
396
+ "eval_span_precision": 0.8228346456692913,
397
+ "eval_span_recall": 0.8038461538461539,
398
+ "eval_span_f1": 0.8132295719844358,
399
+ "eval_gold_spans": 520.0,
400
+ "eval_pred_spans": 508.0,
401
+ "eval_class_account_number_precision": 0.9152542372881356,
402
+ "eval_class_account_number_recall": 0.9557522123893806,
403
+ "eval_class_account_number_f1": 0.9350649350649352,
404
+ "eval_class_account_number_gold_spans": 113.0,
405
+ "eval_class_account_number_pred_spans": 118.0,
406
+ "eval_class_ip_address_precision": 1.0,
407
+ "eval_class_ip_address_recall": 1.0,
408
+ "eval_class_ip_address_f1": 1.0,
409
+ "eval_class_ip_address_gold_spans": 4.0,
410
+ "eval_class_ip_address_pred_spans": 4.0,
411
+ "eval_class_personal_handle_precision": 0.896551724137931,
412
+ "eval_class_personal_handle_recall": 0.9285714285714286,
413
+ "eval_class_personal_handle_f1": 0.912280701754386,
414
+ "eval_class_personal_handle_gold_spans": 28.0,
415
+ "eval_class_personal_handle_pred_spans": 29.0,
416
+ "eval_class_private_address_precision": 0.6041666666666666,
417
+ "eval_class_private_address_recall": 0.6041666666666666,
418
+ "eval_class_private_address_f1": 0.6041666666666666,
419
+ "eval_class_private_address_gold_spans": 48.0,
420
+ "eval_class_private_address_pred_spans": 48.0,
421
+ "eval_class_private_date_precision": 0.9117647058823529,
422
+ "eval_class_private_date_recall": 0.9393939393939394,
423
+ "eval_class_private_date_f1": 0.9253731343283583,
424
+ "eval_class_private_date_gold_spans": 33.0,
425
+ "eval_class_private_date_pred_spans": 34.0,
426
+ "eval_class_private_email_precision": 0.9487179487179487,
427
+ "eval_class_private_email_recall": 0.9487179487179487,
428
+ "eval_class_private_email_f1": 0.9487179487179487,
429
+ "eval_class_private_email_gold_spans": 39.0,
430
+ "eval_class_private_email_pred_spans": 39.0,
431
+ "eval_class_private_person_precision": 0.6433566433566433,
432
+ "eval_class_private_person_recall": 0.5644171779141104,
433
+ "eval_class_private_person_f1": 0.6013071895424836,
434
+ "eval_class_private_person_gold_spans": 163.0,
435
+ "eval_class_private_person_pred_spans": 143.0,
436
+ "eval_class_private_phone_precision": 1.0,
437
+ "eval_class_private_phone_recall": 1.0,
438
+ "eval_class_private_phone_f1": 1.0,
439
+ "eval_class_private_phone_gold_spans": 69.0,
440
+ "eval_class_private_phone_pred_spans": 69.0,
441
+ "eval_class_private_url_precision": 0.9166666666666666,
442
+ "eval_class_private_url_recall": 0.9565217391304348,
443
+ "eval_class_private_url_f1": 0.9361702127659574,
444
+ "eval_class_private_url_gold_spans": 23.0,
445
+ "eval_class_private_url_pred_spans": 24.0,
446
+ "eval_runtime": 6.8956,
447
+ "eval_samples_per_second": 322.962,
448
+ "eval_steps_per_second": 2.61,
449
+ "epoch": 3.0,
450
+ "step": 966
451
+ },
452
+ {
453
+ "loss": 0.04707308769226074,
454
+ "grad_norm": 0.3684696853160858,
455
+ "learning_rate": 0.00044011029159184015,
456
+ "epoch": 3.027950310559006,
457
+ "step": 975
458
+ },
459
+ {
460
+ "loss": 0.04691335678100586,
461
+ "grad_norm": 0.47226452827453613,
462
+ "learning_rate": 0.00043564106158720383,
463
+ "epoch": 3.1055900621118013,
464
+ "step": 1000
465
+ },
466
+ {
467
+ "loss": 0.044693613052368165,
468
+ "grad_norm": 0.6678075194358826,
469
+ "learning_rate": 0.00043103548933101155,
470
+ "epoch": 3.1832298136645965,
471
+ "step": 1025
472
+ },
473
+ {
474
+ "loss": 0.04987412452697754,
475
+ "grad_norm": 0.675778865814209,
476
+ "learning_rate": 0.00042629695734050255,
477
+ "epoch": 3.260869565217391,
478
+ "step": 1050
479
+ },
480
+ {
481
+ "loss": 0.046450929641723634,
482
+ "grad_norm": 1.2661679983139038,
483
+ "learning_rate": 0.00042142894578388486,
484
+ "epoch": 3.3385093167701863,
485
+ "step": 1075
486
+ },
487
+ {
488
+ "loss": 0.050359611511230466,
489
+ "grad_norm": 0.7080872058868408,
490
+ "learning_rate": 0.00041643502992436013,
491
+ "epoch": 3.4161490683229814,
492
+ "step": 1100
493
+ },
494
+ {
495
+ "loss": 0.06056289672851563,
496
+ "grad_norm": 0.7029626369476318,
497
+ "learning_rate": 0.0004113188774943071,
498
+ "epoch": 3.4937888198757765,
499
+ "step": 1125
500
+ },
501
+ {
502
+ "loss": 0.045447168350219724,
503
+ "grad_norm": 0.558189332485199,
504
+ "learning_rate": 0.000406084246001551,
505
+ "epoch": 3.571428571428571,
506
+ "step": 1150
507
+ },
508
+ {
509
+ "loss": 0.045512027740478515,
510
+ "grad_norm": 0.5974974036216736,
511
+ "learning_rate": 0.00040073497996969933,
512
+ "epoch": 3.6490683229813663,
513
+ "step": 1175
514
+ },
515
+ {
516
+ "loss": 0.05025949954986572,
517
+ "grad_norm": 0.5707910656929016,
518
+ "learning_rate": 0.000395275008114569,
519
+ "epoch": 3.7267080745341614,
520
+ "step": 1200
521
+ },
522
+ {
523
+ "loss": 0.06197848796844482,
524
+ "grad_norm": 1.0855952501296997,
525
+ "learning_rate": 0.00038970834045878026,
526
+ "epoch": 3.8043478260869565,
527
+ "step": 1225
528
+ },
529
+ {
530
+ "loss": 0.05249831199645996,
531
+ "grad_norm": 0.3965921401977539,
532
+ "learning_rate": 0.0003840390653866341,
533
+ "epoch": 3.8819875776397517,
534
+ "step": 1250
535
+ },
536
+ {
537
+ "loss": 0.045794315338134765,
538
+ "grad_norm": 0.5562620162963867,
539
+ "learning_rate": 0.00037827134664143947,
540
+ "epoch": 3.9596273291925463,
541
+ "step": 1275
542
+ },
543
+ {
544
+ "eval_loss": 0.07150626927614212,
545
+ "eval_token_accuracy": 0.9900501867413632,
546
+ "eval_span_precision": 0.8294573643410853,
547
+ "eval_span_recall": 0.823076923076923,
548
+ "eval_span_f1": 0.8262548262548264,
549
+ "eval_gold_spans": 520.0,
550
+ "eval_pred_spans": 516.0,
551
+ "eval_class_account_number_precision": 0.9649122807017544,
552
+ "eval_class_account_number_recall": 0.9734513274336283,
553
+ "eval_class_account_number_f1": 0.9691629955947135,
554
+ "eval_class_account_number_gold_spans": 113.0,
555
+ "eval_class_account_number_pred_spans": 114.0,
556
+ "eval_class_ip_address_precision": 1.0,
557
+ "eval_class_ip_address_recall": 1.0,
558
+ "eval_class_ip_address_f1": 1.0,
559
+ "eval_class_ip_address_gold_spans": 4.0,
560
+ "eval_class_ip_address_pred_spans": 4.0,
561
+ "eval_class_personal_handle_precision": 0.7586206896551724,
562
+ "eval_class_personal_handle_recall": 0.7857142857142857,
563
+ "eval_class_personal_handle_f1": 0.7719298245614034,
564
+ "eval_class_personal_handle_gold_spans": 28.0,
565
+ "eval_class_personal_handle_pred_spans": 29.0,
566
+ "eval_class_private_address_precision": 0.6585365853658537,
567
+ "eval_class_private_address_recall": 0.5625,
568
+ "eval_class_private_address_f1": 0.6067415730337078,
569
+ "eval_class_private_address_gold_spans": 48.0,
570
+ "eval_class_private_address_pred_spans": 41.0,
571
+ "eval_class_private_date_precision": 1.0,
572
+ "eval_class_private_date_recall": 1.0,
573
+ "eval_class_private_date_f1": 1.0,
574
+ "eval_class_private_date_gold_spans": 33.0,
575
+ "eval_class_private_date_pred_spans": 33.0,
576
+ "eval_class_private_email_precision": 0.975,
577
+ "eval_class_private_email_recall": 1.0,
578
+ "eval_class_private_email_f1": 0.9873417721518987,
579
+ "eval_class_private_email_gold_spans": 39.0,
580
+ "eval_class_private_email_pred_spans": 40.0,
581
+ "eval_class_private_person_precision": 0.6257668711656442,
582
+ "eval_class_private_person_recall": 0.6257668711656442,
583
+ "eval_class_private_person_f1": 0.6257668711656442,
584
+ "eval_class_private_person_gold_spans": 163.0,
585
+ "eval_class_private_person_pred_spans": 163.0,
586
+ "eval_class_private_phone_precision": 1.0,
587
+ "eval_class_private_phone_recall": 1.0,
588
+ "eval_class_private_phone_f1": 1.0,
589
+ "eval_class_private_phone_gold_spans": 69.0,
590
+ "eval_class_private_phone_pred_spans": 69.0,
591
+ "eval_class_private_url_precision": 0.9565217391304348,
592
+ "eval_class_private_url_recall": 0.9565217391304348,
593
+ "eval_class_private_url_f1": 0.9565217391304348,
594
+ "eval_class_private_url_gold_spans": 23.0,
595
+ "eval_class_private_url_pred_spans": 23.0,
596
+ "eval_runtime": 6.9918,
597
+ "eval_samples_per_second": 318.517,
598
+ "eval_steps_per_second": 2.574,
599
+ "epoch": 4.0,
600
+ "step": 1288
601
+ },
602
+ {
603
+ "loss": 0.03683102369308472,
604
+ "grad_norm": 0.3308059573173523,
605
+ "learning_rate": 0.00037240942026749233,
606
+ "epoch": 4.037267080745342,
607
+ "step": 1300
608
+ },
609
+ {
610
+ "loss": 0.033050749301910404,
611
+ "grad_norm": 0.9377778768539429,
612
+ "learning_rate": 0.0003664575914989552,
613
+ "epoch": 4.114906832298137,
614
+ "step": 1325
615
+ },
616
+ {
617
+ "loss": 0.031362690925598145,
618
+ "grad_norm": 0.5735832452774048,
619
+ "learning_rate": 0.00036042023159791974,
620
+ "epoch": 4.192546583850931,
621
+ "step": 1350
622
+ },
623
+ {
624
+ "loss": 0.03477231979370117,
625
+ "grad_norm": 0.631287693977356,
626
+ "learning_rate": 0.0003543017746439766,
627
+ "epoch": 4.270186335403727,
628
+ "step": 1375
629
+ },
630
+ {
631
+ "loss": 0.03894762754440308,
632
+ "grad_norm": 1.0935797691345215,
633
+ "learning_rate": 0.0003481067142776489,
634
+ "epoch": 4.3478260869565215,
635
+ "step": 1400
636
+ },
637
+ {
638
+ "loss": 0.03512312889099121,
639
+ "grad_norm": 0.6073715090751648,
640
+ "learning_rate": 0.0003418396004000829,
641
+ "epoch": 4.425465838509317,
642
+ "step": 1425
643
+ },
644
+ {
645
+ "loss": 0.04299846649169922,
646
+ "grad_norm": 0.6249358654022217,
647
+ "learning_rate": 0.0003355050358314172,
648
+ "epoch": 4.503105590062112,
649
+ "step": 1450
650
+ },
651
+ {
652
+ "loss": 0.03905656576156616,
653
+ "grad_norm": 0.7248182892799377,
654
+ "learning_rate": 0.0003291076729302869,
655
+ "epoch": 4.580745341614907,
656
+ "step": 1475
657
+ },
658
+ {
659
+ "loss": 0.039488759040832516,
660
+ "grad_norm": 0.9138495922088623,
661
+ "learning_rate": 0.00032265221017694474,
662
+ "epoch": 4.658385093167702,
663
+ "step": 1500
664
+ },
665
+ {
666
+ "loss": 0.03172698974609375,
667
+ "grad_norm": 0.6601109504699707,
668
+ "learning_rate": 0.00031614338872250835,
669
+ "epoch": 4.736024844720497,
670
+ "step": 1525
671
+ },
672
+ {
673
+ "loss": 0.038228404521942136,
674
+ "grad_norm": 1.141296625137329,
675
+ "learning_rate": 0.0003095859889068687,
676
+ "epoch": 4.813664596273292,
677
+ "step": 1550
678
+ },
679
+ {
680
+ "loss": 0.03626953125,
681
+ "grad_norm": 1.3895987272262573,
682
+ "learning_rate": 0.0003029848267478158,
683
+ "epoch": 4.891304347826087,
684
+ "step": 1575
685
+ },
686
+ {
687
+ "loss": 0.03073249340057373,
688
+ "grad_norm": 0.5478597283363342,
689
+ "learning_rate": 0.00029634475040396285,
690
+ "epoch": 4.9689440993788825,
691
+ "step": 1600
692
+ },
693
+ {
694
+ "eval_loss": 0.07776352018117905,
695
+ "eval_token_accuracy": 0.9894082633053222,
696
+ "eval_span_precision": 0.7934782608695652,
697
+ "eval_span_recall": 0.8423076923076923,
698
+ "eval_span_f1": 0.8171641791044777,
699
+ "eval_gold_spans": 520.0,
700
+ "eval_pred_spans": 552.0,
701
+ "eval_class_account_number_precision": 0.9649122807017544,
702
+ "eval_class_account_number_recall": 0.9734513274336283,
703
+ "eval_class_account_number_f1": 0.9691629955947135,
704
+ "eval_class_account_number_gold_spans": 113.0,
705
+ "eval_class_account_number_pred_spans": 114.0,
706
+ "eval_class_ip_address_precision": 1.0,
707
+ "eval_class_ip_address_recall": 1.0,
708
+ "eval_class_ip_address_f1": 1.0,
709
+ "eval_class_ip_address_gold_spans": 4.0,
710
+ "eval_class_ip_address_pred_spans": 4.0,
711
+ "eval_class_personal_handle_precision": 0.8620689655172413,
712
+ "eval_class_personal_handle_recall": 0.8928571428571429,
713
+ "eval_class_personal_handle_f1": 0.8771929824561403,
714
+ "eval_class_personal_handle_gold_spans": 28.0,
715
+ "eval_class_personal_handle_pred_spans": 29.0,
716
+ "eval_class_private_address_precision": 0.6842105263157895,
717
+ "eval_class_private_address_recall": 0.5416666666666666,
718
+ "eval_class_private_address_f1": 0.6046511627906976,
719
+ "eval_class_private_address_gold_spans": 48.0,
720
+ "eval_class_private_address_pred_spans": 38.0,
721
+ "eval_class_private_date_precision": 1.0,
722
+ "eval_class_private_date_recall": 1.0,
723
+ "eval_class_private_date_f1": 1.0,
724
+ "eval_class_private_date_gold_spans": 33.0,
725
+ "eval_class_private_date_pred_spans": 33.0,
726
+ "eval_class_private_email_precision": 0.975,
727
+ "eval_class_private_email_recall": 1.0,
728
+ "eval_class_private_email_f1": 0.9873417721518987,
729
+ "eval_class_private_email_gold_spans": 39.0,
730
+ "eval_class_private_email_pred_spans": 40.0,
731
+ "eval_class_private_person_precision": 0.5445544554455446,
732
+ "eval_class_private_person_recall": 0.6748466257668712,
733
+ "eval_class_private_person_f1": 0.6027397260273972,
734
+ "eval_class_private_person_gold_spans": 163.0,
735
+ "eval_class_private_person_pred_spans": 202.0,
736
+ "eval_class_private_phone_precision": 1.0,
737
+ "eval_class_private_phone_recall": 1.0,
738
+ "eval_class_private_phone_f1": 1.0,
739
+ "eval_class_private_phone_gold_spans": 69.0,
740
+ "eval_class_private_phone_pred_spans": 69.0,
741
+ "eval_class_private_url_precision": 0.9565217391304348,
742
+ "eval_class_private_url_recall": 0.9565217391304348,
743
+ "eval_class_private_url_f1": 0.9565217391304348,
744
+ "eval_class_private_url_gold_spans": 23.0,
745
+ "eval_class_private_url_pred_spans": 23.0,
746
+ "eval_runtime": 7.1537,
747
+ "eval_samples_per_second": 311.307,
748
+ "eval_steps_per_second": 2.516,
749
+ "epoch": 5.0,
750
+ "step": 1610
751
+ },
752
+ {
753
+ "loss": 0.02931797981262207,
754
+ "grad_norm": 0.8306529521942139,
755
+ "learning_rate": 0.00028967063661406285,
756
+ "epoch": 5.046583850931677,
757
+ "step": 1625
758
+ },
759
+ {
760
+ "loss": 0.02403106212615967,
761
+ "grad_norm": 0.3057140111923218,
762
+ "learning_rate": 0.00028296738711533655,
763
+ "epoch": 5.124223602484472,
764
+ "step": 1650
765
+ },
766
+ {
767
+ "loss": 0.024544227123260497,
768
+ "grad_norm": 0.6968570351600647,
769
+ "learning_rate": 0.0002762399250434392,
770
+ "epoch": 5.201863354037267,
771
+ "step": 1675
772
+ },
773
+ {
774
+ "loss": 0.021933317184448242,
775
+ "grad_norm": 0.8571075797080994,
776
+ "learning_rate": 0.0002694931913167124,
777
+ "epoch": 5.279503105590062,
778
+ "step": 1700
779
+ },
780
+ {
781
+ "loss": 0.031152501106262206,
782
+ "grad_norm": 0.5366156101226807,
783
+ "learning_rate": 0.00026273214100737516,
784
+ "epoch": 5.357142857142857,
785
+ "step": 1725
786
+ },
787
+ {
788
+ "loss": 0.02120798110961914,
789
+ "grad_norm": 0.5963289141654968,
790
+ "learning_rate": 0.0002559617397023193,
791
+ "epoch": 5.434782608695652,
792
+ "step": 1750
793
+ },
794
+ {
795
+ "loss": 0.023069217205047607,
796
+ "grad_norm": 0.36561551690101624,
797
+ "learning_rate": 0.00024918695985618394,
798
+ "epoch": 5.512422360248447,
799
+ "step": 1775
800
+ },
801
+ {
802
+ "loss": 0.018797587156295776,
803
+ "grad_norm": 0.834469199180603,
804
+ "learning_rate": 0.00024241277713938443,
805
+ "epoch": 5.590062111801243,
806
+ "step": 1800
807
+ },
808
+ {
809
+ "loss": 0.027416987419128416,
810
+ "grad_norm": 0.3487168848514557,
811
+ "learning_rate": 0.0002356441667837804,
812
+ "epoch": 5.667701863354037,
813
+ "step": 1825
814
+ },
815
+ {
816
+ "loss": 0.029009580612182617,
817
+ "grad_norm": 0.49782446026802063,
818
+ "learning_rate": 0.0002288860999286656,
819
+ "epoch": 5.745341614906832,
820
+ "step": 1850
821
+ },
822
+ {
823
+ "loss": 0.02573819875717163,
824
+ "grad_norm": 0.5541417598724365,
825
+ "learning_rate": 0.00022214353996976349,
826
+ "epoch": 5.8229813664596275,
827
+ "step": 1875
828
+ },
829
+ {
830
+ "loss": 0.024653725624084473,
831
+ "grad_norm": 0.9394005537033081,
832
+ "learning_rate": 0.00021542143891391025,
833
+ "epoch": 5.900621118012422,
834
+ "step": 1900
835
+ },
836
+ {
837
+ "loss": 0.028091182708740236,
838
+ "grad_norm": 0.9744410514831543,
839
+ "learning_rate": 0.0002087247337421019,
840
+ "epoch": 5.978260869565218,
841
+ "step": 1925
842
+ },
843
+ {
844
+ "eval_loss": 0.07920802384614944,
845
+ "eval_token_accuracy": 0.9906921101774043,
846
+ "eval_span_precision": 0.8631790744466801,
847
+ "eval_span_recall": 0.825,
848
+ "eval_span_f1": 0.8436578171091444,
849
+ "eval_gold_spans": 520.0,
850
+ "eval_pred_spans": 497.0,
851
+ "eval_class_account_number_precision": 0.956140350877193,
852
+ "eval_class_account_number_recall": 0.9646017699115044,
853
+ "eval_class_account_number_f1": 0.9603524229074888,
854
+ "eval_class_account_number_gold_spans": 113.0,
855
+ "eval_class_account_number_pred_spans": 114.0,
856
+ "eval_class_ip_address_precision": 1.0,
857
+ "eval_class_ip_address_recall": 1.0,
858
+ "eval_class_ip_address_f1": 1.0,
859
+ "eval_class_ip_address_gold_spans": 4.0,
860
+ "eval_class_ip_address_pred_spans": 4.0,
861
+ "eval_class_personal_handle_precision": 0.9259259259259259,
862
+ "eval_class_personal_handle_recall": 0.8928571428571429,
863
+ "eval_class_personal_handle_f1": 0.9090909090909091,
864
+ "eval_class_personal_handle_gold_spans": 28.0,
865
+ "eval_class_personal_handle_pred_spans": 27.0,
866
+ "eval_class_private_address_precision": 0.7209302325581395,
867
+ "eval_class_private_address_recall": 0.6458333333333334,
868
+ "eval_class_private_address_f1": 0.6813186813186812,
869
+ "eval_class_private_address_gold_spans": 48.0,
870
+ "eval_class_private_address_pred_spans": 43.0,
871
+ "eval_class_private_date_precision": 0.9393939393939394,
872
+ "eval_class_private_date_recall": 0.9393939393939394,
873
+ "eval_class_private_date_f1": 0.9393939393939394,
874
+ "eval_class_private_date_gold_spans": 33.0,
875
+ "eval_class_private_date_pred_spans": 33.0,
876
+ "eval_class_private_email_precision": 0.95,
877
+ "eval_class_private_email_recall": 0.9743589743589743,
878
+ "eval_class_private_email_f1": 0.9620253164556962,
879
+ "eval_class_private_email_gold_spans": 39.0,
880
+ "eval_class_private_email_pred_spans": 40.0,
881
+ "eval_class_private_person_precision": 0.6944444444444444,
882
+ "eval_class_private_person_recall": 0.6134969325153374,
883
+ "eval_class_private_person_f1": 0.6514657980456026,
884
+ "eval_class_private_person_gold_spans": 163.0,
885
+ "eval_class_private_person_pred_spans": 144.0,
886
+ "eval_class_private_phone_precision": 1.0,
887
+ "eval_class_private_phone_recall": 1.0,
888
+ "eval_class_private_phone_f1": 1.0,
889
+ "eval_class_private_phone_gold_spans": 69.0,
890
+ "eval_class_private_phone_pred_spans": 69.0,
891
+ "eval_class_private_url_precision": 0.9565217391304348,
892
+ "eval_class_private_url_recall": 0.9565217391304348,
893
+ "eval_class_private_url_f1": 0.9565217391304348,
894
+ "eval_class_private_url_gold_spans": 23.0,
895
+ "eval_class_private_url_pred_spans": 23.0,
896
+ "eval_runtime": 6.6716,
897
+ "eval_samples_per_second": 333.804,
898
+ "eval_steps_per_second": 2.698,
899
+ "epoch": 6.0,
900
+ "step": 1932
901
+ },
902
+ {
903
+ "loss": 0.01689502239227295,
904
+ "grad_norm": 0.5681946277618408,
905
+ "learning_rate": 0.00020205834278357742,
906
+ "epoch": 6.055900621118012,
907
+ "step": 1950
908
+ },
909
+ {
910
+ "loss": 0.015278645753860474,
911
+ "grad_norm": 0.3505108058452606,
912
+ "learning_rate": 0.00019542716210359964,
913
+ "epoch": 6.133540372670807,
914
+ "step": 1975
915
+ },
916
+ {
917
+ "loss": 0.015209052562713623,
918
+ "grad_norm": 0.3547140955924988,
919
+ "learning_rate": 0.00018883606190758903,
920
+ "epoch": 6.211180124223603,
921
+ "step": 2000
922
+ },
923
+ {
924
+ "loss": 0.01439091444015503,
925
+ "grad_norm": 0.3973138928413391,
926
+ "learning_rate": 0.00018228988296424876,
927
+ "epoch": 6.288819875776397,
928
+ "step": 2025
929
+ },
930
+ {
931
+ "loss": 0.014540103673934936,
932
+ "grad_norm": 0.5407560467720032,
933
+ "learning_rate": 0.00017579343305030985,
934
+ "epoch": 6.366459627329193,
935
+ "step": 2050
936
+ },
937
+ {
938
+ "loss": 0.017405078411102295,
939
+ "grad_norm": 0.3447468876838684,
940
+ "learning_rate": 0.00016935148341950694,
941
+ "epoch": 6.444099378881988,
942
+ "step": 2075
943
+ },
944
+ {
945
+ "loss": 0.015654172897338867,
946
+ "grad_norm": 0.30423322319984436,
947
+ "learning_rate": 0.0001629687652983776,
948
+ "epoch": 6.521739130434782,
949
+ "step": 2100
950
+ },
951
+ {
952
+ "loss": 0.017501153945922852,
953
+ "grad_norm": 0.47501835227012634,
954
+ "learning_rate": 0.0001566499664114601,
955
+ "epoch": 6.599378881987578,
956
+ "step": 2125
957
+ },
958
+ {
959
+ "loss": 0.014991183280944825,
960
+ "grad_norm": 0.5283868312835693,
961
+ "learning_rate": 0.00015039972753843963,
962
+ "epoch": 6.6770186335403725,
963
+ "step": 2150
964
+ },
965
+ {
966
+ "loss": 0.021359992027282716,
967
+ "grad_norm": 0.9889101982116699,
968
+ "learning_rate": 0.00014422263910577376,
969
+ "epoch": 6.754658385093168,
970
+ "step": 2175
971
+ },
972
+ {
973
+ "loss": 0.01291258454322815,
974
+ "grad_norm": 0.6929994225502014,
975
+ "learning_rate": 0.0001381232378152987,
976
+ "epoch": 6.832298136645963,
977
+ "step": 2200
978
+ },
979
+ {
980
+ "loss": 0.018135499954223634,
981
+ "grad_norm": 0.7863985896110535,
982
+ "learning_rate": 0.00013210600331229317,
983
+ "epoch": 6.909937888198757,
984
+ "step": 2225
985
+ },
986
+ {
987
+ "loss": 0.022398817539215087,
988
+ "grad_norm": 0.45059671998023987,
989
+ "learning_rate": 0.000126175354895447,
990
+ "epoch": 6.987577639751553,
991
+ "step": 2250
992
+ },
993
+ {
994
+ "eval_loss": 0.08392658829689026,
995
+ "eval_token_accuracy": 0.9905753968253969,
996
+ "eval_span_precision": 0.8574257425742574,
997
+ "eval_span_recall": 0.8326923076923077,
998
+ "eval_span_f1": 0.8448780487804879,
999
+ "eval_gold_spans": 520.0,
1000
+ "eval_pred_spans": 505.0,
1001
+ "eval_class_account_number_precision": 0.9736842105263158,
1002
+ "eval_class_account_number_recall": 0.9823008849557522,
1003
+ "eval_class_account_number_f1": 0.9779735682819383,
1004
+ "eval_class_account_number_gold_spans": 113.0,
1005
+ "eval_class_account_number_pred_spans": 114.0,
1006
+ "eval_class_ip_address_precision": 1.0,
1007
+ "eval_class_ip_address_recall": 1.0,
1008
+ "eval_class_ip_address_f1": 1.0,
1009
+ "eval_class_ip_address_gold_spans": 4.0,
1010
+ "eval_class_ip_address_pred_spans": 4.0,
1011
+ "eval_class_personal_handle_precision": 0.9259259259259259,
1012
+ "eval_class_personal_handle_recall": 0.8928571428571429,
1013
+ "eval_class_personal_handle_f1": 0.9090909090909091,
1014
+ "eval_class_personal_handle_gold_spans": 28.0,
1015
+ "eval_class_personal_handle_pred_spans": 27.0,
1016
+ "eval_class_private_address_precision": 0.7857142857142857,
1017
+ "eval_class_private_address_recall": 0.6875,
1018
+ "eval_class_private_address_f1": 0.7333333333333334,
1019
+ "eval_class_private_address_gold_spans": 48.0,
1020
+ "eval_class_private_address_pred_spans": 42.0,
1021
+ "eval_class_private_date_precision": 1.0,
1022
+ "eval_class_private_date_recall": 1.0,
1023
+ "eval_class_private_date_f1": 1.0,
1024
+ "eval_class_private_date_gold_spans": 33.0,
1025
+ "eval_class_private_date_pred_spans": 33.0,
1026
+ "eval_class_private_email_precision": 0.926829268292683,
1027
+ "eval_class_private_email_recall": 0.9743589743589743,
1028
+ "eval_class_private_email_f1": 0.9500000000000001,
1029
+ "eval_class_private_email_gold_spans": 39.0,
1030
+ "eval_class_private_email_pred_spans": 41.0,
1031
+ "eval_class_private_person_precision": 0.6466666666666666,
1032
+ "eval_class_private_person_recall": 0.5950920245398773,
1033
+ "eval_class_private_person_f1": 0.6198083067092651,
1034
+ "eval_class_private_person_gold_spans": 163.0,
1035
+ "eval_class_private_person_pred_spans": 150.0,
1036
+ "eval_class_private_phone_precision": 1.0,
1037
+ "eval_class_private_phone_recall": 1.0,
1038
+ "eval_class_private_phone_f1": 1.0,
1039
+ "eval_class_private_phone_gold_spans": 69.0,
1040
+ "eval_class_private_phone_pred_spans": 69.0,
1041
+ "eval_class_private_url_precision": 0.92,
1042
+ "eval_class_private_url_recall": 1.0,
1043
+ "eval_class_private_url_f1": 0.9583333333333334,
1044
+ "eval_class_private_url_gold_spans": 23.0,
1045
+ "eval_class_private_url_pred_spans": 25.0,
1046
+ "eval_runtime": 6.6789,
1047
+ "eval_samples_per_second": 333.44,
1048
+ "eval_steps_per_second": 2.695,
1049
+ "epoch": 7.0,
1050
+ "step": 2254
1051
+ },
1052
+ {
1053
+ "loss": 0.013271838426589966,
1054
+ "grad_norm": 0.24804139137268066,
1055
+ "learning_rate": 0.00012033564827115123,
1056
+ "epoch": 7.065217391304348,
1057
+ "step": 2275
1058
+ },
1059
+ {
1060
+ "loss": 0.009375990629196166,
1061
+ "grad_norm": 0.34166398644447327,
1062
+ "learning_rate": 0.0001145911723544914,
1063
+ "epoch": 7.142857142857143,
1064
+ "step": 2300
1065
+ },
1066
+ {
1067
+ "loss": 0.014302024841308594,
1068
+ "grad_norm": 0.8205583691596985,
1069
+ "learning_rate": 0.00010894614611929701,
1070
+ "epoch": 7.220496894409938,
1071
+ "step": 2325
1072
+ },
1073
+ {
1074
+ "loss": 0.013667027950286865,
1075
+ "grad_norm": 0.10090164840221405,
1076
+ "learning_rate": 0.00010340471549955708,
1077
+ "epoch": 7.298136645962733,
1078
+ "step": 2350
1079
+ },
1080
+ {
1081
+ "loss": 0.008717143535614013,
1082
+ "grad_norm": 0.38398560881614685,
1083
+ "learning_rate": 9.797095034447984e-05,
1084
+ "epoch": 7.375776397515528,
1085
+ "step": 2375
1086
+ },
1087
+ {
1088
+ "loss": 0.009974555373191833,
1089
+ "grad_norm": 0.4515543580055237,
1090
+ "learning_rate": 9.264884142943237e-05,
1091
+ "epoch": 7.453416149068323,
1092
+ "step": 2400
1093
+ },
1094
+ {
1095
+ "loss": 0.011132798194885253,
1096
+ "grad_norm": 0.18586605787277222,
1097
+ "learning_rate": 8.744229752495425e-05,
1098
+ "epoch": 7.5310559006211175,
1099
+ "step": 2425
1100
+ },
1101
+ {
1102
+ "loss": 0.013178889751434325,
1103
+ "grad_norm": 0.8527100682258606,
1104
+ "learning_rate": 8.235514252600023e-05,
1105
+ "epoch": 7.608695652173913,
1106
+ "step": 2450
1107
+ },
1108
+ {
1109
+ "loss": 0.008569929599761963,
1110
+ "grad_norm": 0.08932485431432724,
1111
+ "learning_rate": 7.739111264351752e-05,
1112
+ "epoch": 7.686335403726708,
1113
+ "step": 2475
1114
+ },
1115
+ {
1116
+ "loss": 0.01057699203491211,
1117
+ "grad_norm": 0.7609089612960815,
1118
+ "learning_rate": 7.255385366042274e-05,
1119
+ "epoch": 7.763975155279503,
1120
+ "step": 2500
1121
+ },
1122
+ {
1123
+ "loss": 0.007927331924438476,
1124
+ "grad_norm": 0.30278390645980835,
1125
+ "learning_rate": 6.784691825399267e-05,
1126
+ "epoch": 7.841614906832298,
1127
+ "step": 2525
1128
+ },
1129
+ {
1130
+ "loss": 0.012843213081359862,
1131
+ "grad_norm": 0.3382692039012909,
1132
+ "learning_rate": 6.327376338663493e-05,
1133
+ "epoch": 7.919254658385093,
1134
+ "step": 2550
1135
+ },
1136
+ {
1137
+ "loss": 0.010109683275222778,
1138
+ "grad_norm": 1.0424972772598267,
1139
+ "learning_rate": 5.883774776695652e-05,
1140
+ "epoch": 7.996894409937888,
1141
+ "step": 2575
1142
+ },
1143
+ {
1144
+ "eval_loss": 0.09201034158468246,
1145
+ "eval_token_accuracy": 0.9901960784313726,
1146
+ "eval_span_precision": 0.8404669260700389,
1147
+ "eval_span_recall": 0.8307692307692308,
1148
+ "eval_span_f1": 0.8355899419729208,
1149
+ "eval_gold_spans": 520.0,
1150
+ "eval_pred_spans": 514.0,
1151
+ "eval_class_account_number_precision": 0.9736842105263158,
1152
+ "eval_class_account_number_recall": 0.9823008849557522,
1153
+ "eval_class_account_number_f1": 0.9779735682819383,
1154
+ "eval_class_account_number_gold_spans": 113.0,
1155
+ "eval_class_account_number_pred_spans": 114.0,
1156
+ "eval_class_ip_address_precision": 1.0,
1157
+ "eval_class_ip_address_recall": 1.0,
1158
+ "eval_class_ip_address_f1": 1.0,
1159
+ "eval_class_ip_address_gold_spans": 4.0,
1160
+ "eval_class_ip_address_pred_spans": 4.0,
1161
+ "eval_class_personal_handle_precision": 0.8214285714285714,
1162
+ "eval_class_personal_handle_recall": 0.8214285714285714,
1163
+ "eval_class_personal_handle_f1": 0.8214285714285714,
1164
+ "eval_class_personal_handle_gold_spans": 28.0,
1165
+ "eval_class_personal_handle_pred_spans": 28.0,
1166
+ "eval_class_private_address_precision": 0.7380952380952381,
1167
+ "eval_class_private_address_recall": 0.6458333333333334,
1168
+ "eval_class_private_address_f1": 0.6888888888888889,
1169
+ "eval_class_private_address_gold_spans": 48.0,
1170
+ "eval_class_private_address_pred_spans": 42.0,
1171
+ "eval_class_private_date_precision": 1.0,
1172
+ "eval_class_private_date_recall": 1.0,
1173
+ "eval_class_private_date_f1": 1.0,
1174
+ "eval_class_private_date_gold_spans": 33.0,
1175
+ "eval_class_private_date_pred_spans": 33.0,
1176
+ "eval_class_private_email_precision": 0.926829268292683,
1177
+ "eval_class_private_email_recall": 0.9743589743589743,
1178
+ "eval_class_private_email_f1": 0.9500000000000001,
1179
+ "eval_class_private_email_gold_spans": 39.0,
1180
+ "eval_class_private_email_pred_spans": 41.0,
1181
+ "eval_class_private_person_precision": 0.6329113924050633,
1182
+ "eval_class_private_person_recall": 0.6134969325153374,
1183
+ "eval_class_private_person_f1": 0.6230529595015576,
1184
+ "eval_class_private_person_gold_spans": 163.0,
1185
+ "eval_class_private_person_pred_spans": 158.0,
1186
+ "eval_class_private_phone_precision": 1.0,
1187
+ "eval_class_private_phone_recall": 1.0,
1188
+ "eval_class_private_phone_f1": 1.0,
1189
+ "eval_class_private_phone_gold_spans": 69.0,
1190
+ "eval_class_private_phone_pred_spans": 69.0,
1191
+ "eval_class_private_url_precision": 0.92,
1192
+ "eval_class_private_url_recall": 1.0,
1193
+ "eval_class_private_url_f1": 0.9583333333333334,
1194
+ "eval_class_private_url_gold_spans": 23.0,
1195
+ "eval_class_private_url_pred_spans": 25.0,
1196
+ "eval_runtime": 6.9909,
1197
+ "eval_samples_per_second": 318.559,
1198
+ "eval_steps_per_second": 2.575,
1199
+ "epoch": 8.0,
1200
+ "step": 2576
1201
+ },
1202
+ {
1203
+ "loss": 0.009849590063095093,
1204
+ "grad_norm": 0.1411687433719635,
1205
+ "learning_rate": 5.454212938299255e-05,
1206
+ "epoch": 8.074534161490684,
1207
+ "step": 2600
1208
+ },
1209
+ {
1210
+ "loss": 0.007954022884368896,
1211
+ "grad_norm": 0.3560449182987213,
1212
+ "learning_rate": 5.0390063109409914e-05,
1213
+ "epoch": 8.152173913043478,
1214
+ "step": 2625
1215
+ },
1216
+ {
1217
+ "loss": 0.005342819690704346,
1218
+ "grad_norm": 0.36226364970207214,
1219
+ "learning_rate": 4.6384598390439706e-05,
1220
+ "epoch": 8.229813664596273,
1221
+ "step": 2650
1222
+ },
1223
+ {
1224
+ "loss": 0.00969659686088562,
1225
+ "grad_norm": 0.5297871232032776,
1226
+ "learning_rate": 4.252867700024374e-05,
1227
+ "epoch": 8.307453416149068,
1228
+ "step": 2675
1229
+ },
1230
+ {
1231
+ "loss": 0.006519293785095215,
1232
+ "grad_norm": 0.5881589651107788,
1233
+ "learning_rate": 3.882513088235676e-05,
1234
+ "epoch": 8.385093167701863,
1235
+ "step": 2700
1236
+ },
1237
+ {
1238
+ "loss": 0.00985595166683197,
1239
+ "grad_norm": 0.50627201795578,
1240
+ "learning_rate": 3.5276680069793705e-05,
1241
+ "epoch": 8.462732919254659,
1242
+ "step": 2725
1243
+ },
1244
+ {
1245
+ "loss": 0.004720044732093811,
1246
+ "grad_norm": 0.17046724259853363,
1247
+ "learning_rate": 3.18859306873483e-05,
1248
+ "epoch": 8.540372670807454,
1249
+ "step": 2750
1250
+ },
1251
+ {
1252
+ "loss": 0.011722609996795655,
1253
+ "grad_norm": 0.294168084859848,
1254
+ "learning_rate": 2.8655373037550058e-05,
1255
+ "epoch": 8.618012422360248,
1256
+ "step": 2775
1257
+ },
1258
+ {
1259
+ "loss": 0.006650465726852417,
1260
+ "grad_norm": 0.7373964190483093,
1261
+ "learning_rate": 2.5587379771686403e-05,
1262
+ "epoch": 8.695652173913043,
1263
+ "step": 2800
1264
+ },
1265
+ {
1266
+ "loss": 0.01016199827194214,
1267
+ "grad_norm": 0.4546540081501007,
1268
+ "learning_rate": 2.2684204147231934e-05,
1269
+ "epoch": 8.773291925465838,
1270
+ "step": 2825
1271
+ },
1272
+ {
1273
+ "loss": 0.009012922048568725,
1274
+ "grad_norm": 0.9172267317771912,
1275
+ "learning_rate": 1.9947978372965804e-05,
1276
+ "epoch": 8.850931677018634,
1277
+ "step": 2850
1278
+ },
1279
+ {
1280
+ "loss": 0.005773799419403076,
1281
+ "grad_norm": 0.29897835850715637,
1282
+ "learning_rate": 1.7380712042992184e-05,
1283
+ "epoch": 8.928571428571429,
1284
+ "step": 2875
1285
+ },
1286
+ {
1287
+ "eval_loss": 0.09796658158302307,
1288
+ "eval_token_accuracy": 0.9907504668534081,
1289
+ "eval_span_precision": 0.8565815324165029,
1290
+ "eval_span_recall": 0.8384615384615385,
1291
+ "eval_span_f1": 0.847424684159378,
1292
+ "eval_gold_spans": 520.0,
1293
+ "eval_pred_spans": 509.0,
1294
+ "eval_class_account_number_precision": 0.9736842105263158,
1295
+ "eval_class_account_number_recall": 0.9823008849557522,
1296
+ "eval_class_account_number_f1": 0.9779735682819383,
1297
+ "eval_class_account_number_gold_spans": 113.0,
1298
+ "eval_class_account_number_pred_spans": 114.0,
1299
+ "eval_class_ip_address_precision": 1.0,
1300
+ "eval_class_ip_address_recall": 1.0,
1301
+ "eval_class_ip_address_f1": 1.0,
1302
+ "eval_class_ip_address_gold_spans": 4.0,
1303
+ "eval_class_ip_address_pred_spans": 4.0,
1304
+ "eval_class_personal_handle_precision": 0.8928571428571429,
1305
+ "eval_class_personal_handle_recall": 0.8928571428571429,
1306
+ "eval_class_personal_handle_f1": 0.8928571428571429,
1307
+ "eval_class_personal_handle_gold_spans": 28.0,
1308
+ "eval_class_personal_handle_pred_spans": 28.0,
1309
+ "eval_class_private_address_precision": 0.7441860465116279,
1310
+ "eval_class_private_address_recall": 0.6666666666666666,
1311
+ "eval_class_private_address_f1": 0.7032967032967034,
1312
+ "eval_class_private_address_gold_spans": 48.0,
1313
+ "eval_class_private_address_pred_spans": 43.0,
1314
+ "eval_class_private_date_precision": 1.0,
1315
+ "eval_class_private_date_recall": 1.0,
1316
+ "eval_class_private_date_f1": 1.0,
1317
+ "eval_class_private_date_gold_spans": 33.0,
1318
+ "eval_class_private_date_pred_spans": 33.0,
1319
+ "eval_class_private_email_precision": 0.926829268292683,
1320
+ "eval_class_private_email_recall": 0.9743589743589743,
1321
+ "eval_class_private_email_f1": 0.9500000000000001,
1322
+ "eval_class_private_email_gold_spans": 39.0,
1323
+ "eval_class_private_email_pred_spans": 41.0,
1324
+ "eval_class_private_person_precision": 0.6644736842105263,
1325
+ "eval_class_private_person_recall": 0.6196319018404908,
1326
+ "eval_class_private_person_f1": 0.6412698412698412,
1327
+ "eval_class_private_person_gold_spans": 163.0,
1328
+ "eval_class_private_person_pred_spans": 152.0,
1329
+ "eval_class_private_phone_precision": 1.0,
1330
+ "eval_class_private_phone_recall": 1.0,
1331
+ "eval_class_private_phone_f1": 1.0,
1332
+ "eval_class_private_phone_gold_spans": 69.0,
1333
+ "eval_class_private_phone_pred_spans": 69.0,
1334
+ "eval_class_private_url_precision": 0.92,
1335
+ "eval_class_private_url_recall": 1.0,
1336
+ "eval_class_private_url_f1": 0.9583333333333334,
1337
+ "eval_class_private_url_gold_spans": 23.0,
1338
+ "eval_class_private_url_pred_spans": 25.0,
1339
+ "eval_runtime": 7.2699,
1340
+ "eval_samples_per_second": 306.33,
1341
+ "eval_steps_per_second": 2.476,
1342
+ "epoch": 9.0,
1343
+ "step": 2898
1344
+ },
1345
+ {
1346
+ "loss": 0.008111558556556701,
1347
+ "grad_norm": 0.1876034289598465,
1348
+ "learning_rate": 1.498429066081311e-05,
1349
+ "epoch": 9.006211180124224,
1350
+ "step": 2900
1351
+ },
1352
+ {
1353
+ "loss": 0.007400209903717041,
1354
+ "grad_norm": 0.23115915060043335,
1355
+ "learning_rate": 1.276047425453955e-05,
1356
+ "epoch": 9.083850931677018,
1357
+ "step": 2925
1358
+ },
1359
+ {
1360
+ "loss": 0.007166728377342224,
1361
+ "grad_norm": 0.30637550354003906,
1362
+ "learning_rate": 1.0710896084255684e-05,
1363
+ "epoch": 9.161490683229813,
1364
+ "step": 2950
1365
+ },
1366
+ {
1367
+ "loss": 0.006242750883102417,
1368
+ "grad_norm": 0.2221984565258026,
1369
+ "learning_rate": 8.837061442487193e-06,
1370
+ "epoch": 9.23913043478261,
1371
+ "step": 2975
1372
+ },
1373
+ {
1374
+ "loss": 0.008013453483581543,
1375
+ "grad_norm": 0.38621774315834045,
1376
+ "learning_rate": 7.1403465486540584e-06,
1377
+ "epoch": 9.316770186335404,
1378
+ "step": 3000
1379
+ },
1380
+ {
1381
+ "loss": 0.008517981171607972,
1382
+ "grad_norm": 0.2457902729511261,
1383
+ "learning_rate": 5.6219975383197675e-06,
1384
+ "epoch": 9.394409937888199,
1385
+ "step": 3025
1386
+ },
1387
+ {
1388
+ "loss": 0.006109429001808167,
1389
+ "grad_norm": 0.5641500353813171,
1390
+ "learning_rate": 4.2831295479791034e-06,
1391
+ "epoch": 9.472049689440993,
1392
+ "step": 3050
1393
+ },
1394
+ {
1395
+ "loss": 0.005958260297775269,
1396
+ "grad_norm": 0.18688508868217468,
1397
+ "learning_rate": 3.1247258960573067e-06,
1398
+ "epoch": 9.549689440993788,
1399
+ "step": 3075
1400
+ },
1401
+ {
1402
+ "loss": 0.009225964546203613,
1403
+ "grad_norm": 0.2731750011444092,
1404
+ "learning_rate": 2.1476373607214882e-06,
1405
+ "epoch": 9.627329192546584,
1406
+ "step": 3100
1407
+ },
1408
+ {
1409
+ "loss": 0.0062656128406524654,
1410
+ "grad_norm": 0.12572576105594635,
1411
+ "learning_rate": 1.3525815550349262e-06,
1412
+ "epoch": 9.70496894409938,
1413
+ "step": 3125
1414
+ },
1415
+ {
1416
+ "loss": 0.0059302222728729245,
1417
+ "grad_norm": 0.28346896171569824,
1418
+ "learning_rate": 7.401423999132406e-07,
1419
+ "epoch": 9.782608695652174,
1420
+ "step": 3150
1421
+ },
1422
+ {
1423
+ "loss": 0.005810813903808594,
1424
+ "grad_norm": 0.4496349096298218,
1425
+ "learning_rate": 3.107696952694139e-07,
1426
+ "epoch": 9.860248447204969,
1427
+ "step": 3175
1428
+ },
1429
+ {
1430
+ "loss": 0.008135276436805726,
1431
+ "grad_norm": 0.7214242815971375,
1432
+ "learning_rate": 6.477878966271167e-08,
1433
+ "epoch": 9.937888198757763,
1434
+ "step": 3200
1435
+ },
1436
+ {
1437
+ "eval_loss": 0.09789847582578659,
1438
+ "eval_token_accuracy": 0.9908088235294118,
1439
+ "eval_span_precision": 0.8551859099804305,
1440
+ "eval_span_recall": 0.8403846153846154,
1441
+ "eval_span_f1": 0.8477206595538312,
1442
+ "eval_gold_spans": 520.0,
1443
+ "eval_pred_spans": 511.0,
1444
+ "eval_class_account_number_precision": 0.9736842105263158,
1445
+ "eval_class_account_number_recall": 0.9823008849557522,
1446
+ "eval_class_account_number_f1": 0.9779735682819383,
1447
+ "eval_class_account_number_gold_spans": 113.0,
1448
+ "eval_class_account_number_pred_spans": 114.0,
1449
+ "eval_class_ip_address_precision": 1.0,
1450
+ "eval_class_ip_address_recall": 1.0,
1451
+ "eval_class_ip_address_f1": 1.0,
1452
+ "eval_class_ip_address_gold_spans": 4.0,
1453
+ "eval_class_ip_address_pred_spans": 4.0,
1454
+ "eval_class_personal_handle_precision": 0.8928571428571429,
1455
+ "eval_class_personal_handle_recall": 0.8928571428571429,
1456
+ "eval_class_personal_handle_f1": 0.8928571428571429,
1457
+ "eval_class_personal_handle_gold_spans": 28.0,
1458
+ "eval_class_personal_handle_pred_spans": 28.0,
1459
+ "eval_class_private_address_precision": 0.7441860465116279,
1460
+ "eval_class_private_address_recall": 0.6666666666666666,
1461
+ "eval_class_private_address_f1": 0.7032967032967034,
1462
+ "eval_class_private_address_gold_spans": 48.0,
1463
+ "eval_class_private_address_pred_spans": 43.0,
1464
+ "eval_class_private_date_precision": 1.0,
1465
+ "eval_class_private_date_recall": 1.0,
1466
+ "eval_class_private_date_f1": 1.0,
1467
+ "eval_class_private_date_gold_spans": 33.0,
1468
+ "eval_class_private_date_pred_spans": 33.0,
1469
+ "eval_class_private_email_precision": 0.926829268292683,
1470
+ "eval_class_private_email_recall": 0.9743589743589743,
1471
+ "eval_class_private_email_f1": 0.9500000000000001,
1472
+ "eval_class_private_email_gold_spans": 39.0,
1473
+ "eval_class_private_email_pred_spans": 41.0,
1474
+ "eval_class_private_person_precision": 0.6623376623376623,
1475
+ "eval_class_private_person_recall": 0.6257668711656442,
1476
+ "eval_class_private_person_f1": 0.6435331230283912,
1477
+ "eval_class_private_person_gold_spans": 163.0,
1478
+ "eval_class_private_person_pred_spans": 154.0,
1479
+ "eval_class_private_phone_precision": 1.0,
1480
+ "eval_class_private_phone_recall": 1.0,
1481
+ "eval_class_private_phone_f1": 1.0,
1482
+ "eval_class_private_phone_gold_spans": 69.0,
1483
+ "eval_class_private_phone_pred_spans": 69.0,
1484
+ "eval_class_private_url_precision": 0.92,
1485
+ "eval_class_private_url_recall": 1.0,
1486
+ "eval_class_private_url_f1": 0.9583333333333334,
1487
+ "eval_class_private_url_gold_spans": 23.0,
1488
+ "eval_class_private_url_pred_spans": 25.0,
1489
+ "eval_runtime": 7.0696,
1490
+ "eval_samples_per_second": 315.012,
1491
+ "eval_steps_per_second": 2.546,
1492
+ "epoch": 10.0,
1493
+ "step": 3220
1494
+ },
1495
+ {
1496
+ "train_runtime": 2753.5271,
1497
+ "train_samples_per_second": 149.615,
1498
+ "train_steps_per_second": 1.169,
1499
+ "total_flos": 1.291714856788951e+17,
1500
+ "train_loss": 0.07032274794504509,
1501
+ "epoch": 10.0,
1502
+ "step": 3220
1503
+ },
1504
+ {
1505
+ "eval_loss": 0.09874702990055084,
1506
+ "eval_token_accuracy": 0.9907504668534081,
1507
+ "eval_span_precision": 0.8582677165354331,
1508
+ "eval_span_recall": 0.8384615384615385,
1509
+ "eval_span_f1": 0.8482490272373542,
1510
+ "eval_gold_spans": 520.0,
1511
+ "eval_pred_spans": 508.0,
1512
+ "eval_class_account_number_precision": 0.9736842105263158,
1513
+ "eval_class_account_number_recall": 0.9823008849557522,
1514
+ "eval_class_account_number_f1": 0.9779735682819383,
1515
+ "eval_class_account_number_gold_spans": 113.0,
1516
+ "eval_class_account_number_pred_spans": 114.0,
1517
+ "eval_class_ip_address_precision": 1.0,
1518
+ "eval_class_ip_address_recall": 1.0,
1519
+ "eval_class_ip_address_f1": 1.0,
1520
+ "eval_class_ip_address_gold_spans": 4.0,
1521
+ "eval_class_ip_address_pred_spans": 4.0,
1522
+ "eval_class_personal_handle_precision": 0.8571428571428571,
1523
+ "eval_class_personal_handle_recall": 0.8571428571428571,
1524
+ "eval_class_personal_handle_f1": 0.8571428571428571,
1525
+ "eval_class_personal_handle_gold_spans": 28.0,
1526
+ "eval_class_personal_handle_pred_spans": 28.0,
1527
+ "eval_class_private_address_precision": 0.7619047619047619,
1528
+ "eval_class_private_address_recall": 0.6666666666666666,
1529
+ "eval_class_private_address_f1": 0.7111111111111111,
1530
+ "eval_class_private_address_gold_spans": 48.0,
1531
+ "eval_class_private_address_pred_spans": 42.0,
1532
+ "eval_class_private_date_precision": 1.0,
1533
+ "eval_class_private_date_recall": 1.0,
1534
+ "eval_class_private_date_f1": 1.0,
1535
+ "eval_class_private_date_gold_spans": 33.0,
1536
+ "eval_class_private_date_pred_spans": 33.0,
1537
+ "eval_class_private_email_precision": 0.926829268292683,
1538
+ "eval_class_private_email_recall": 0.9743589743589743,
1539
+ "eval_class_private_email_f1": 0.9500000000000001,
1540
+ "eval_class_private_email_gold_spans": 39.0,
1541
+ "eval_class_private_email_pred_spans": 41.0,
1542
+ "eval_class_private_person_precision": 0.6710526315789473,
1543
+ "eval_class_private_person_recall": 0.6257668711656442,
1544
+ "eval_class_private_person_f1": 0.6476190476190476,
1545
+ "eval_class_private_person_gold_spans": 163.0,
1546
+ "eval_class_private_person_pred_spans": 152.0,
1547
+ "eval_class_private_phone_precision": 1.0,
1548
+ "eval_class_private_phone_recall": 1.0,
1549
+ "eval_class_private_phone_f1": 1.0,
1550
+ "eval_class_private_phone_gold_spans": 69.0,
1551
+ "eval_class_private_phone_pred_spans": 69.0,
1552
+ "eval_class_private_url_precision": 0.92,
1553
+ "eval_class_private_url_recall": 1.0,
1554
+ "eval_class_private_url_f1": 0.9583333333333334,
1555
+ "eval_class_private_url_gold_spans": 23.0,
1556
+ "eval_class_private_url_pred_spans": 25.0,
1557
+ "eval_runtime": 7.0151,
1558
+ "eval_samples_per_second": 317.46,
1559
+ "eval_steps_per_second": 2.566,
1560
+ "epoch": 10.0,
1561
+ "step": 3220
1562
+ },
1563
+ {
1564
+ "test_loss": 0.08586616814136505,
1565
+ "test_token_accuracy": 0.9924174456631841,
1566
+ "test_span_precision": 0.9009708737864077,
1567
+ "test_span_recall": 0.8560885608856088,
1568
+ "test_span_f1": 0.8779564806054873,
1569
+ "test_gold_spans": 542.0,
1570
+ "test_pred_spans": 515.0,
1571
+ "test_class_account_number_precision": 0.9752066115702479,
1572
+ "test_class_account_number_recall": 0.9833333333333333,
1573
+ "test_class_account_number_f1": 0.979253112033195,
1574
+ "test_class_account_number_gold_spans": 120.0,
1575
+ "test_class_account_number_pred_spans": 121.0,
1576
+ "test_class_ip_address_precision": 1.0,
1577
+ "test_class_ip_address_recall": 1.0,
1578
+ "test_class_ip_address_f1": 1.0,
1579
+ "test_class_ip_address_gold_spans": 9.0,
1580
+ "test_class_ip_address_pred_spans": 9.0,
1581
+ "test_class_personal_handle_precision": 0.9743589743589743,
1582
+ "test_class_personal_handle_recall": 0.9743589743589743,
1583
+ "test_class_personal_handle_f1": 0.9743589743589743,
1584
+ "test_class_personal_handle_gold_spans": 39.0,
1585
+ "test_class_personal_handle_pred_spans": 39.0,
1586
+ "test_class_private_address_precision": 0.8275862068965517,
1587
+ "test_class_private_address_recall": 0.7384615384615385,
1588
+ "test_class_private_address_f1": 0.7804878048780489,
1589
+ "test_class_private_address_gold_spans": 65.0,
1590
+ "test_class_private_address_pred_spans": 58.0,
1591
+ "test_class_private_date_precision": 0.9166666666666666,
1592
+ "test_class_private_date_recall": 0.88,
1593
+ "test_class_private_date_f1": 0.8979591836734694,
1594
+ "test_class_private_date_gold_spans": 25.0,
1595
+ "test_class_private_date_pred_spans": 24.0,
1596
+ "test_class_private_email_precision": 1.0,
1597
+ "test_class_private_email_recall": 1.0,
1598
+ "test_class_private_email_f1": 1.0,
1599
+ "test_class_private_email_gold_spans": 38.0,
1600
+ "test_class_private_email_pred_spans": 38.0,
1601
+ "test_class_private_person_precision": 0.7348484848484849,
1602
+ "test_class_private_person_recall": 0.6381578947368421,
1603
+ "test_class_private_person_f1": 0.6830985915492959,
1604
+ "test_class_private_person_gold_spans": 152.0,
1605
+ "test_class_private_person_pred_spans": 132.0,
1606
+ "test_class_private_phone_precision": 1.0,
1607
+ "test_class_private_phone_recall": 1.0,
1608
+ "test_class_private_phone_f1": 1.0,
1609
+ "test_class_private_phone_gold_spans": 76.0,
1610
+ "test_class_private_phone_pred_spans": 76.0,
1611
+ "test_class_private_url_precision": 1.0,
1612
+ "test_class_private_url_recall": 1.0,
1613
+ "test_class_private_url_f1": 1.0,
1614
+ "test_class_private_url_gold_spans": 18.0,
1615
+ "test_class_private_url_pred_spans": 18.0,
1616
+ "test_runtime": 6.4275,
1617
+ "test_samples_per_second": 350.37,
1618
+ "test_steps_per_second": 2.8,
1619
+ "epoch": 10.0,
1620
+ "step": 3220
1621
+ }
1622
+ ]
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36d54436cb0f35aecb8cfba81f9eb72967d7a3950cff8140be9a6a9f2bbba92d
3
+ size 2798994626
test_privacy_filter_ko.ipynb ADDED
@@ -0,0 +1,299 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cells": [
3
+ {
4
+ "cell_type": "markdown",
5
+ "metadata": {},
6
+ "source": "# Privacy Filter — Korean — Test Notebook\n\nTest the LoRA-fine-tuned `FrameByFrame/privacy-filter-korean` model on Korean and English PII detection.\n\n**Capabilities (9 categories):**\n- `private_person` — personal names (Korean / Western / handles)\n- `private_address` — physical / postal addresses\n- `private_phone` — phone numbers\n- `private_email` — email addresses\n- `private_date` — birthdays / personally-identifying dates\n- `private_url` — personal URLs\n- `account_number` — bank, card, RRN, passport, etc.\n- `personal_handle` — usernames / handles\n- `ip_address` — IP addresses"
7
+ },
8
+ {
9
+ "cell_type": "markdown",
10
+ "metadata": {},
11
+ "source": [
12
+ "## 1. Install & Load Model"
13
+ ]
14
+ },
15
+ {
16
+ "cell_type": "code",
17
+ "execution_count": null,
18
+ "metadata": {},
19
+ "outputs": [],
20
+ "source": [
21
+ "# Uncomment if needed\n",
22
+ "# !pip install transformers peft torch"
23
+ ]
24
+ },
25
+ {
26
+ "cell_type": "code",
27
+ "execution_count": null,
28
+ "metadata": {},
29
+ "outputs": [],
30
+ "source": "from transformers import AutoTokenizer, AutoModelForTokenClassification\nimport torch\nimport time\n\nMODEL_ID = \"FrameByFrame/privacy-filter-korean\"\n\ntokenizer = AutoTokenizer.from_pretrained(MODEL_ID, trust_remote_code=True)\nmodel = AutoModelForTokenClassification.from_pretrained(\n MODEL_ID, trust_remote_code=True, torch_dtype=torch.bfloat16\n)\nmodel.eval()\nif torch.cuda.is_available():\n model.cuda()\n\nprint(f\"Model loaded. Categories: {sorted(set(v.split('-', 1)[-1] for v in model.config.id2label.values() if v != 'O'))}\")"
31
+ },
32
+ {
33
+ "cell_type": "markdown",
34
+ "metadata": {},
35
+ "source": [
36
+ "## 2. Helper — extract spans + show"
37
+ ]
38
+ },
39
+ {
40
+ "cell_type": "code",
41
+ "execution_count": null,
42
+ "metadata": {},
43
+ "outputs": [],
44
+ "source": [
45
+ "import json\n",
46
+ "\n",
47
+ "\n",
48
+ "def extract_pii(text: str, max_length: int = 512):\n",
49
+ " \"\"\"Run the model on `text` and decode BIOES into character-offset spans.\"\"\"\n",
50
+ " enc = tokenizer(\n",
51
+ " text,\n",
52
+ " truncation=True,\n",
53
+ " max_length=max_length,\n",
54
+ " return_offsets_mapping=True,\n",
55
+ " return_tensors=\"pt\",\n",
56
+ " )\n",
57
+ " offsets = enc.pop(\"offset_mapping\")[0].tolist()\n",
58
+ " enc = {k: v.to(model.device) for k, v in enc.items()}\n",
59
+ " with torch.no_grad():\n",
60
+ " logits = model(**enc).logits\n",
61
+ " pred_ids = logits.argmax(-1)[0].tolist()\n",
62
+ " id2label = model.config.id2label\n",
63
+ "\n",
64
+ " spans = []\n",
65
+ " active = None # (label, start, end)\n",
66
+ " for tok_idx, lid in enumerate(pred_ids):\n",
67
+ " label = id2label[int(lid)]\n",
68
+ " if label == \"O\":\n",
69
+ " if active is not None:\n",
70
+ " spans.append(active)\n",
71
+ " active = None\n",
72
+ " continue\n",
73
+ " prefix, cat = label.split(\"-\", 1)\n",
74
+ " c_start, c_end = offsets[tok_idx]\n",
75
+ " if prefix == \"S\":\n",
76
+ " if active is not None:\n",
77
+ " spans.append(active)\n",
78
+ " active = None\n",
79
+ " spans.append((cat, c_start, c_end))\n",
80
+ " elif prefix == \"B\":\n",
81
+ " if active is not None:\n",
82
+ " spans.append(active)\n",
83
+ " active = (cat, c_start, c_end)\n",
84
+ " elif prefix in (\"I\", \"E\"):\n",
85
+ " if active and active[0] == cat:\n",
86
+ " active = (active[0], active[1], c_end)\n",
87
+ " else:\n",
88
+ " if active is not None:\n",
89
+ " spans.append(active)\n",
90
+ " active = None\n",
91
+ " if prefix == \"E\":\n",
92
+ " spans.append((cat, c_start, c_end))\n",
93
+ " if active is not None:\n",
94
+ " spans.append(active)\n",
95
+ "\n",
96
+ " return [\n",
97
+ " {\"label\": cat, \"start\": s, \"end\": e, \"text\": text[s:e].strip()}\n",
98
+ " for cat, s, e in spans\n",
99
+ " if text[s:e].strip()\n",
100
+ " ]\n",
101
+ "\n",
102
+ "\n",
103
+ "def show(text: str):\n",
104
+ " \"\"\"Detect spans and pretty-print with timing.\"\"\"\n",
105
+ " t0 = time.time()\n",
106
+ " spans = extract_pii(text)\n",
107
+ " ms = round((time.time() - t0) * 1000)\n",
108
+ " icon = \"🚫\" if spans else \"✅\"\n",
109
+ " print(f\"{icon} [{ms}ms] {text[:100]}\")\n",
110
+ " if spans:\n",
111
+ " print(json.dumps(spans, indent=2, ensure_ascii=False))\n",
112
+ " else:\n",
113
+ " print(\" (no PII detected)\")\n",
114
+ " print()\n",
115
+ "\n",
116
+ "\n",
117
+ "def redact(text: str) -> str:\n",
118
+ " \"\"\"Replace each detected span with [LABEL] in reverse order so offsets stay valid.\"\"\"\n",
119
+ " spans = sorted(extract_pii(text), key=lambda s: s[\"start\"], reverse=True)\n",
120
+ " out = text\n",
121
+ " for s in spans:\n",
122
+ " out = out[: s[\"start\"]] + f\"[{s['label'].upper()}]\" + out[s[\"end\"]:]\n",
123
+ " return out"
124
+ ]
125
+ },
126
+ {
127
+ "cell_type": "markdown",
128
+ "metadata": {},
129
+ "source": [
130
+ "## 3. Korean — Chat-style PII"
131
+ ]
132
+ },
133
+ {
134
+ "cell_type": "code",
135
+ "execution_count": null,
136
+ "metadata": {},
137
+ "outputs": [],
138
+ "source": [
139
+ "show(\"김민수의 전화번호는 010-1234-5678이고 이메일은 minsu@example.com입니다.\")\n",
140
+ "show(\"서울특별시 강남구 테헤란로 123에 사는 박지영씨에게 연락주세요.\")\n",
141
+ "show(\"오늘 날씨가 좋네요.\") # safe — no PII\n",
142
+ "show(\"제 생일은 1990년 5월 14일입니다. 카드번호 1234-5678-9012-3456 잊지 마세요.\")"
143
+ ]
144
+ },
145
+ {
146
+ "cell_type": "markdown",
147
+ "metadata": {},
148
+ "source": [
149
+ "## 4. Korean — Form-style document\n",
150
+ "\n",
151
+ "This is the format-matched style (`이름:`, `주소:` clues) — Privacy Filter handles it strongly because the base model was trained on similar structured PII data."
152
+ ]
153
+ },
154
+ {
155
+ "cell_type": "code",
156
+ "execution_count": null,
157
+ "metadata": {},
158
+ "outputs": [],
159
+ "source": [
160
+ "show(\"\"\"고객 정보\n",
161
+ "이름: 이수진\n",
162
+ "생년월일: 1985년 3월 12일\n",
163
+ "주소: 부산광역시 해운대구 우동 1457\n",
164
+ "연락처: 010-9876-5432\n",
165
+ "이메일: lee.sj@daum.net\"\"\")"
166
+ ]
167
+ },
168
+ {
169
+ "cell_type": "markdown",
170
+ "metadata": {},
171
+ "source": [
172
+ "## 5. Korean — Banking / multi-PII"
173
+ ]
174
+ },
175
+ {
176
+ "cell_type": "code",
177
+ "execution_count": null,
178
+ "metadata": {},
179
+ "outputs": [],
180
+ "source": [
181
+ "show(\"신한은행 계좌번호 110-234-567890 (예금주 박민수), 등록 주소 인천광역시 연수구 송도과학로 100, 비상연락 010-2345-6789, 가입일 2018.04.22.\")"
182
+ ]
183
+ },
184
+ {
185
+ "cell_type": "markdown",
186
+ "metadata": {},
187
+ "source": [
188
+ "## 6. English — names, addresses, accounts"
189
+ ]
190
+ },
191
+ {
192
+ "cell_type": "code",
193
+ "execution_count": null,
194
+ "metadata": {},
195
+ "outputs": [],
196
+ "source": [
197
+ "show(\"John Smith works at Google. Email: john@google.com, phone: 555-1234.\")\n",
198
+ "show(\"Wire to acct 110-234-567890, contact minsu@example.com\")\n",
199
+ "show(\"My SSN is 123-45-6789 and I live at 456 Oak Street, Springfield.\")\n",
200
+ "show(\"The weather is nice today.\") # safe"
201
+ ]
202
+ },
203
+ {
204
+ "cell_type": "markdown",
205
+ "metadata": {},
206
+ "source": [
207
+ "## 7. Redaction"
208
+ ]
209
+ },
210
+ {
211
+ "cell_type": "code",
212
+ "execution_count": null,
213
+ "metadata": {},
214
+ "outputs": [],
215
+ "source": [
216
+ "samples = [\n",
217
+ " \"김민수님의 번호는 010-1234-5678입니다.\",\n",
218
+ " \"서울특별시 강남구 테헤란로 123에 사는 박지영씨에게 연락주세요.\",\n",
219
+ " \"My account is 110-234-567890 and email is minsu@example.com.\",\n",
220
+ "]\n",
221
+ "for s in samples:\n",
222
+ " print(f\" in: {s}\")\n",
223
+ " print(f\" out: {redact(s)}\")\n",
224
+ " print()"
225
+ ]
226
+ },
227
+ {
228
+ "cell_type": "markdown",
229
+ "metadata": {},
230
+ "source": [
231
+ "## 8. Latency benchmark"
232
+ ]
233
+ },
234
+ {
235
+ "cell_type": "code",
236
+ "execution_count": null,
237
+ "metadata": {},
238
+ "outputs": [],
239
+ "source": [
240
+ "test_cases = [\n",
241
+ " (\"오늘 점심 뭐 먹지?\", 0),\n",
242
+ " (\"010-1234-5678로 전화해\", 1),\n",
243
+ " (\"What time is it?\", 0),\n",
244
+ " (\"주민등록번호 901201-1234567\", 1),\n",
245
+ " (\"김민수의 전화번호는 010-1234-5678이고 이메일은 minsu@example.com입니다.\", 3),\n",
246
+ " (\"서울특별시 강남구 테헤란로 123에 사는 박지영씨에게 연락주세요.\", 4),\n",
247
+ " (\"신한은행 계좌번호 110-234-567890 (예금주 박민수)\", 2),\n",
248
+ " (\"My account is 110-234-567890 and email is minsu@example.com.\", 2),\n",
249
+ "]\n",
250
+ "\n",
251
+ "total_ms = 0\n",
252
+ "correct_count = 0\n",
253
+ "for text, expected_n in test_cases:\n",
254
+ " t0 = time.time()\n",
255
+ " spans = extract_pii(text)\n",
256
+ " ms = round((time.time() - t0) * 1000)\n",
257
+ " total_ms += ms\n",
258
+ " n = len(spans)\n",
259
+ " icon = \"✅\" if n == expected_n else (\"~\" if abs(n - expected_n) <= 1 else \"❌\")\n",
260
+ " correct_count += int(n == expected_n)\n",
261
+ " print(f\"{icon} [{ms:>4}ms] expected={expected_n} got={n} | {text[:70]}\")\n",
262
+ "\n",
263
+ "print(f\"\\nExact-count match: {correct_count}/{len(test_cases)}\")\n",
264
+ "print(f\"Avg latency: {total_ms/len(test_cases):.0f}ms\")"
265
+ ]
266
+ },
267
+ {
268
+ "cell_type": "markdown",
269
+ "metadata": {},
270
+ "source": [
271
+ "## 9. Custom Test\n",
272
+ "\n",
273
+ "Try your own inputs:"
274
+ ]
275
+ },
276
+ {
277
+ "cell_type": "code",
278
+ "execution_count": null,
279
+ "metadata": {},
280
+ "outputs": [],
281
+ "source": [
282
+ "show(\"여기에 한국어 텍스트를 넣으세요\")"
283
+ ]
284
+ }
285
+ ],
286
+ "metadata": {
287
+ "kernelspec": {
288
+ "display_name": "Python 3",
289
+ "language": "python",
290
+ "name": "python3"
291
+ },
292
+ "language_info": {
293
+ "name": "python",
294
+ "version": "3.11"
295
+ }
296
+ },
297
+ "nbformat": 4,
298
+ "nbformat_minor": 5
299
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e714c627d94fd333b14f9ff32436219a4d7ac969719efe340fdc3385e1c7cd3e
3
+ size 27868272
tokenizer_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "eos_token": "<|endoftext|>",
4
+ "is_local": true,
5
+ "local_files_only": false,
6
+ "model_input_names": [
7
+ "input_ids",
8
+ "attention_mask"
9
+ ],
10
+ "model_max_length": 128000,
11
+ "pad_token": "<|endoftext|>",
12
+ "tokenizer_class": "TokenizersBackend"
13
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:195eed197af2b70c2fd47223db6e41715897a271a1bcbac65be2af77ec79752c
3
+ size 4920
training_summary.json ADDED
@@ -0,0 +1,258 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "checkpoint": "/models/privacy-filter",
3
+ "output_dir": "/workspace/data/checkpoints/ko_pii_hf_ddp_v6_lora",
4
+ "label_space_json": "/workspace/data/generated/ko_pii_opf_v4/label_space.json",
5
+ "token_labels": [
6
+ "O",
7
+ "B-private_person",
8
+ "I-private_person",
9
+ "E-private_person",
10
+ "S-private_person",
11
+ "B-personal_handle",
12
+ "I-personal_handle",
13
+ "E-personal_handle",
14
+ "S-personal_handle",
15
+ "B-private_phone",
16
+ "I-private_phone",
17
+ "E-private_phone",
18
+ "S-private_phone",
19
+ "B-private_email",
20
+ "I-private_email",
21
+ "E-private_email",
22
+ "S-private_email",
23
+ "B-private_address",
24
+ "I-private_address",
25
+ "E-private_address",
26
+ "S-private_address",
27
+ "B-private_date",
28
+ "I-private_date",
29
+ "E-private_date",
30
+ "S-private_date",
31
+ "B-private_url",
32
+ "I-private_url",
33
+ "E-private_url",
34
+ "S-private_url",
35
+ "B-account_number",
36
+ "I-account_number",
37
+ "E-account_number",
38
+ "S-account_number",
39
+ "B-ip_address",
40
+ "I-ip_address",
41
+ "E-ip_address",
42
+ "S-ip_address"
43
+ ],
44
+ "classifier_remap": {
45
+ "exact_rows_copied": 29,
46
+ "fallback_rows_copied": 8,
47
+ "random_rows_kept": 0
48
+ },
49
+ "lora": {
50
+ "r": 16,
51
+ "alpha": 32,
52
+ "dropout": 0.05,
53
+ "target_modules": [
54
+ "q_proj",
55
+ "k_proj",
56
+ "v_proj",
57
+ "o_proj"
58
+ ],
59
+ "trainable_params": 613541,
60
+ "total_params": 1400102970,
61
+ "trainable_pct": 0.0438
62
+ },
63
+ "train_dataset": {
64
+ "split": "train",
65
+ "records": 41197,
66
+ "tokens": 650201,
67
+ "spans": 11465,
68
+ "spans_without_token_overlap": 0,
69
+ "truncated_examples": 0,
70
+ "max_tokens": 63,
71
+ "records_per_path": {
72
+ "/workspace/data/generated/ko_pii_opf_v4/train.jsonl": 41197
73
+ }
74
+ },
75
+ "validation_dataset": {
76
+ "split": "validation",
77
+ "records": 2227,
78
+ "tokens": 34272,
79
+ "spans": 520,
80
+ "spans_without_token_overlap": 0,
81
+ "truncated_examples": 0,
82
+ "max_tokens": 45,
83
+ "records_per_path": {
84
+ "/workspace/data/generated/ko_pii_opf_v4/validation.jsonl": 2227
85
+ }
86
+ },
87
+ "test_dataset": {
88
+ "split": "test",
89
+ "records": 2252,
90
+ "tokens": 34553,
91
+ "spans": 542,
92
+ "spans_without_token_overlap": 0,
93
+ "truncated_examples": 0,
94
+ "max_tokens": 48,
95
+ "records_per_path": {
96
+ "/workspace/data/generated/ko_pii_opf_v4/test.jsonl": 2252
97
+ }
98
+ },
99
+ "train_metrics": {
100
+ "train_runtime": 2753.5271,
101
+ "train_samples_per_second": 149.615,
102
+ "train_steps_per_second": 1.169,
103
+ "total_flos": 1.291714856788951e+17,
104
+ "train_loss": 0.07032274794504509,
105
+ "epoch": 10.0
106
+ },
107
+ "validation_metrics": {
108
+ "eval_loss": 0.09874702990055084,
109
+ "eval_token_accuracy": 0.9907504668534081,
110
+ "eval_span_precision": 0.8582677165354331,
111
+ "eval_span_recall": 0.8384615384615385,
112
+ "eval_span_f1": 0.8482490272373542,
113
+ "eval_gold_spans": 520.0,
114
+ "eval_pred_spans": 508.0,
115
+ "eval_class_account_number_precision": 0.9736842105263158,
116
+ "eval_class_account_number_recall": 0.9823008849557522,
117
+ "eval_class_account_number_f1": 0.9779735682819383,
118
+ "eval_class_account_number_gold_spans": 113.0,
119
+ "eval_class_account_number_pred_spans": 114.0,
120
+ "eval_class_ip_address_precision": 1.0,
121
+ "eval_class_ip_address_recall": 1.0,
122
+ "eval_class_ip_address_f1": 1.0,
123
+ "eval_class_ip_address_gold_spans": 4.0,
124
+ "eval_class_ip_address_pred_spans": 4.0,
125
+ "eval_class_personal_handle_precision": 0.8571428571428571,
126
+ "eval_class_personal_handle_recall": 0.8571428571428571,
127
+ "eval_class_personal_handle_f1": 0.8571428571428571,
128
+ "eval_class_personal_handle_gold_spans": 28.0,
129
+ "eval_class_personal_handle_pred_spans": 28.0,
130
+ "eval_class_private_address_precision": 0.7619047619047619,
131
+ "eval_class_private_address_recall": 0.6666666666666666,
132
+ "eval_class_private_address_f1": 0.7111111111111111,
133
+ "eval_class_private_address_gold_spans": 48.0,
134
+ "eval_class_private_address_pred_spans": 42.0,
135
+ "eval_class_private_date_precision": 1.0,
136
+ "eval_class_private_date_recall": 1.0,
137
+ "eval_class_private_date_f1": 1.0,
138
+ "eval_class_private_date_gold_spans": 33.0,
139
+ "eval_class_private_date_pred_spans": 33.0,
140
+ "eval_class_private_email_precision": 0.926829268292683,
141
+ "eval_class_private_email_recall": 0.9743589743589743,
142
+ "eval_class_private_email_f1": 0.9500000000000001,
143
+ "eval_class_private_email_gold_spans": 39.0,
144
+ "eval_class_private_email_pred_spans": 41.0,
145
+ "eval_class_private_person_precision": 0.6710526315789473,
146
+ "eval_class_private_person_recall": 0.6257668711656442,
147
+ "eval_class_private_person_f1": 0.6476190476190476,
148
+ "eval_class_private_person_gold_spans": 163.0,
149
+ "eval_class_private_person_pred_spans": 152.0,
150
+ "eval_class_private_phone_precision": 1.0,
151
+ "eval_class_private_phone_recall": 1.0,
152
+ "eval_class_private_phone_f1": 1.0,
153
+ "eval_class_private_phone_gold_spans": 69.0,
154
+ "eval_class_private_phone_pred_spans": 69.0,
155
+ "eval_class_private_url_precision": 0.92,
156
+ "eval_class_private_url_recall": 1.0,
157
+ "eval_class_private_url_f1": 0.9583333333333334,
158
+ "eval_class_private_url_gold_spans": 23.0,
159
+ "eval_class_private_url_pred_spans": 25.0,
160
+ "eval_runtime": 7.0151,
161
+ "eval_samples_per_second": 317.46,
162
+ "eval_steps_per_second": 2.566,
163
+ "epoch": 10.0
164
+ },
165
+ "test_metrics": {
166
+ "test_loss": 0.08586616814136505,
167
+ "test_token_accuracy": 0.9924174456631841,
168
+ "test_span_precision": 0.9009708737864077,
169
+ "test_span_recall": 0.8560885608856088,
170
+ "test_span_f1": 0.8779564806054873,
171
+ "test_gold_spans": 542.0,
172
+ "test_pred_spans": 515.0,
173
+ "test_class_account_number_precision": 0.9752066115702479,
174
+ "test_class_account_number_recall": 0.9833333333333333,
175
+ "test_class_account_number_f1": 0.979253112033195,
176
+ "test_class_account_number_gold_spans": 120.0,
177
+ "test_class_account_number_pred_spans": 121.0,
178
+ "test_class_ip_address_precision": 1.0,
179
+ "test_class_ip_address_recall": 1.0,
180
+ "test_class_ip_address_f1": 1.0,
181
+ "test_class_ip_address_gold_spans": 9.0,
182
+ "test_class_ip_address_pred_spans": 9.0,
183
+ "test_class_personal_handle_precision": 0.9743589743589743,
184
+ "test_class_personal_handle_recall": 0.9743589743589743,
185
+ "test_class_personal_handle_f1": 0.9743589743589743,
186
+ "test_class_personal_handle_gold_spans": 39.0,
187
+ "test_class_personal_handle_pred_spans": 39.0,
188
+ "test_class_private_address_precision": 0.8275862068965517,
189
+ "test_class_private_address_recall": 0.7384615384615385,
190
+ "test_class_private_address_f1": 0.7804878048780489,
191
+ "test_class_private_address_gold_spans": 65.0,
192
+ "test_class_private_address_pred_spans": 58.0,
193
+ "test_class_private_date_precision": 0.9166666666666666,
194
+ "test_class_private_date_recall": 0.88,
195
+ "test_class_private_date_f1": 0.8979591836734694,
196
+ "test_class_private_date_gold_spans": 25.0,
197
+ "test_class_private_date_pred_spans": 24.0,
198
+ "test_class_private_email_precision": 1.0,
199
+ "test_class_private_email_recall": 1.0,
200
+ "test_class_private_email_f1": 1.0,
201
+ "test_class_private_email_gold_spans": 38.0,
202
+ "test_class_private_email_pred_spans": 38.0,
203
+ "test_class_private_person_precision": 0.7348484848484849,
204
+ "test_class_private_person_recall": 0.6381578947368421,
205
+ "test_class_private_person_f1": 0.6830985915492959,
206
+ "test_class_private_person_gold_spans": 152.0,
207
+ "test_class_private_person_pred_spans": 132.0,
208
+ "test_class_private_phone_precision": 1.0,
209
+ "test_class_private_phone_recall": 1.0,
210
+ "test_class_private_phone_f1": 1.0,
211
+ "test_class_private_phone_gold_spans": 76.0,
212
+ "test_class_private_phone_pred_spans": 76.0,
213
+ "test_class_private_url_precision": 1.0,
214
+ "test_class_private_url_recall": 1.0,
215
+ "test_class_private_url_f1": 1.0,
216
+ "test_class_private_url_gold_spans": 18.0,
217
+ "test_class_private_url_pred_spans": 18.0,
218
+ "test_runtime": 6.4275,
219
+ "test_samples_per_second": 350.37,
220
+ "test_steps_per_second": 2.8,
221
+ "epoch": 10.0
222
+ },
223
+ "args": {
224
+ "train_dataset": [
225
+ "/workspace/data/generated/ko_pii_opf_v4/train.jsonl"
226
+ ],
227
+ "validation_dataset": "/workspace/data/generated/ko_pii_opf_v4/validation.jsonl",
228
+ "test_dataset": "/workspace/data/generated/ko_pii_opf_v4/test.jsonl",
229
+ "label_space_json": "/workspace/data/generated/ko_pii_opf_v4/label_space.json",
230
+ "checkpoint": "/models/privacy-filter",
231
+ "output_dir": "/workspace/data/checkpoints/ko_pii_hf_ddp_v6_lora",
232
+ "max_length": 512,
233
+ "epochs": 10.0,
234
+ "early_stopping_patience": 3,
235
+ "per_device_train_batch_size": 64,
236
+ "per_device_eval_batch_size": 64,
237
+ "gradient_accumulation_steps": 1,
238
+ "learning_rate": 0.0005,
239
+ "lr_scheduler_type": "cosine",
240
+ "weight_decay": 0.01,
241
+ "warmup_ratio": 0.1,
242
+ "max_grad_norm": 1.0,
243
+ "logging_steps": 25,
244
+ "save_total_limit": 2,
245
+ "dataloader_num_workers": 4,
246
+ "seed": 42,
247
+ "overwrite_output": true,
248
+ "resume_from_checkpoint": null,
249
+ "max_train_examples": null,
250
+ "max_validation_examples": null,
251
+ "max_test_examples": null,
252
+ "use_lora": true,
253
+ "lora_r": 16,
254
+ "lora_alpha": 32,
255
+ "lora_dropout": 0.05,
256
+ "lora_target_modules": "q_proj,k_proj,v_proj,o_proj"
257
+ }
258
+ }