PEFT
qlora
sft
trl
qwen3
tmf921
intent-based-networking
network-slicing
rtx-6000-ada
ml-intern
nraptisss commited on
Commit
e902bb8
·
verified ·
1 Parent(s): 300682c

Upload paper/outline.md

Browse files
Files changed (1) hide show
  1. paper/outline.md +178 -0
paper/outline.md ADDED
@@ -0,0 +1,178 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Paper Outline
2
+
3
+ Working title:
4
+
5
+ > **A Research-Grade Benchmark and QLoRA Baseline for Multi-Standard Telecom Intent-to-Configuration Translation**
6
+
7
+ Alternative titles:
8
+
9
+ 1. **From Natural Language Intent to Telecom Configuration: Dataset, OOD Evaluation, and QLoRA Baselines Across TMF921, 3GPP, CAMARA, O-RAN, and ETSI ZSM**
10
+ 2. **Multi-Standard Telecom Intent-to-Configuration Translation: An Open Dataset, OOD Benchmark, and Qwen3 QLoRA Baseline**
11
+ 3. **Benchmarking LLMs for Intent-Based Network Configuration Across Telecom Standards**
12
+
13
+ ---
14
+
15
+ ## One-sentence thesis
16
+
17
+ We introduce an open research dataset, OOD benchmark, normalized evaluator, and Qwen3-8B QLoRA baseline for multi-standard telecom intent-to-configuration translation, showing strong JSON validity and structural correctness while identifying O1 NRM and A1 policy value fidelity as the main remaining bottlenecks.
18
+
19
+ ---
20
+
21
+ ## Core contributions
22
+
23
+ 1. **Research-grade dataset and benchmark**
24
+ - Derived from `nraptisss/TMF921-intent-to-config-augmented`.
25
+ - Adds benchmark splits for in-distribution, template-OOD, use-case-OOD, sector-OOD, and adversarial evaluation.
26
+ - Adds provenance, template/scenario IDs, token-length audit, validation flags, and rare-class-aware training split.
27
+
28
+ 2. **Reproducible training pipeline**
29
+ - Single RTX 6000 Ada QLoRA recipe for Qwen3-8B.
30
+ - Includes nohup-based resumability, GPU preflight, adapter merging, and Hub-ready artifacts.
31
+
32
+ 3. **Evaluation methodology**
33
+ - Raw JSON exact/field metrics.
34
+ - Normalized evaluator that removes volatile IDs, hrefs, timestamps, schema links, and generated identifiers.
35
+ - Stratified reporting by split, target layer, slice type, and lifecycle operation.
36
+
37
+ 4. **Baseline model and empirical results**
38
+ - Qwen3-8B QLoRA achieves near-perfect JSON validity.
39
+ - Normalized key F1 is about 98% across non-adversarial ID/OOD splits.
40
+ - Normalized field F1 is about 77–80% across non-adversarial ID/OOD splits.
41
+ - Adversarial rejection reaches 96.97% normalized exact match.
42
+
43
+ 5. **Weak-layer diagnostic experiment**
44
+ - A stage-2 weak-layer continuation experiment did not materially improve O1/A1 value fidelity.
45
+ - This negative result suggests better semantic supervision/evaluation is needed, not merely more weak-layer exposure.
46
+
47
+ ---
48
+
49
+ ## Proposed abstract draft
50
+
51
+ Intent-Based Networking aims to let operators express high-level network goals in natural language and automatically translate them into executable configurations. However, open research resources for training and evaluating such systems across multiple telecom standards remain limited. We present a research-grade benchmark and reproducible QLoRA baseline for multi-standard telecom intent-to-configuration translation. Starting from a 41,815-example TMF921 intent-to-configuration corpus, we construct a benchmark with in-distribution, template-OOD, use-case-OOD, sector-OOD, and adversarial splits, plus provenance, validation, and token-length metadata. We fine-tune Qwen3-8B using QLoRA on a single RTX 6000 Ada GPU and evaluate with both raw JSON metrics and a normalized evaluator that removes volatile identifiers, hrefs, timestamps, and schema links. The model achieves near-perfect JSON validity across all splits, normalized structural key F1 around 98%, normalized field F1 of 79.6% in-distribution, 78.7% on template-OOD, 79.1% on use-case-OOD, 77.0% on sector-OOD, and 96.97% normalized exact match on adversarial rejection. A targeted weak-layer continuation experiment shows that simple oversampling does not solve O1 NRM and A1 policy value-fidelity errors, motivating layer-specific semantic evaluators and canonical scenario generation. We release the dataset, training code, evaluation scripts, metrics, and qualitative examples to support reproducible research in telecom LLMs and intent-based network management.
52
+
53
+ ---
54
+
55
+ ## Paper structure
56
+
57
+ ### 1. Introduction
58
+
59
+ Purpose:
60
+
61
+ - Motivate intent-based networking and natural-language-to-config translation.
62
+ - Explain need for open data and reproducible baselines.
63
+ - Introduce multi-standard challenge: TMF921, 3GPP, ETSI ZSM, CAMARA, O-RAN A1, O1 NRM.
64
+ - State contributions.
65
+
66
+ Key points:
67
+
68
+ - Frontier LLMs may perform well, but telecom systems need open, auditable, deployable models.
69
+ - Raw JSON exact match is not enough due to volatile fields.
70
+ - OOD evaluation is necessary because synthetic template datasets can inflate in-distribution performance.
71
+
72
+ ### 2. Related Work
73
+
74
+ Subsections:
75
+
76
+ - LLMs for telecommunications.
77
+ - Telecom benchmarks: TeleQnA, SPEC5G, ORANBench, TSpec-LLM, TelecomGPT-related resources.
78
+ - Intent-based networking and network slicing automation.
79
+ - Structured-output / JSON-constrained generation.
80
+ - QLoRA and parameter-efficient adaptation.
81
+
82
+ ### 3. Dataset and Benchmark Construction
83
+
84
+ Subsections:
85
+
86
+ - Source dataset summary.
87
+ - Data audit and motivation for research splits.
88
+ - Split construction:
89
+ - `train_base`
90
+ - `train_sota`
91
+ - `validation`
92
+ - `test_in_distribution`
93
+ - `test_template_ood`
94
+ - `test_use_case_ood`
95
+ - `test_sector_ood`
96
+ - `test_adversarial`
97
+ - Metadata and validation columns.
98
+ - Token length audit.
99
+ - Limitations of synthetic data.
100
+
101
+ ### 4. Training Method
102
+
103
+ Subsections:
104
+
105
+ - Base model: Qwen3-8B.
106
+ - QLoRA setup:
107
+ - NF4 4-bit quantization.
108
+ - double quantization.
109
+ - LoRA rank 64.
110
+ - all-linear target modules.
111
+ - assistant-only SFT loss.
112
+ - Hardware: RTX 6000 Ada.
113
+ - Reproducible training scripts and run management.
114
+ - Stage-2 continuation setup.
115
+
116
+ ### 5. Evaluation Methodology
117
+
118
+ Subsections:
119
+
120
+ - Raw metrics:
121
+ - JSON parse rate.
122
+ - canonical exact match.
123
+ - flattened field precision/recall/F1.
124
+ - KPI text-presence diagnostics.
125
+ - Normalized metrics:
126
+ - remove volatile IDs/hrefs/timestamps/descriptions/schema links.
127
+ - sort lists deterministically.
128
+ - normalized field F1.
129
+ - normalized key F1.
130
+ - OOD evaluation protocol.
131
+ - Adversarial evaluation.
132
+ - Known limitations of current evaluator.
133
+
134
+ ### 6. Results
135
+
136
+ Subsections:
137
+
138
+ - Main stage-1 metrics.
139
+ - OOD generalization.
140
+ - Per-target-layer performance.
141
+ - Raw vs normalized evaluation.
142
+ - Adversarial rejection.
143
+
144
+ ### 7. Weak-Layer and Error Analysis
145
+
146
+ Subsections:
147
+
148
+ - O1 NRM: high structural correctness but low value fidelity.
149
+ - A1 policy: moderate value fidelity.
150
+ - Lifecycle report/monitor: measurement mismatch issues.
151
+ - Qualitative examples.
152
+ - Stage-2 weak-layer continuation negative result.
153
+
154
+ ### 8. Limitations
155
+
156
+ Must state clearly:
157
+
158
+ - Synthetic benchmark, not real operator logs.
159
+ - No official standards validators yet.
160
+ - Normalized JSON metrics are a proxy for semantic correctness.
161
+ - O1/A1 need layer-specific semantic evaluators.
162
+ - No human expert validation yet.
163
+ - Low-level values may require canonical scenario generation.
164
+
165
+ ### 9. Conclusion
166
+
167
+ Reiterate:
168
+
169
+ - Open dataset + OOD benchmark + reproducible QLoRA baseline.
170
+ - Strong JSON validity and structural correctness.
171
+ - Honest weak-layer findings.
172
+ - Future work: semantic validators and Gen4 canonical data.
173
+
174
+ ---
175
+
176
+ ## Main conclusion to defend
177
+
178
+ This is not yet a production-certified telecom configuration system. It is a strong open research benchmark and baseline that advances reproducible study of multi-standard telecom intent-to-configuration translation.