nraptisss's picture
Upload README.md
4d20f09 verified
# TMF921 Intent-to-Configuration Translation β€” Training Pipeline
> **The first open-source fine-tuning pipeline for translating natural language network intents into spec-compliant 5G/6G configurations across 6 telecom standards.**
## What This Does
Fine-tunes an LLM (default: Qwen3-8B) via **4-bit QLoRA** on the [TMF921-intent-to-config-augmented](https://huggingface.co/datasets/nraptisss/TMF921-intent-to-config-augmented) dataset (41.8K samples) to translate natural language network intents into structured JSON configurations for:
| Target Standard | Description |
|---|---|
| **TMF921** | TM Forum Intent Management API v4 |
| **3GPP TS 28.312** | Intent-Driven Management (Rel-18) |
| **CAMARA** | GSMA Open Gateway NetworkSliceBooking |
| **ETSI ZSM** | Zero-touch Service Management intents |
| **O-RAN A1 Policy** | Non-RT RIC A1 policy objects |
| **3GPP O1 NRM** | O1 interface NR Network Resource Model |
Plus **TMF921 lifecycle operations** (activate, modify, suspend, resume, terminate, scale, monitor, report) and **adversarial rejection** (ambiguous, out-of-scope, contradictory intents).
## Quick Start
### 1. Clone & Install
```bash
git clone https://huggingface.co/nraptisss/intent-translation-training
cd intent-translation-training
pip install -r requirements.txt
```
### 2. Train (defaults β€” RTX 6000 Ada 50GB)
```bash
python train.py
```
This will:
- Load Qwen3-8B in 4-bit NF4 quantization
- Apply QLoRA (r=32, alpha=64) to all linear layers
- Train for 3 epochs with cosine LR schedule
- Save checkpoints and best model to `./output/`
### 3. Evaluate
```bash
python evaluate.py --adapter_path ./output --num_samples 200
```
### 4. One-command pipeline
```bash
chmod +x run.sh
./run.sh
```
## Training Configuration
| Parameter | Default | Notes |
|---|---|---|
| Base model | `Qwen/Qwen3-8B` | Any HF causal LM works |
| Quantization | 4-bit NF4 + double quant | ~4.5 GB VRAM for weights |
| LoRA rank | 32 | target_modules="all-linear" |
| LoRA alpha | 64 | 2Γ— rank |
| Batch size | 4 Γ— 8 grad_accum = 32 effective | |
| Learning rate | 1e-4 | 10Γ— base for LoRA |
| Epochs | 3 | ~3,700 steps total |
| Max length | 4096 tokens | Covers 99%+ of samples |
| Loss | assistant_only_loss=True | Train only on config outputs |
| Precision | bf16 | |
| Flash attention | Yes (flash_attention_2) | |
| Gradient checkpointing | Yes | Saves ~40% VRAM |
### VRAM Estimate (RTX 6000 Ada 50GB)
| Component | VRAM |
|---|---|
| Model weights (4-bit) | ~4.5 GB |
| LoRA adapters | ~0.5 GB |
| Activations + gradients (checkpointed) | ~12 GB |
| Optimizer states (AdamW, LoRA params only) | ~1 GB |
| KV cache / batch | ~8 GB |
| **Total estimated** | **~26 GB** |
Leaves ~24 GB headroom. You can increase batch_size to 8 if desired.
## Customization
### Different base model
```bash
python train.py --base_model Qwen/Qwen2.5-7B-Instruct
python train.py --base_model meta-llama/Llama-3.1-8B-Instruct
python train.py --base_model microsoft/phi-4
```
### Push to Hugging Face Hub
```bash
export HF_TOKEN="hf_..."
python train.py --push_to_hub --hub_model_id your-username/model-name
```
### Adjust hyperparameters
```bash
python train.py --lr 2e-4 --epochs 5 --lora_r 64 --lora_alpha 128 --batch_size 2 --grad_accum 16
```
### No flash attention (older GPUs)
```bash
python train.py --no_flash_attn
```
## Evaluation Metrics
The evaluation script measures:
| Metric | Description |
|---|---|
| **JSON Validity Rate** | % of outputs that are valid JSON |
| **Structure Correctness** | % with correct root keys per standard |
| **KPI Field Accuracy** | % containing correct latency, throughput, reliability, UE values |
| **All KPIs Correct** | % where ALL 5 KPI fields match |
| **Adversarial Accuracy** | % of bad intents correctly rejected |
Results are broken down per target layer (TMF921, 3GPP, CAMARA, etc.).
## File Structure
```
β”œβ”€β”€ train.py # QLoRA fine-tuning script
β”œβ”€β”€ evaluate.py # Evaluation with per-layer metrics
β”œβ”€β”€ run.sh # One-command train+eval pipeline
β”œβ”€β”€ requirements.txt # Python dependencies
└── README.md # This file
```
## Dataset
[nraptisss/TMF921-intent-to-config-augmented](https://huggingface.co/datasets/nraptisss/TMF921-intent-to-config-augmented) β€” 41,815 samples:
- **39,153** standard intent→config pairs across 6 target layers
- **1,552** lifecycle operation samples (8 TMF921 operations)
- **141** adversarial samples (ambiguous, out-of-scope, contradictory)
- **18 sectors**, **147 use cases**, **55 regions**
- **6 slice types**: eMBB, URLLC, mMTC, V2X, MPS, HMTC
## References
- [ORION: Intent-Aware Orchestration in Open RAN](https://arxiv.org/abs/2603.03667)
- [TelecomGPT](https://arxiv.org/abs/2407.09424)
- [ORANSight-2.0](https://arxiv.org/abs/2503.05200)
- [NEFMind](https://arxiv.org/abs/2508.09240)
- [TMF921 Intent Management API v5.0](https://www.tmforum.org/resources/specification/tmf921-intent-management-api-rest-specification-r21-0-0/)
- [3GPP TS 28.312 Rel-18](https://www.3gpp.org/dynareport/28312.htm)
## License
Apache 2.0