File size: 9,638 Bytes
d8ba5fb
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
# /// script
# dependencies = [
#   "torch>=2.2.0",
#   "transformers>=4.40.0",
#   "trl>=1.2.0",
#   "peft>=0.10.0",
#   "accelerate>=0.27.0",
#   "bitsandbytes>=0.43.0",
#   "datasets>=2.18.0",
#   "huggingface-hub>=0.22.0",
#   "trackio",
#   "pydantic>=2.0",
#   "numpy>=1.24",
#   "pandas>=2.0",
#   "matplotlib>=3.8",
#   "tqdm>=4.60",
#   "networkx>=3.0",
#   "scipy>=1.10",
#   "fastapi>=0.100",
#   "uvicorn>=0.22",
#   "httpx>=0.24",
#   "pyyaml>=6.0",
# ]
# ///
"""
PolyGuard SFT Training Job β€” runs on Hugging Face Jobs cloud GPU.

This script:
  1. Clones the project from GitHub
  2. Installs the polyguard-rl package
  3. Generates the SFT dataset from the PolyGuard environment
  4. Fine-tunes Qwen2.5-1.5B-Instruct with LoRA via TRL SFTTrainer
  5. Pushes the LoRA adapter + tokenizer to the Hugging Face Hub

Submit via CLI:
  hf jobs uv run \
    --flavor a10g-large \
    --timeout 3h \
    --secrets HF_TOKEN \
    "https://huggingface.co/TheJackBright/polyguard-training-scripts/resolve/main/hf_sft_train.py"

Environment variables (passed as --env or via job secrets):
  HF_TOKEN          : HF write token (required for Hub push)
  GITHUB_REPO       : override GitHub repo URL (optional)
  MODEL_NAME        : base model (default: Qwen/Qwen2.5-1.5B-Instruct)
  HUB_MODEL_ID      : output model repo on Hub (default: TheJackBright/polyguard-qwen-sft)
  N_EPISODES        : SFT dataset episodes (default: 500)
  EPOCHS            : training epochs (default: 3)
  BATCH_SIZE        : per-device batch size (default: 2)
  GRAD_ACCUM        : gradient accumulation steps (default: 8)
  MAX_LENGTH        : max token length (default: 1024)
  LEARNING_RATE     : learning rate (default: 2e-4)
  LORA_RANK         : LoRA rank (default: 16)
"""
from __future__ import annotations

import inspect
import json
import os
import random
import subprocess
import sys
from pathlib import Path
from typing import Any, Dict, List, Optional

# ─── Config from environment ────────────────────────────────────────────────
GITHUB_REPO   = os.environ.get("GITHUB_REPO", "https://github.com/Vishwa-docs/Meta_PyTorch_Scalar_OpenEnv-Hackathon.git")
MODEL_NAME    = os.environ.get("MODEL_NAME", "Qwen/Qwen2.5-1.5B-Instruct")
HUB_MODEL_ID  = os.environ.get("HUB_MODEL_ID", "TheJackBright/polyguard-qwen-sft")
N_EPISODES    = int(os.environ.get("N_EPISODES", "500"))
EPOCHS        = int(os.environ.get("EPOCHS", "3"))
BATCH_SIZE    = int(os.environ.get("BATCH_SIZE", "2"))
GRAD_ACCUM    = int(os.environ.get("GRAD_ACCUM", "8"))
MAX_LENGTH    = int(os.environ.get("MAX_LENGTH", "1024"))
LEARNING_RATE = float(os.environ.get("LEARNING_RATE", "2e-4"))
LORA_RANK     = int(os.environ.get("LORA_RANK", "16"))
SEED          = 42
OUTPUT_DIR    = "/tmp/polyguard_sft_output"
DATA_PATH     = "/tmp/polyguard_sft_data.jsonl"
DATA_FMT_PATH = "/tmp/polyguard_sft_data_formatted.jsonl"

SYSTEM_PROMPT = (
    "You are a clinical pharmacist agent performing polypharmacy medication review. "
    "Analyze drug-drug interactions, Beers criteria violations, and propose safe interventions. "
    "Respond only with a structured <decision>...</decision> XML action."
)

print("=" * 60)
print("PolyGuard SFT Training on HF Jobs")
print(f"  Model:        {MODEL_NAME}")
print(f"  Hub output:   {HUB_MODEL_ID}")
print(f"  Episodes:     {N_EPISODES}")
print(f"  Epochs:       {EPOCHS}")
print(f"  Batch size:   {BATCH_SIZE} x {GRAD_ACCUM} grad accum")
print(f"  Max length:   {MAX_LENGTH}")
print(f"  LoRA rank:    {LORA_RANK}")
print("=" * 60)


# ─── Step 1: Clone repo and install polyguard-rl ────────────────────────────
print("\n[1/5] Cloning project from GitHub...")
clone_dir = Path("/tmp/polyguard_project")
if not clone_dir.exists():
    subprocess.run(
        ["git", "clone", "--depth=1", GITHUB_REPO, str(clone_dir)],
        check=True,
    )
    print(f"  Cloned to {clone_dir}")
else:
    print(f"  Already cloned at {clone_dir}")

polyguard_rl_dir = clone_dir / "polyguard-rl"
print(f"\n[2/5] Installing polyguard-rl from {polyguard_rl_dir}...")
subprocess.run(
    [sys.executable, "-m", "pip", "install", "-e", str(polyguard_rl_dir), "--quiet"],
    check=True,
)
print("  polyguard-rl installed.")

# Add to sys.path so relative imports work during dataset generation
if str(polyguard_rl_dir) not in sys.path:
    sys.path.insert(0, str(polyguard_rl_dir))


# ─── Step 2: Generate SFT dataset ───────────────────────────────────────────
print(f"\n[3/5] Generating SFT dataset ({N_EPISODES} episodes)...")

os.chdir(polyguard_rl_dir)

from app.training.sft_dataset import generate_sft_dataset, build_external_ddi_sft_examples, format_for_training  # noqa: E402

examples = generate_sft_dataset(
    n_episodes=N_EPISODES,
    seed=SEED,
    output_path=DATA_PATH,
)

print(f"  Generated {len(examples)} episodes.")

# Format for TRL (convert to messages format)
formatted = format_for_training(examples, system_prompt=SYSTEM_PROMPT)
print(f"  Formatted {len(formatted)} training rows.")

with open(DATA_FMT_PATH, "w") as f:
    for row in formatted:
        f.write(json.dumps(row) + "\n")

print(f"  Saved formatted dataset to {DATA_FMT_PATH}")


# ─── Step 3: Load model with LoRA ───────────────────────────────────────────
print(f"\n[4/5] Loading model and running SFT training...")

import torch
from datasets import Dataset
from peft import LoraConfig, get_peft_model
from transformers import AutoModelForCausalLM, AutoTokenizer
from trl import SFTConfig, SFTTrainer

device_map = "auto" if torch.cuda.is_available() else "cpu"
dtype = torch.bfloat16 if (torch.cuda.is_available() and torch.cuda.is_bf16_supported()) else torch.float16

print(f"  CUDA available: {torch.cuda.is_available()}")
if torch.cuda.is_available():
    print(f"  GPU: {torch.cuda.get_device_name(0)}")
    print(f"  VRAM: {torch.cuda.get_device_properties(0).total_memory / 1e9:.1f} GB")

print(f"  Loading tokenizer from {MODEL_NAME}...")
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, trust_remote_code=True)
if tokenizer.pad_token is None:
    tokenizer.pad_token = tokenizer.eos_token
    tokenizer.pad_token_id = tokenizer.eos_token_id

print(f"  Loading base model...")
model = AutoModelForCausalLM.from_pretrained(
    MODEL_NAME,
    torch_dtype=dtype,
    device_map=device_map,
    trust_remote_code=True,
)

lora_config = LoraConfig(
    r=LORA_RANK,
    lora_alpha=LORA_RANK * 2,
    target_modules=["q_proj", "k_proj", "v_proj", "o_proj", "gate_proj", "up_proj", "down_proj"],
    lora_dropout=0.05,
    bias="none",
    task_type="CAUSAL_LM",
)
model = get_peft_model(model, lora_config)
model.print_trainable_parameters()


# ─── Step 4: Build dataset and trainer ──────────────────────────────────────
random.seed(SEED)
ds_full = Dataset.from_list(formatted).shuffle(seed=SEED)
split = ds_full.train_test_split(test_size=0.1, seed=SEED)

sft_config_kwargs: Dict[str, Any] = {
    "output_dir": OUTPUT_DIR,
    "num_train_epochs": EPOCHS,
    "per_device_train_batch_size": BATCH_SIZE,
    "gradient_accumulation_steps": GRAD_ACCUM,
    "learning_rate": LEARNING_RATE,
    "warmup_ratio": 0.05,
    "weight_decay": 0.01,
    "bf16": dtype == torch.bfloat16,
    "fp16": dtype == torch.float16,
    "logging_steps": 10,
    "save_steps": 100,
    "save_total_limit": 2,
    "max_grad_norm": 1.0,
    "seed": SEED,
    "report_to": ["trackio"],
    "run_name": "polyguard-sft-qwen",
    "project": "polyguard-training",
    "push_to_hub": True,
    "hub_model_id": HUB_MODEL_ID,
    "hub_strategy": "every_save",
    "eval_strategy": "steps",
    "eval_steps": 50,
}

# Adapt to TRL version
sft_params = set(inspect.signature(SFTConfig).parameters)
if "max_length" in sft_params:
    sft_config_kwargs["max_length"] = MAX_LENGTH
elif "max_seq_length" in sft_params:
    sft_config_kwargs["max_seq_length"] = MAX_LENGTH
if "eos_token" in sft_params:
    sft_config_kwargs["eos_token"] = "<|im_end|>"

sft_config = SFTConfig(**{k: v for k, v in sft_config_kwargs.items() if k in sft_params})

trainer_kwargs: Dict[str, Any] = {
    "model": model,
    "args": sft_config,
    "train_dataset": split["train"],
    "eval_dataset": split["test"],
}
trainer_params = set(inspect.signature(SFTTrainer).parameters)
if "processing_class" in trainer_params:
    trainer_kwargs["processing_class"] = tokenizer
elif "tokenizer" in trainer_params:
    trainer_kwargs["tokenizer"] = tokenizer

trainer = SFTTrainer(**{k: v for k, v in trainer_kwargs.items() if v is not None})

print(f"\n  Training dataset: {len(split['train'])} rows")
print(f"  Eval dataset:     {len(split['test'])} rows")
print("\n  Starting training...\n")


# ─── Step 5: Train and push to Hub ─────────────────────────────────────────
train_result = trainer.train()

print("\n  Training complete.")
print(f"  Train loss:   {train_result.training_loss:.4f}")
print(f"  Train steps:  {train_result.global_step}")

print(f"\n[5/5] Pushing model to Hub: {HUB_MODEL_ID}...")
trainer.push_to_hub()
tokenizer.push_to_hub(HUB_MODEL_ID)

print("\n" + "=" * 60)
print("SFT training complete!")
print(f"Model saved to: https://huggingface.co/{HUB_MODEL_ID}")
print("=" * 60)