polyguard-training-scripts / hf_sft_train.py
TheJackBright's picture
Upload hf_sft_train.py with huggingface_hub
d8ba5fb verified
# /// script
# dependencies = [
# "torch>=2.2.0",
# "transformers>=4.40.0",
# "trl>=1.2.0",
# "peft>=0.10.0",
# "accelerate>=0.27.0",
# "bitsandbytes>=0.43.0",
# "datasets>=2.18.0",
# "huggingface-hub>=0.22.0",
# "trackio",
# "pydantic>=2.0",
# "numpy>=1.24",
# "pandas>=2.0",
# "matplotlib>=3.8",
# "tqdm>=4.60",
# "networkx>=3.0",
# "scipy>=1.10",
# "fastapi>=0.100",
# "uvicorn>=0.22",
# "httpx>=0.24",
# "pyyaml>=6.0",
# ]
# ///
"""
PolyGuard SFT Training Job β€” runs on Hugging Face Jobs cloud GPU.
This script:
1. Clones the project from GitHub
2. Installs the polyguard-rl package
3. Generates the SFT dataset from the PolyGuard environment
4. Fine-tunes Qwen2.5-1.5B-Instruct with LoRA via TRL SFTTrainer
5. Pushes the LoRA adapter + tokenizer to the Hugging Face Hub
Submit via CLI:
hf jobs uv run \
--flavor a10g-large \
--timeout 3h \
--secrets HF_TOKEN \
"https://huggingface.co/TheJackBright/polyguard-training-scripts/resolve/main/hf_sft_train.py"
Environment variables (passed as --env or via job secrets):
HF_TOKEN : HF write token (required for Hub push)
GITHUB_REPO : override GitHub repo URL (optional)
MODEL_NAME : base model (default: Qwen/Qwen2.5-1.5B-Instruct)
HUB_MODEL_ID : output model repo on Hub (default: TheJackBright/polyguard-qwen-sft)
N_EPISODES : SFT dataset episodes (default: 500)
EPOCHS : training epochs (default: 3)
BATCH_SIZE : per-device batch size (default: 2)
GRAD_ACCUM : gradient accumulation steps (default: 8)
MAX_LENGTH : max token length (default: 1024)
LEARNING_RATE : learning rate (default: 2e-4)
LORA_RANK : LoRA rank (default: 16)
"""
from __future__ import annotations
import inspect
import json
import os
import random
import subprocess
import sys
from pathlib import Path
from typing import Any, Dict, List, Optional
# ─── Config from environment ────────────────────────────────────────────────
GITHUB_REPO = os.environ.get("GITHUB_REPO", "https://github.com/Vishwa-docs/Meta_PyTorch_Scalar_OpenEnv-Hackathon.git")
MODEL_NAME = os.environ.get("MODEL_NAME", "Qwen/Qwen2.5-1.5B-Instruct")
HUB_MODEL_ID = os.environ.get("HUB_MODEL_ID", "TheJackBright/polyguard-qwen-sft")
N_EPISODES = int(os.environ.get("N_EPISODES", "500"))
EPOCHS = int(os.environ.get("EPOCHS", "3"))
BATCH_SIZE = int(os.environ.get("BATCH_SIZE", "2"))
GRAD_ACCUM = int(os.environ.get("GRAD_ACCUM", "8"))
MAX_LENGTH = int(os.environ.get("MAX_LENGTH", "1024"))
LEARNING_RATE = float(os.environ.get("LEARNING_RATE", "2e-4"))
LORA_RANK = int(os.environ.get("LORA_RANK", "16"))
SEED = 42
OUTPUT_DIR = "/tmp/polyguard_sft_output"
DATA_PATH = "/tmp/polyguard_sft_data.jsonl"
DATA_FMT_PATH = "/tmp/polyguard_sft_data_formatted.jsonl"
SYSTEM_PROMPT = (
"You are a clinical pharmacist agent performing polypharmacy medication review. "
"Analyze drug-drug interactions, Beers criteria violations, and propose safe interventions. "
"Respond only with a structured <decision>...</decision> XML action."
)
print("=" * 60)
print("PolyGuard SFT Training on HF Jobs")
print(f" Model: {MODEL_NAME}")
print(f" Hub output: {HUB_MODEL_ID}")
print(f" Episodes: {N_EPISODES}")
print(f" Epochs: {EPOCHS}")
print(f" Batch size: {BATCH_SIZE} x {GRAD_ACCUM} grad accum")
print(f" Max length: {MAX_LENGTH}")
print(f" LoRA rank: {LORA_RANK}")
print("=" * 60)
# ─── Step 1: Clone repo and install polyguard-rl ────────────────────────────
print("\n[1/5] Cloning project from GitHub...")
clone_dir = Path("/tmp/polyguard_project")
if not clone_dir.exists():
subprocess.run(
["git", "clone", "--depth=1", GITHUB_REPO, str(clone_dir)],
check=True,
)
print(f" Cloned to {clone_dir}")
else:
print(f" Already cloned at {clone_dir}")
polyguard_rl_dir = clone_dir / "polyguard-rl"
print(f"\n[2/5] Installing polyguard-rl from {polyguard_rl_dir}...")
subprocess.run(
[sys.executable, "-m", "pip", "install", "-e", str(polyguard_rl_dir), "--quiet"],
check=True,
)
print(" polyguard-rl installed.")
# Add to sys.path so relative imports work during dataset generation
if str(polyguard_rl_dir) not in sys.path:
sys.path.insert(0, str(polyguard_rl_dir))
# ─── Step 2: Generate SFT dataset ───────────────────────────────────────────
print(f"\n[3/5] Generating SFT dataset ({N_EPISODES} episodes)...")
os.chdir(polyguard_rl_dir)
from app.training.sft_dataset import generate_sft_dataset, build_external_ddi_sft_examples, format_for_training # noqa: E402
examples = generate_sft_dataset(
n_episodes=N_EPISODES,
seed=SEED,
output_path=DATA_PATH,
)
print(f" Generated {len(examples)} episodes.")
# Format for TRL (convert to messages format)
formatted = format_for_training(examples, system_prompt=SYSTEM_PROMPT)
print(f" Formatted {len(formatted)} training rows.")
with open(DATA_FMT_PATH, "w") as f:
for row in formatted:
f.write(json.dumps(row) + "\n")
print(f" Saved formatted dataset to {DATA_FMT_PATH}")
# ─── Step 3: Load model with LoRA ───────────────────────────────────────────
print(f"\n[4/5] Loading model and running SFT training...")
import torch
from datasets import Dataset
from peft import LoraConfig, get_peft_model
from transformers import AutoModelForCausalLM, AutoTokenizer
from trl import SFTConfig, SFTTrainer
device_map = "auto" if torch.cuda.is_available() else "cpu"
dtype = torch.bfloat16 if (torch.cuda.is_available() and torch.cuda.is_bf16_supported()) else torch.float16
print(f" CUDA available: {torch.cuda.is_available()}")
if torch.cuda.is_available():
print(f" GPU: {torch.cuda.get_device_name(0)}")
print(f" VRAM: {torch.cuda.get_device_properties(0).total_memory / 1e9:.1f} GB")
print(f" Loading tokenizer from {MODEL_NAME}...")
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, trust_remote_code=True)
if tokenizer.pad_token is None:
tokenizer.pad_token = tokenizer.eos_token
tokenizer.pad_token_id = tokenizer.eos_token_id
print(f" Loading base model...")
model = AutoModelForCausalLM.from_pretrained(
MODEL_NAME,
torch_dtype=dtype,
device_map=device_map,
trust_remote_code=True,
)
lora_config = LoraConfig(
r=LORA_RANK,
lora_alpha=LORA_RANK * 2,
target_modules=["q_proj", "k_proj", "v_proj", "o_proj", "gate_proj", "up_proj", "down_proj"],
lora_dropout=0.05,
bias="none",
task_type="CAUSAL_LM",
)
model = get_peft_model(model, lora_config)
model.print_trainable_parameters()
# ─── Step 4: Build dataset and trainer ──────────────────────────────────────
random.seed(SEED)
ds_full = Dataset.from_list(formatted).shuffle(seed=SEED)
split = ds_full.train_test_split(test_size=0.1, seed=SEED)
sft_config_kwargs: Dict[str, Any] = {
"output_dir": OUTPUT_DIR,
"num_train_epochs": EPOCHS,
"per_device_train_batch_size": BATCH_SIZE,
"gradient_accumulation_steps": GRAD_ACCUM,
"learning_rate": LEARNING_RATE,
"warmup_ratio": 0.05,
"weight_decay": 0.01,
"bf16": dtype == torch.bfloat16,
"fp16": dtype == torch.float16,
"logging_steps": 10,
"save_steps": 100,
"save_total_limit": 2,
"max_grad_norm": 1.0,
"seed": SEED,
"report_to": ["trackio"],
"run_name": "polyguard-sft-qwen",
"project": "polyguard-training",
"push_to_hub": True,
"hub_model_id": HUB_MODEL_ID,
"hub_strategy": "every_save",
"eval_strategy": "steps",
"eval_steps": 50,
}
# Adapt to TRL version
sft_params = set(inspect.signature(SFTConfig).parameters)
if "max_length" in sft_params:
sft_config_kwargs["max_length"] = MAX_LENGTH
elif "max_seq_length" in sft_params:
sft_config_kwargs["max_seq_length"] = MAX_LENGTH
if "eos_token" in sft_params:
sft_config_kwargs["eos_token"] = "<|im_end|>"
sft_config = SFTConfig(**{k: v for k, v in sft_config_kwargs.items() if k in sft_params})
trainer_kwargs: Dict[str, Any] = {
"model": model,
"args": sft_config,
"train_dataset": split["train"],
"eval_dataset": split["test"],
}
trainer_params = set(inspect.signature(SFTTrainer).parameters)
if "processing_class" in trainer_params:
trainer_kwargs["processing_class"] = tokenizer
elif "tokenizer" in trainer_params:
trainer_kwargs["tokenizer"] = tokenizer
trainer = SFTTrainer(**{k: v for k, v in trainer_kwargs.items() if v is not None})
print(f"\n Training dataset: {len(split['train'])} rows")
print(f" Eval dataset: {len(split['test'])} rows")
print("\n Starting training...\n")
# ─── Step 5: Train and push to Hub ─────────────────────────────────────────
train_result = trainer.train()
print("\n Training complete.")
print(f" Train loss: {train_result.training_loss:.4f}")
print(f" Train steps: {train_result.global_step}")
print(f"\n[5/5] Pushing model to Hub: {HUB_MODEL_ID}...")
trainer.push_to_hub()
tokenizer.push_to_hub(HUB_MODEL_ID)
print("\n" + "=" * 60)
print("SFT training complete!")
print(f"Model saved to: https://huggingface.co/{HUB_MODEL_ID}")
print("=" * 60)