Text Generation
PEFT
Safetensors
English
qlora
lora
structured-output

qwen3-4b-h100-v5-hard-ep2

This LoRA adapter is highly optimized for complex structured outputs (JSON/XML/YAML/CSV/TOML). Trained on H100 with a blend of two datasets (10k rows total) and heavily preprocessed to remove CoT traces and markdown artifacts.

Training Configuration

  • Base model: Qwen/Qwen3-4B-Instruct-2507
  • Max sequence length: 2560
  • Epochs: 2
  • Learning rate: 2e-5
  • Effective Batch size: 32 (BS=16, GradAccum=2)
  • Strategy: Phase 1 & Phase 2 mixed, custom output stripping applied.
Downloads last month
257
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for shinich001/qwen3-4b-h100-v5-hard-ep2

Adapter
(5273)
this model

Dataset used to train shinich001/qwen3-4b-h100-v5-hard-ep2