PEFT
Safetensors
lora
orchestrator
qwen3
tool-use
agent

Qwen3-8B Orchestrator LoRA โ€” With Failures

LoRA fine-tuned Qwen/Qwen3-8B for agentic orchestration tasks (tool use, multi-turn reasoning, web search).

Training Details

Parameter Value
Base Model Qwen/Qwen3-8B
LoRA Rank 128
LoRA Alpha 256
Learning Rate 2.38e-05
Epochs 1
Validation Loss 0.2352
Training Samples 52,204
Dataset GLM-4.7-flash SFT traces (includes failed traces)
Context Length 16,384 tokens
Quantization 4-bit (QLoRA)

Evaluation Results

Benchmark Score
SimpleQA (200) 30.0% accuracy
GAIA (165) 10.9% accuracy
HLE (180) 6.1% accuracy
DeepResearch (100) 0.2340 score

Usage

from peft import AutoPeftModelForCausalLM
from transformers import AutoTokenizer

model = AutoPeftModelForCausalLM.from_pretrained(
    "akenginorhun/qwen3-8b-orchestrator-lora-with-failures",
    device_map="auto",
)
tokenizer = AutoTokenizer.from_pretrained(
    "akenginorhun/qwen3-8b-orchestrator-lora-with-failures"
)

W&B Sweep

Downloads last month
1
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for akenginorhun/qwen3-8b-orchestrator-lora-with-failures

Finetuned
Qwen/Qwen3-8B
Adapter
(1071)
this model