Qwen3.5-9B-NSC-ACE-200-Merged

This is the merged full-weight release of DJLougen/Qwen3.5-9B-NSC-ACE-200, merged into Qwen/Qwen3.5-9B.

NSC-ACE improves structured agentic tool-calling behavior, especially tool-call format reliability and required argument quality.

Highlights

Metric Base NSC-ACE
Held-out Acta composite structural score 0.804 0.947
Held-out Acta tool-call rate 82.5% 97.5%
Held-out Acta reasoning tag rate 47.5% 97.5%
BFCL subset exact required call accuracy 67.5% 75.0%
BFCL required argument name accuracy 82.7% 91.8%
BFCL required argument value accuracy 72.5% 81.6%

Loading

import torch
from transformers import AutoModelForImageTextToText, AutoTokenizer

model_id = "DJLougen/Qwen3.5-9B-NSC-ACE-200-Merged"
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
model = AutoModelForImageTextToText.from_pretrained(
    model_id,
    torch_dtype=torch.bfloat16,
    device_map="auto",
    trust_remote_code=True,
)

Notes

This is a merged model for easier deployment and downstream quantization. The original PEFT adapter remains available at DJLougen/Qwen3.5-9B-NSC-ACE-200.

Downloads last month
35
Safetensors
Model size
9B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for GestaltLabs/Qwen3.5-9B-NSC-ACE

Finetuned
Qwen/Qwen3.5-9B
Finetuned
(253)
this model
Finetunes
1 model

Dataset used to train GestaltLabs/Qwen3.5-9B-NSC-ACE