TRACER: Trace-Based Adaptive Cost-Efficient Routing for LLM Classification
Paper • 2604.14531 • Published • 6
Hugging Face Inference Endpoints Images repository allows AI Builders to collaborate and engage creating awesome inference deployments
from trl.experimental.ssd import SSDConfig, SSDTrainer
trainer = SSDTrainer(
model="Qwen/Qwen3-4B-Instruct",
args=SSDConfig(temperature=0.6, top_k=20, top_p=0.95),
train_dataset=dataset,
)
trainer.train()use_transformers_paged, and key fixes for VLM response parsing.pip install --upgrade trlhf-mem v0.4.1 now also estimates KV cache memory requirements for any context length and batch size with the --experimental flag!uvx hf-mem --model-id ... --experimental will automatically pull the required information from the Hugging Face Hub to include the KV cache estimation, when applicable.--max-model-len, --batch-size and --kv-cache-dtype arguments (à la vLLM) manually if preferred.