bingbangboom/Qwen3006B-transcriber-beta

Post processor for local ASR.

  • Developed by: bingbangboom
  • License: apache-2.0
  • Finetuned from model : unsloth/qwen3-0.6b-unsloth-bnb-4bit

Recommended Settings

  > Temperature = 0.1
  > top_k = 10
  > top_p = 0.95
  > min_p = 0.05
  > repeat_penalty = 1.0
  > Prompt format (for chat) = {input transcript}
  > Prompt format (for use in Handy) = ${output}

Note

No System Prompt required. 
You need to disable thinking for the model by adding {%- set enable_thinking = false %} in the Jinja Prompt Template.
LMStudio: Go to model gallery, click the model entry, then in inference settings scroll to the bottom to Prompt Template and paste at top.

Available Model files:

  • Qwen3.5-0.8B.F16.gguf
  • Qwen3.5-0.8B.Q8_0.gguf
  • Qwen3.5-0.8B.Q5_K_M.ggu
  • Qwen3.5-0.8B.Q4_K_M.gguf
  • Lora merged safetensor

This qwen3 model was trained 2x faster with Unsloth and Huggingface's TRL library.

Downloads last month
1,086
Safetensors
Model size
0.6B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train bingbangboom/Qwen3006B-transcriber-beta

Collection including bingbangboom/Qwen3006B-transcriber-beta