Qwen3-TTS-12Hz-0.6B-Base-QTS

Qwen3-TTS-12Hz-0.6B-Base-QTS is a distribution repository for model artifacts produced by yet-another-ai/qts.

This Hugging Face repository is intended to contain stable, downloadable runtime artifacts only:

  • one shared qwen3-tts-vocoder.onnx
  • one or more GGUF variants such as qwen3-tts-0.6b-f16.gguf
  • optional additional GGUF variants such as qwen3-tts-0.6b-q8_0.gguf

It is not the source-of-truth repository for code, export logic, or developer documentation. Those live in yet-another-ai/qts.

Relationship To qts

Recommended maintenance flow:

  1. Change behavior in the GitHub repository first.
  2. Export artifacts from a known Git commit.
  3. Publish only the built model files to this Hugging Face repository, preferably from the tagged GitHub Actions release workflow in yet-another-ai/qts.
  4. Keep this model card aligned with the GitHub docs, but do not treat this repository as a second source repository.

Included Files

Expected root layout:

qwen3-tts-0.6b-f16.gguf
qwen3-tts-vocoder.onnx
README.md
SHA256SUMS

Notes:

  • qwen3-tts-vocoder.onnx is shared across all GGUF variants in this repository.
  • The Rust runtime in qts expects the GGUF and vocoder files to live in the same directory by default.
  • Not every release must ship every quantization variant.
  • For the current artifact set, q8_0 is the recommended default download and f16 is the reference-quality export.

Current Quantization Support

At the moment, the qts exporter supports:

  • f16

Other quantization types may appear in future releases once the export and validation pipeline is ready.

Usage With qts

See the source repository for current usage and export documentation:

Typical local layout:

models/
  qwen3-tts-0.6b-f16.gguf
  qwen3-tts-vocoder.onnx

Example CLI usage:

cargo run -p qwen3-tts-cli -- synthesize \
  --model-dir /path/to/models \
  --text "hello" \
  --out target/hello.wav

Provenance

Current source repository snapshot:

  • GitHub commit: 0f66e61429aede9ff2b5d03fdd3de6d8d25535f7

Current artifact checksums:

  • qwen3-tts-0.6b-f16.gguf 33255963ba6a6f19f83830db82e0db2e5729b1d8cdbc2c2635ab2d386efe4a3c
  • qwen3-tts-vocoder.onnx 6d5e135a3d56cd2ede0b268ca8890155d09a53ca08586571ab8d914755963034

For future releases, it is recommended to record:

  • source GitHub commit SHA from yet-another-ai/qts
  • exported file list
  • SHA256 checksums
  • any release-specific notes such as added or removed quantization variants

Base Model

Base upstream model:

Downloads last month
453
GGUF
Model size
0.9B params
Architecture
qwen3-tts
Hardware compatibility
Log In to add your hardware

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for dsh0416/Qwen3-TTS-12Hz-0.6B-Base-QTS

Quantized
(12)
this model