kani-tts-2-pt (LLM part) โ€” GGUF

GGUF conversions of nineninesix/kani-tts-2-pt LLM part only.

Files

  • kani-tts-2-pt.F16.gguf
  • kani-tts-2-pt.Q8_0.gguf
  • kani-tts-2-pt.Q6_K.gguf
  • kani-tts-2-pt.Q5_K_M.gguf
  • kani-tts-2-pt.Q4_K_M.gguf
  • kani-tts-2-pt.Q3_K_M.gguf
  • kani-tts-2-pt.Q2_K.gguf

Notes

  • Converted with llama.cpp convert_hf_to_gguf.py and quantized with llama-quantize.
  • This checkpoint uses a custom HF architecture name (KaniTTS2ForCausalLM) but is compatible with the existing lfm2 GGUF arch in llama.cpp.
  • Non-LLM tensors (speaker conditioning projection) and unsupported learnable-RoPE parameters were omitted to keep this repo focused on the LLM part.
Downloads last month
147
GGUF
Model size
68.6M params
Architecture
nemo_nano_codec
Hardware compatibility
Log In to add your hardware

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

32-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for hans00/kani-tts-2-GGUF

Quantized
(2)
this model