24GB Models
Collection
Models optimized for 24GB VRAM • 8 items • Updated
Q4F Q8A: Q4_K ffn, Q8_0 attn, Q8_0 output, Q8_0 embeds
Fits ≥24K Q8 CTX on a 24GiB GPU
We're not able to determine the quantization variants.
Base model
Qwen/Qwen3-32B