batiai/DeepSeek-V4-Flash-GGUF
Text Generation • 284B • Updated • 1.2k • 3
Largest open-weight LLMs, BatiAI-quantized. Mac-runnable from M4 Max 128GB to Mac Studio M3 Ultra 512GB.
Note 284B-A13B MoE • CSA+HCA hybrid attention • SWE-Bench Pro top tier • via batiai/bati.cpp
Note 128B Dense • SWE-Bench Verified 77.6% • Modified MIT • measured 6.8 t/s on M4 Max IQ3