Mistral Medium 3.5 128B GGUF โ€” Quantized by BatiAI

BatiFlow Ollama

Quantizations of Mistral Medium 3.5 128B (Dense, Modified MIT) โ€” frontier coding model on a Mac. Built and verified by BatiAI for BatiFlow.

Why Mistral Medium 3.5?

  • SWE-Bench Verified: 77.6% โ€” second only to Gemini 3.1 Pro Preview (78.8%)
  • Tau3-Telecom: 91.4% โ€” top-tier agentic benchmark
  • 128B Dense โ€” Mistral's first flagship merged model (instruction + reasoning + coding in one weight set)
  • Modified MIT โ€” open weights, commercial-friendly
  • 256K context window + multimodal vision capabilities
  • Configurable reasoning effort per request
  • Released April 29, 2026

Quick Start

ollama pull batiai/mistral-medium-3.5:iq4

Available Quantizations

Quant Size VRAM target Recommended For
IQ3_XXS ~50GB ~64GB 96GB+ Mac (compact)
IQ4_XS ~70GB ~84GB 128GB Mac (recommended)
Q5_K_M ~90GB ~108GB 192GB+ Mac (highest quality)

RAM Requirements

Your Mac RAM IQ3 (~50GB) IQ4 (~70GB) Q5 (~90GB)
64GB โš ๏ธ Heavy swap โŒ โŒ
96GB โœ… Fast โš ๏ธ Tight โŒ
128GB โœ… โœ… Recommended โš ๏ธ Tight
192GB+ โœ… โœ… โœ… Best
512GB โœ… โœ… โœ… Headroom

Coding Powerhouse on a Mac

Benchmark Mistral Medium 3.5
SWE-Bench Verified 77.6%
Tau3-Telecom (agentic) 91.4%
Context window 256K

A coding model that goes head-to-head with cloud frontier (Gemini 3.1 Pro), runnable entirely on-device on a 128GB MacBook Pro.

Model Comparison โ€” Frontier Class on Mac

Your Mac Best Pick Speed Use Case
32GB batiai/nemotron3-nano:iq4 โ€” General agentic
96GB batiai/mistral-medium-3.5:iq3 (this) โ€” Coding agents
128GB batiai/mistral-medium-3.5:iq4 (this) โ€” Frontier coding
128GB batiai/minimax-m2.7:iq3 36.7 t/s 229B Dense, GDPval
192GB+ batiai/mistral-medium-3.5:q5 โ€” Max quality
512GB batiai/kimi-k2.6:iq4 โ€” 1T MoE, SWE-Bench Pro 58.6

Why BatiAI Quantization?

BatiAI Third-party
Source Official Mistral weights Re-quantized
imatrix โœ… wikitext-2 200 chunks Standard
Vision (text-only quants) (text-only quants)
BatiAI signed โœ… โ€”

Technical Details

About BatiFlow

BatiFlow โ€” free, on-device AI automation for Mac. 5MB app, 100% local, unlimited. 60+ tools โ€” KakaoTalk, iMessage, Slack, Calendar, browser, file system.

License

Quantized from mistralai/Mistral-Medium-3.5-128B. License: Modified MIT.

Benchmarks

Machine Quant Cold start Prompt eval Token gen Tested
MacBook Pro M4 Max 128GB IQ3_XXS 35.316s 36.7 t/s 6.82 t/s 2026-05-04
Downloads last month
697
GGUF
Model size
125B params
Architecture
mistral3
Hardware compatibility
Log In to add your hardware

3-bit

4-bit

5-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for batiai/Mistral-Medium-3.5-128B-GGUF

Quantized
(18)
this model

Collection including batiai/Mistral-Medium-3.5-128B-GGUF