How to use from the
Use from the
llama-cpp-python library
# !pip install llama-cpp-python

from llama_cpp import Llama

llm = Llama.from_pretrained(
	repo_id="batiai/MiniMax-M2.7-GGUF",
	filename="",
)
llm.create_chat_completion(
	messages = [
		{
			"role": "user",
			"content": "What is the capital of France?"
		}
	]
)

MiniMax M2.7 GGUF β€” Quantized by BatiAI

BatiFlow Ollama

IQ3_XXS quantization of MiniMaxAI/MiniMax-M2.7 (229B Dense) for on-device AI on Mac. Built and verified by BatiAI for BatiFlow.

Why MiniMax M2.7?

  • 229B Dense β€” one of the largest open models
  • Outperforms GPT-5.3 on GDPval-AA (ELO 1495)
  • Toolathon: 46.3% accuracy (global top tier)
  • Agent Teams, complex Skills, dynamic tool search
  • Runs on a 128GB MacBook Pro β€” no cloud needed

Quick Start

ollama pull batiai/minimax-m2.7:iq3

Available Quantizations

Quant Size VRAM M4 Max (128GB) Recommended For
IQ3_XXS 82GB 104GB 36.7 t/s 128GB+ Mac

Benchmarks β€” MacBook Pro M4 Max (128GB)

Metric IQ3_XXS
Token gen (short) 22.1 t/s
Token gen (long, 300 tokens) 36.7 t/s
Prompt eval 14.8 t/s
VRAM 104 GB (97% GPU / 3% CPU)
Cold start 42 seconds
Korean output βœ…
Tool call JSON βœ…
Basic math (2+2) βœ…

RAM Requirements

Your Mac RAM IQ3_XXS (82GB)
64GB or less ❌ Won't fit
96GB ⚠️ Heavy swap, unusable
128GB βœ… 36.7 t/s
192GB+ βœ… Fast, with headroom

229B on a Laptop

This is a 229B parameter dense model running entirely on-device β€” no cloud, no API, no costs. IQ3_XXS quantization compresses from 457GB (BF16) to 82GB while maintaining Korean, tool calling, and reasoning capabilities.

Model Comparison β€” Which BatiAI Model for Your Mac?

Your Mac Best Model Speed
16GB batiai/gemma4-e4b:q4 57 t/s
24GB batiai/gemma4-26b:iq4 85 t/s
36GB batiai/qwen3.5-35b:iq4 26.6 t/s
48GB batiai/gemma4-31b:iq4 22.8 t/s
128GB batiai/minimax-m2.7:iq3 36.7 t/s

Why BatiAI Quantization?

BatiAI Third-party (unsloth, etc.)
Source Quantized from official MiniMax weights Re-quantized from other GGUFs
Tested on Real MacBook Pro M4 Max (128GB) Often untested on consumer hardware
Tool Calling βœ… Verified Often untested
Korean βœ… Validated Not tested
imatrix βœ… Calibrated for quality Standard or none

Technical Details

About BatiFlow

BatiFlow β€” free, on-device AI automation for Mac. 5MB app, 100% local, unlimited. 57+ built-in tools for calendar, notes, reminders, files, email, browser, messaging.

License

Quantized from MiniMaxAI/MiniMax-M2.7. License: Modified-MIT.

Downloads last month
1,305
GGUF
Model size
229B params
Architecture
minimax-m2
Hardware compatibility
Log In to add your hardware

3-bit

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for batiai/MiniMax-M2.7-GGUF

Quantized
(103)
this model

Collection including batiai/MiniMax-M2.7-GGUF