metadata
language:
- en
- ko
- ja
- zh
license: other
license_name: modified-mit
license_link: https://huggingface.co/MiniMaxAI/MiniMax-M2.7/blob/main/LICENSE
tags:
- gguf
- minimax
- quantized
- apple-silicon
- ollama
- batiai
- on-device
- 229b
base_model: MiniMaxAI/MiniMax-M2.7
pipeline_tag: text-generation
library_name: llama.cpp
MiniMax M2.7 GGUF β Quantized by BatiAI
IQ3_XXS quantization of MiniMaxAI/MiniMax-M2.7 (229B Dense) for on-device AI on Mac. Built and verified by BatiAI for BatiFlow.
Why MiniMax M2.7?
- 229B Dense β one of the largest open models
- Outperforms GPT-5.3 on GDPval-AA (ELO 1495)
- Toolathon: 46.3% accuracy (global top tier)
- Agent Teams, complex Skills, dynamic tool search
- Runs on a 128GB MacBook Pro β no cloud needed
Quick Start
ollama pull batiai/minimax-m2.7:iq3
Available Quantizations
| Quant | Size | VRAM | M4 Max (128GB) | Recommended For |
|---|---|---|---|---|
| IQ3_XXS | 82GB | 104GB | 36.7 t/s | 128GB+ Mac |
Benchmarks β MacBook Pro M4 Max (128GB)
| Metric | IQ3_XXS |
|---|---|
| Token gen (short) | 22.1 t/s |
| Token gen (long, 300 tokens) | 36.7 t/s |
| Prompt eval | 14.8 t/s |
| VRAM | 104 GB (97% GPU / 3% CPU) |
| Cold start | 42 seconds |
| Korean output | β |
| Tool call JSON | β |
| Basic math (2+2) | β |
RAM Requirements
| Your Mac RAM | IQ3_XXS (82GB) |
|---|---|
| 64GB or less | β Won't fit |
| 96GB | β οΈ Heavy swap, unusable |
| 128GB | β 36.7 t/s |
| 192GB+ | β Fast, with headroom |
229B on a Laptop
This is a 229B parameter dense model running entirely on-device β no cloud, no API, no costs. IQ3_XXS quantization compresses from 457GB (BF16) to 82GB while maintaining Korean, tool calling, and reasoning capabilities.
Model Comparison β Which BatiAI Model for Your Mac?
| Your Mac | Best Model | Speed |
|---|---|---|
| 16GB | batiai/gemma4-e4b:q4 |
57 t/s |
| 24GB | batiai/gemma4-26b:iq4 |
85 t/s |
| 36GB | batiai/qwen3.5-35b:iq4 |
26.6 t/s |
| 48GB | batiai/gemma4-31b:iq4 |
22.8 t/s |
| 128GB | batiai/minimax-m2.7:iq3 |
36.7 t/s |
Why BatiAI Quantization?
| BatiAI | Third-party (unsloth, etc.) | |
|---|---|---|
| Source | Quantized from official MiniMax weights | Re-quantized from other GGUFs |
| Tested on | Real MacBook Pro M4 Max (128GB) | Often untested on consumer hardware |
| Tool Calling | β Verified | Often untested |
| Korean | β Validated | Not tested |
| imatrix | β Calibrated for quality | Standard or none |
Technical Details
- Original Model: MiniMaxAI/MiniMax-M2.7
- Architecture: Dense (229B params, all active)
- License: Modified-MIT
- Quantized with: llama.cpp
- Quantized by: BatiAI
About BatiFlow
BatiFlow β free, on-device AI automation for Mac. 5MB app, 100% local, unlimited. 57+ built-in tools for calendar, notes, reminders, files, email, browser, messaging.
License
Quantized from MiniMaxAI/MiniMax-M2.7. License: Modified-MIT.