hero775 commited on
Commit
4ddb16a
Β·
verified Β·
1 Parent(s): fd4ebdc

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +157 -0
README.md ADDED
@@ -0,0 +1,157 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: modified-mit
4
+ license_link: https://huggingface.co/MiniMaxAI/MiniMax-M2.7/blob/main/LICENSE
5
+ base_model: MiniMaxAI/MiniMax-M2.7
6
+ base_model_relation: quantized
7
+ tags:
8
+ - gguf
9
+ - quantized
10
+ - llama.cpp
11
+ - batiai
12
+ - batiflow
13
+ - moe
14
+ - agentic
15
+ - minimax
16
+ library_name: gguf
17
+ pipeline_tag: text-generation
18
+ ---
19
+
20
+ # MiniMax-M2.7 GGUF β€” Quantized by BatiAI
21
+
22
+ <p align="center">
23
+ <a href="https://flow.bati.ai"><img src="https://img.shields.io/badge/BatiFlow-macOS%20AI%20Automation-blue?style=for-the-badge&logo=apple" alt="BatiFlow"></a>
24
+ <a href="https://ollama.com/batiai/minimax-m2.7"><img src="https://img.shields.io/badge/Ollama-batiai/minimax--m2.7-green?style=for-the-badge" alt="Ollama"></a>
25
+ </p>
26
+
27
+ > Optimized GGUF quantizations of **[MiniMaxAI/MiniMax-M2.7](https://huggingface.co/MiniMaxAI/MiniMax-M2.7)** for on-device AI on Mac.
28
+ > Built and verified by [BatiAI](https://bati.ai) for [BatiFlow](https://flow.bati.ai) β€” free, unlimited, on-device AI automation for Mac. Just 5MB.
29
+
30
+ ## About MiniMax-M2.7
31
+
32
+ MiniMax-M2.7 is a **229B parameter MoE** agentic model (10B active per forward pass) released April 2026. It is the first model deeply participating in its own evolution:
33
+
34
+ - **ELO 1495 on GDPval-AA** β€” highest among open-source models, surpassing GPT-5.3
35
+ - **66.6% medal rate on MLE Bench Lite** (22 ML competitions), second only to Opus-4.6 and GPT-5.4
36
+ - Built for **complex agent harnesses**, multi-round Word/Excel/PPT editing, and advanced productivity tasks
37
+ - Self-evolution: internal version autonomously optimized its programming scaffold over 100+ rounds, achieving 30% performance improvement
38
+
39
+ ## Quick Start
40
+
41
+ ### Ollama (Recommended)
42
+
43
+ ```bash
44
+ # IQ3_XXS β€” 82GB, requires 128GB+ Mac
45
+ ollama pull batiai/minimax-m2.7:q3
46
+
47
+ # IQ4_XS β€” 114GB, requires 192GB+ Mac
48
+ ollama pull batiai/minimax-m2.7:q4
49
+
50
+ ollama run batiai/minimax-m2.7:q3
51
+ ```
52
+
53
+ ### llama.cpp
54
+
55
+ ```bash
56
+ # Download specific quant
57
+ huggingface-cli download batiai/MiniMax-M2.7-GGUF MiniMaxAI-MiniMax-M2.7-IQ3_XXS.gguf --local-dir .
58
+
59
+ # Run
60
+ ./llama-cli -m MiniMaxAI-MiniMax-M2.7-IQ3_XXS.gguf -p "Hello"
61
+ ```
62
+
63
+ ### BatiFlow App
64
+
65
+ [Download BatiFlow](https://github.com/batiai/batiflow-releases/releases/latest) β†’ Settings β†’ AI β†’ Ollama β†’ Select `batiai/minimax-m2.7`
66
+
67
+ ## Available Quantizations
68
+
69
+ | File | Quant Type | Size | Min RAM | Target Hardware |
70
+ |------|-----------|------|---------|----------------|
71
+ | `MiniMaxAI-MiniMax-M2.7-IQ3_XXS.gguf` | **IQ3_XXS** (imatrix) | **82GB** | 128GB | MacBook Pro M4 Max 128GB |
72
+ | `MiniMaxAI-MiniMax-M2.7-IQ4_XS.gguf` | **IQ4_XS** (imatrix) | **114GB** | 192GB | Mac Studio M4 Ultra 192GB |
73
+
74
+ All quants use **importance matrix** (imatrix) generated from wikitext-2-raw calibration data to preserve quality at low bit-widths. MoE models particularly benefit from imatrix quantization since expert utilization varies significantly.
75
+
76
+ ## Why BatiAI Quantizations?
77
+
78
+ | | BatiAI | Third-party (re-quantizations) |
79
+ |---|---|---|
80
+ | **Source** | Quantized directly from official MiniMax BF16 weights | Re-quantized from other GGUF files |
81
+ | **imatrix** | Applied to all IQ quants (quality-optimized) | Often missing |
82
+ | **Tested on** | Real Apple Silicon hardware | Untested on consumer hardware |
83
+ | **Tool Calling** | Verified with BatiFlow's 57 tool functions | Often broken on MoE models |
84
+ | **Korean** | Validated Korean text generation | Not tested |
85
+ | **Metadata** | `general.author: BatiAI` signed | Varies |
86
+
87
+ ## Technical Details
88
+
89
+ - **Original Model**: [MiniMaxAI/MiniMax-M2.7](https://huggingface.co/MiniMaxAI/MiniMax-M2.7)
90
+ - **Architecture**: MiniMax-M2 (MoE, 256 experts, 8 activated per token)
91
+ - **Parameters**: 228.7B total, ~10B active per forward pass
92
+ - **Context Window**: 196,608 tokens (192K native)
93
+ - **License**: modified-MIT (commercial use permitted)
94
+ - **Quantized with**: [llama.cpp](https://github.com/ggml-org/llama.cpp) (latest build, April 2026)
95
+ - **Quantized by**: [BatiAI](https://bati.ai)
96
+ - **GGUF Metadata**: `general.author: BatiAI`, `general.url: https://flow.bati.ai`
97
+
98
+ ## How We Quantize
99
+
100
+ ```
101
+ MiniMaxAI/MiniMax-M2.7 official BF16 safetensors (~230GB)
102
+ ↓ llama.cpp convert_hf_to_gguf.py
103
+ BF16 GGUF (~457GB)
104
+ ↓ llama-imatrix with wikitext-2 calibration (~6h CPU)
105
+ imatrix.dat
106
+ ↓ llama-quantize --imatrix
107
+ IQ3_XXS (82GB) + IQ4_XS (114GB)
108
+ ↓ BatiAI metadata signing
109
+ Published GGUFs
110
+ ```
111
+
112
+ No third-party intermediaries. Direct from source, verified on real hardware.
113
+
114
+ ## About BatiFlow
115
+
116
+ <p align="center">
117
+ <a href="https://flow.bati.ai"><strong>flow.bati.ai</strong></a>
118
+ </p>
119
+
120
+ [BatiFlow](https://flow.bati.ai) is a **macOS-native AI desktop automation app** β€” just 5MB, built with Swift.
121
+
122
+ ### Why BatiFlow?
123
+
124
+ | | |
125
+ |---|---|
126
+ | **Free & Unlimited** | On-device AI via Ollama β€” no API costs, no usage limits, no subscriptions |
127
+ | **100% Private** | All data stays on your Mac. Nothing is sent to the cloud. |
128
+ | **Ultra Lightweight** | Native macOS app, only 5MB. No Electron, no bloat. |
129
+ | **Instant Setup** | Download β†’ install Ollama β†’ start automating. Under 5 minutes. |
130
+
131
+ ### What It Does
132
+
133
+ - **AI Command Bar** β€” type natural language, AI executes actions
134
+ - **KakaoTalk, iMessage, Slack** β€” messaging automation
135
+ - **Chrome browser** β€” navigate, click, fill, screenshot via CDP
136
+ - **57 built-in tools** β€” calendar, notes, reminders, files, email, shell, and more
137
+ - **Skill builder** β€” create reusable automations with YAML
138
+ - **Multilingual** β€” Korean & English UI + AI responses
139
+
140
+ <p align="center">
141
+ <a href="https://github.com/batiai/batiflow-releases/releases/latest">
142
+ <img src="https://img.shields.io/badge/Download-BatiFlow-blue?style=for-the-badge&logo=apple" alt="Download BatiFlow">
143
+ </a>
144
+ <a href="https://flow.bati.ai">
145
+ <img src="https://img.shields.io/badge/Learn%20More-flow.bati.ai-green?style=for-the-badge" alt="Learn More">
146
+ </a>
147
+ </p>
148
+
149
+ ## License
150
+
151
+ This model is quantized from `MiniMaxAI/MiniMax-M2.7` and follows the original model's **modified-MIT license**. Commercial use is permitted per the original license terms.
152
+
153
+ BatiAI quantization pipeline and metadata are provided under MIT License.
154
+
155
+ ---
156
+
157
+ **Benchmarks on Apple Silicon** (updated after verification): See [https://github.com/batiai/batiai-models](https://github.com/batiai/batiai-models) for reproducible benchmarks and quantization pipeline.