Yohann Bearzi
bearzi
AI & ML interests
Uploading mixed-precision MLX quantizations of popular open-weight models, produced with oMLX’s oQ (sensitivity-driven bit allocation). Target averages of 2, 3, 4, 6, and 8 bits are provided where feasible; actual per-layer bits vary by measured sensitivity. Compatible with mlx-lm, mlx-vlm, and oMLX on Apple Silicon. Benchmarks and comparisons are welcome.
Recent Activity
updated a collection about 7 hours ago
Qwen3.5-397B-A17B-oQ updated a collection about 7 hours ago
Qwen3-Coder-Next-oQ updated a collection about 7 hours ago
Qwen3.5-397B-A17B-oQOrganizations
None yet