Embed logo + quickstart + tier matrix
Browse files
README.md
CHANGED
|
@@ -1,104 +1,69 @@
|
|
| 1 |
---
|
| 2 |
-
|
| 3 |
---
|
| 4 |
|
| 5 |
-
|
|
|
|
|
|
|
| 6 |
|
| 7 |
-
|
| 8 |
|
| 9 |
-
|
| 10 |
|
| 11 |
-
|
| 12 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 13 |
|
| 14 |
-
|
| 15 |
|
| 16 |
-
|
| 17 |
-
|---|---|---|---|---|
|
| 18 |
-
| Nano (1.7B) | [Nano-MLX](https://huggingface.co/Outlier-Ai/Outlier-Nano-1.7B-MLX-4bit) | β | 4 GB+ | Any M-series Mac |
|
| 19 |
-
| **Lite (7B)** | [Lite-MLX](https://huggingface.co/Outlier-Ai/Outlier-Lite-7B-MLX-4bit) | [Lite-GGUF](https://huggingface.co/Outlier-Ai/Outlier-Lite-7B-GGUF) | 8 GB+ | **Default tier** |
|
| 20 |
-
| Compact (14B) | [Compact-MLX](https://huggingface.co/Outlier-Ai/Outlier-Compact-14B-MLX-4bit) | [Compact-GGUF](https://huggingface.co/Outlier-Ai/Outlier-Compact-14B-GGUF) | 16 GB+ | Reasoning, code |
|
| 21 |
-
| Max (32B) | β | [Max-GGUF](https://huggingface.co/Outlier-Ai/Outlier-Max-32B-GGUF) | 32 GB+ | Heaviest workloads |
|
| 22 |
|
| 23 |
-
|
| 24 |
|
| 25 |
-
|
| 26 |
|
| 27 |
-
|
|
| 28 |
|---|---|---|
|
| 29 |
-
|
|
| 30 |
-
|
|
| 31 |
-
|
|
| 32 |
-
| [Outlier-150B-V3.2](https://huggingface.co/Outlier-Ai/Outlier-150B-V3.2) | ~83% | `[INCOMPLETE]` provenance |
|
| 33 |
-
|
| 34 |
-
Per Rule 66: `[VERIFIED]` = complete source file + provenance; `[INCOMPLETE]` = number on disk but fields stripped; `[CLAIM]` = historical smoke-test pending re-run.
|
| 35 |
-
|
| 36 |
-
## V3.3 preview
|
| 37 |
-
|
| 38 |
-
- [Outlier-70B-V3.3](https://huggingface.co/Outlier-Ai/Outlier-70B-V3.3) (weights upload pending, Day 18)
|
| 39 |
-
|
| 40 |
-
## Architecture β plain English
|
| 41 |
-
|
| 42 |
-
Outlier flagship models use **ternary mixture-of-experts**: weights are {-1, 0, +1} instead of 16-bit floats, so a matrix multiply becomes a stream of additions and subtractions. Combined with **expert paging** (only the experts a given token routes to are resident in RAM; the rest live on disk), Outlier runs a 150B-parameter model on a single Mac Studio at usable speeds.
|
| 43 |
-
|
| 44 |
-
Three provisional patents filed April 2026 (61 claims total) cover the ternary MoE composition, expert paging, and specialist merging techniques.
|
| 45 |
-
|
| 46 |
-
## Quick start
|
| 47 |
|
| 48 |
-
|
| 49 |
|
| 50 |
-
|
| 51 |
-
|
|
|
|
| 52 |
from mlx_lm import load, generate
|
| 53 |
-
|
| 54 |
-
print(generate(
|
|
|
|
| 55 |
```
|
| 56 |
|
| 57 |
-
|
| 58 |
-
```bash
|
| 59 |
-
ollama run hf.co/Outlier-Ai/Outlier-Lite-7B-GGUF:Q4_K_M
|
| 60 |
-
```
|
| 61 |
|
| 62 |
-
|
| 63 |
-
|
| 64 |
-
|
| 65 |
-
m = AutoModelForCausalLM.from_pretrained(
|
| 66 |
-
"Outlier-Ai/Outlier-40B-V3.2", trust_remote_code=True, torch_dtype="auto"
|
| 67 |
-
)
|
| 68 |
```
|
| 69 |
|
| 70 |
-
##
|
| 71 |
|
| 72 |
-
|
| 73 |
-
- [Outlier Server V3.2](https://huggingface.co/collections/Outlier-Ai/outlier-server-v32-69e2fb4b71984614b3c7a4a3)
|
| 74 |
-
- [Outlier Research](https://huggingface.co/collections/Outlier-Ai/outlier-research-69e2fb3a71984614b3c7a279)
|
| 75 |
|
| 76 |
-
##
|
| 77 |
|
| 78 |
-
Three provisional patents filed April 2026 (61 claims)
|
| 79 |
|
| 80 |
-
##
|
| 81 |
|
| 82 |
-
-
|
| 83 |
-
- GitHub: https://github.com/Outlier-host/Outlier
|
| 84 |
-
- Desktop app (Mac): https://outlier.host/download (notarization pending)
|
| 85 |
-
- Founders deal: $199 for 2 years β first 100 seats
|
| 86 |
-
|
| 87 |
-
## Citation
|
| 88 |
-
|
| 89 |
-
```bibtex
|
| 90 |
-
@misc{outlier2026,
|
| 91 |
-
author = {Kerr, Matt},
|
| 92 |
-
title = {Outlier: Ternary Mixture-of-Experts for Consumer Hardware},
|
| 93 |
-
year = {2026},
|
| 94 |
-
publisher = {Hugging Face},
|
| 95 |
-
howpublished = {\url{https://huggingface.co/Outlier-Ai}}
|
| 96 |
-
}
|
| 97 |
-
```
|
| 98 |
|
| 99 |
-
##
|
| 100 |
|
| 101 |
-
|
| 102 |
-
- **Day 20:** Path F+ v3 (code specialist retrain 500 β 2000 steps).
|
| 103 |
-
- **Day 24:** ICML 2026 Workshop submission; public launch.
|
| 104 |
-
- **Later:** True warehouse paging at 150B; shadow-space BTX merge (Claim B); desktop-app notarization.
|
|
|
|
| 1 |
---
|
| 2 |
+
thumbnail: https://huggingface.co/Outlier-Ai/README/resolve/main/assets/thumbnail.png
|
| 3 |
---
|
| 4 |
|
| 5 |
+
<p align="center">
|
| 6 |
+
<img src="assets/logo.png" alt="Outlier" width="180" />
|
| 7 |
+
</p>
|
| 8 |
|
| 9 |
+
<h1 align="center">Outlier</h1>
|
| 10 |
|
| 11 |
+
<p align="center"><em>Frontier AI on hardware you own.</em></p>
|
| 12 |
|
| 13 |
+
<p align="center">
|
| 14 |
+
<a href="https://huggingface.co/Outlier-Ai/Outlier-Nano-1.7B-MLX-4bit">Nano (1.7B)</a>
|
| 15 |
+
Β·
|
| 16 |
+
<a href="https://huggingface.co/Outlier-Ai/Outlier-Lite-7B-MLX-4bit">Lite (7B)</a>
|
| 17 |
+
Β·
|
| 18 |
+
<a href="https://huggingface.co/Outlier-Ai/Outlier-Compact-14B-MLX-4bit">Compact (14B)</a>
|
| 19 |
+
Β·
|
| 20 |
+
<a href="https://huggingface.co/Outlier-Ai/Outlier-Max-32B-GGUF">Max (32B)</a>
|
| 21 |
+
</p>
|
| 22 |
|
| 23 |
+
---
|
| 24 |
|
| 25 |
+
## What is Outlier
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 26 |
|
| 27 |
+
A family of small, fast, local language models designed for the hardware people actually own β Apple Silicon Macs, consumer NVIDIA GPUs, and CPU-only laptops via GGUF.
|
| 28 |
|
| 29 |
+
Four tiers from 1.7B to 32B, shipped in three formats:
|
| 30 |
|
| 31 |
+
| Format | Use case | Runtime |
|
| 32 |
|---|---|---|
|
| 33 |
+
| **MLX 4-bit** | Apple Silicon Macs | Outlier Desktop app, `mlx_lm` |
|
| 34 |
+
| **AutoAWQ 4-bit** | NVIDIA GPUs (datacenter) | vLLM, transformers+awq |
|
| 35 |
+
| **GGUF Q4_K_M / Q5_K_M** | CPU-portable, any OS | llama.cpp, Ollama, LM Studio |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 36 |
|
| 37 |
+
## Quickstart β Apple Silicon
|
| 38 |
|
| 39 |
+
```bash
|
| 40 |
+
pip install mlx-lm
|
| 41 |
+
python -c "
|
| 42 |
from mlx_lm import load, generate
|
| 43 |
+
model, tok = load('Outlier-Ai/Outlier-Lite-7B-MLX-4bit')
|
| 44 |
+
print(generate(model, tok, prompt='Hello', max_tokens=100))
|
| 45 |
+
"
|
| 46 |
```
|
| 47 |
|
| 48 |
+
## Quickstart β Ollama
|
|
|
|
|
|
|
|
|
|
| 49 |
|
| 50 |
+
```bash
|
| 51 |
+
ollama pull hf.co/Outlier-Ai/Outlier-Lite-7B-GGUF
|
| 52 |
+
ollama run hf.co/Outlier-Ai/Outlier-Lite-7B-GGUF "Hello"
|
|
|
|
|
|
|
|
|
|
| 53 |
```
|
| 54 |
|
| 55 |
+
## Desktop app
|
| 56 |
|
| 57 |
+
Outlier ships a native Mac desktop app with streaming chat, markdown rendering, session history, model switching, and a full settings panel. See [outlier.host](https://outlier.host) (coming soon) or download from the Releases page.
|
|
|
|
|
|
|
| 58 |
|
| 59 |
+
## Research
|
| 60 |
|
| 61 |
+
Three provisional patents filed April 2026 (61 claims total) on the **Path F+ BTX merge architecture** that produces the tier family from base models. Non-provisional consolidation due April 2027.
|
| 62 |
|
| 63 |
+
## License
|
| 64 |
|
| 65 |
+
Apache 2.0 on all models. Base weights inherit Qwen 2.5's license. Post-training artifacts released under the same terms.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 66 |
|
| 67 |
+
## Contact
|
| 68 |
|
| 69 |
+
[mattkerr09@gmail.com](mailto:mattkerr09@gmail.com) β Matt Kerr, solo founder, Michigan.
|
|
|
|
|
|
|
|
|