ur-dad-matt commited on
Commit
738e1ba
Β·
verified Β·
1 Parent(s): cbdcd7b

Embed logo + quickstart + tier matrix

Browse files
Files changed (1) hide show
  1. README.md +42 -77
README.md CHANGED
@@ -1,104 +1,69 @@
1
  ---
2
- title: Outlier
3
  ---
4
 
5
- # Outlier
 
 
6
 
7
- **Patented ternary mixture-of-experts models. Frontier quality on consumer hardware. Apache 2.0.**
8
 
9
- Outlier publishes two tracks of models:
10
 
11
- 1. **Consumer tier** β€” MLX (Apple Silicon) and GGUF (cross-platform) quantized builds ready to run in Ollama, LM Studio, Jan, or the [Outlier desktop app](https://outlier.host).
12
- 2. **Research / server tier** β€” full-precision ternary-MoE overlays on Qwen 2.5 (V3.2, V3.3) for evaluation and further research.
 
 
 
 
 
 
 
13
 
14
- ## Consumer tier
15
 
16
- | Tier | MLX | GGUF | RAM | Target platform |
17
- |---|---|---|---|---|
18
- | Nano (1.7B) | [Nano-MLX](https://huggingface.co/Outlier-Ai/Outlier-Nano-1.7B-MLX-4bit) | β€” | 4 GB+ | Any M-series Mac |
19
- | **Lite (7B)** | [Lite-MLX](https://huggingface.co/Outlier-Ai/Outlier-Lite-7B-MLX-4bit) | [Lite-GGUF](https://huggingface.co/Outlier-Ai/Outlier-Lite-7B-GGUF) | 8 GB+ | **Default tier** |
20
- | Compact (14B) | [Compact-MLX](https://huggingface.co/Outlier-Ai/Outlier-Compact-14B-MLX-4bit) | [Compact-GGUF](https://huggingface.co/Outlier-Ai/Outlier-Compact-14B-GGUF) | 16 GB+ | Reasoning, code |
21
- | Max (32B) | β€” | [Max-GGUF](https://huggingface.co/Outlier-Ai/Outlier-Max-32B-GGUF) | 32 GB+ | Heaviest workloads |
22
 
23
- Fastest path for a Mac user: install the desktop app β€” https://outlier.host.
24
 
25
- ## Research / server tier (V3.2 family)
26
 
27
- | Variant | MMLU | Provenance |
28
  |---|---|---|
29
- | [Outlier-10B-V3.2](https://huggingface.co/Outlier-Ai/Outlier-10B-V3.2) | ~76% | `[CLAIM]` smoke test |
30
- | [Outlier-40B-V3.2](https://huggingface.co/Outlier-Ai/Outlier-40B-V3.2) | **77.80%** | `[VERIFIED]` n=14,042 |
31
- | [Outlier-70B-V3.2](https://huggingface.co/Outlier-Ai/Outlier-70B-V3.2) | **81.49%** | `[VERIFIED]` n=14,042 |
32
- | [Outlier-150B-V3.2](https://huggingface.co/Outlier-Ai/Outlier-150B-V3.2) | ~83% | `[INCOMPLETE]` provenance |
33
-
34
- Per Rule 66: `[VERIFIED]` = complete source file + provenance; `[INCOMPLETE]` = number on disk but fields stripped; `[CLAIM]` = historical smoke-test pending re-run.
35
-
36
- ## V3.3 preview
37
-
38
- - [Outlier-70B-V3.3](https://huggingface.co/Outlier-Ai/Outlier-70B-V3.3) (weights upload pending, Day 18)
39
-
40
- ## Architecture β€” plain English
41
-
42
- Outlier flagship models use **ternary mixture-of-experts**: weights are {-1, 0, +1} instead of 16-bit floats, so a matrix multiply becomes a stream of additions and subtractions. Combined with **expert paging** (only the experts a given token routes to are resident in RAM; the rest live on disk), Outlier runs a 150B-parameter model on a single Mac Studio at usable speeds.
43
-
44
- Three provisional patents filed April 2026 (61 claims total) cover the ternary MoE composition, expert paging, and specialist merging techniques.
45
-
46
- ## Quick start
47
 
48
- **Mac (fastest):** [outlier.host](https://outlier.host) β€” download the desktop app.
49
 
50
- **Mac command-line (MLX):**
51
- ```python
 
52
  from mlx_lm import load, generate
53
- m, t = load("Outlier-Ai/Outlier-Lite-7B-MLX-4bit")
54
- print(generate(m, t, "What is the capital of France?", max_tokens=200))
 
55
  ```
56
 
57
- **Any platform (Ollama):**
58
- ```bash
59
- ollama run hf.co/Outlier-Ai/Outlier-Lite-7B-GGUF:Q4_K_M
60
- ```
61
 
62
- **Server-grade research (Transformers):**
63
- ```python
64
- from transformers import AutoModelForCausalLM, AutoTokenizer
65
- m = AutoModelForCausalLM.from_pretrained(
66
- "Outlier-Ai/Outlier-40B-V3.2", trust_remote_code=True, torch_dtype="auto"
67
- )
68
  ```
69
 
70
- ## Collections
71
 
72
- - [Outlier Consumer Edition](https://huggingface.co/collections/Outlier-Ai/outlier-consumer-edition-69e2fb4a0df119ea1747275e)
73
- - [Outlier Server V3.2](https://huggingface.co/collections/Outlier-Ai/outlier-server-v32-69e2fb4b71984614b3c7a4a3)
74
- - [Outlier Research](https://huggingface.co/collections/Outlier-Ai/outlier-research-69e2fb3a71984614b3c7a279)
75
 
76
- ## Patents
77
 
78
- Three provisional patents filed April 2026 (61 claims). Non-provisional deadline: April 2027. Patent #3 covers pre-attention expert prediction β€” empirical scaling to 40B/70B is pending cluster validation.
79
 
80
- ## Links
81
 
82
- - Website: https://outlier.host
83
- - GitHub: https://github.com/Outlier-host/Outlier
84
- - Desktop app (Mac): https://outlier.host/download (notarization pending)
85
- - Founders deal: $199 for 2 years β€” first 100 seats
86
-
87
- ## Citation
88
-
89
- ```bibtex
90
- @misc{outlier2026,
91
- author = {Kerr, Matt},
92
- title = {Outlier: Ternary Mixture-of-Experts for Consumer Hardware},
93
- year = {2026},
94
- publisher = {Hugging Face},
95
- howpublished = {\url{https://huggingface.co/Outlier-Ai}}
96
- }
97
- ```
98
 
99
- ## Roadmap
100
 
101
- - **Day 18:** V3.3 weights uploads; MMLU re-verification on 10B-V3.2 and 150B-V3.2; Apple Dev ID enrollment.
102
- - **Day 20:** Path F+ v3 (code specialist retrain 500 β†’ 2000 steps).
103
- - **Day 24:** ICML 2026 Workshop submission; public launch.
104
- - **Later:** True warehouse paging at 150B; shadow-space BTX merge (Claim B); desktop-app notarization.
 
1
  ---
2
+ thumbnail: https://huggingface.co/Outlier-Ai/README/resolve/main/assets/thumbnail.png
3
  ---
4
 
5
+ <p align="center">
6
+ <img src="assets/logo.png" alt="Outlier" width="180" />
7
+ </p>
8
 
9
+ <h1 align="center">Outlier</h1>
10
 
11
+ <p align="center"><em>Frontier AI on hardware you own.</em></p>
12
 
13
+ <p align="center">
14
+ <a href="https://huggingface.co/Outlier-Ai/Outlier-Nano-1.7B-MLX-4bit">Nano (1.7B)</a>
15
+ Β·
16
+ <a href="https://huggingface.co/Outlier-Ai/Outlier-Lite-7B-MLX-4bit">Lite (7B)</a>
17
+ Β·
18
+ <a href="https://huggingface.co/Outlier-Ai/Outlier-Compact-14B-MLX-4bit">Compact (14B)</a>
19
+ Β·
20
+ <a href="https://huggingface.co/Outlier-Ai/Outlier-Max-32B-GGUF">Max (32B)</a>
21
+ </p>
22
 
23
+ ---
24
 
25
+ ## What is Outlier
 
 
 
 
 
26
 
27
+ A family of small, fast, local language models designed for the hardware people actually own β€” Apple Silicon Macs, consumer NVIDIA GPUs, and CPU-only laptops via GGUF.
28
 
29
+ Four tiers from 1.7B to 32B, shipped in three formats:
30
 
31
+ | Format | Use case | Runtime |
32
  |---|---|---|
33
+ | **MLX 4-bit** | Apple Silicon Macs | Outlier Desktop app, `mlx_lm` |
34
+ | **AutoAWQ 4-bit** | NVIDIA GPUs (datacenter) | vLLM, transformers+awq |
35
+ | **GGUF Q4_K_M / Q5_K_M** | CPU-portable, any OS | llama.cpp, Ollama, LM Studio |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
36
 
37
+ ## Quickstart β€” Apple Silicon
38
 
39
+ ```bash
40
+ pip install mlx-lm
41
+ python -c "
42
  from mlx_lm import load, generate
43
+ model, tok = load('Outlier-Ai/Outlier-Lite-7B-MLX-4bit')
44
+ print(generate(model, tok, prompt='Hello', max_tokens=100))
45
+ "
46
  ```
47
 
48
+ ## Quickstart β€” Ollama
 
 
 
49
 
50
+ ```bash
51
+ ollama pull hf.co/Outlier-Ai/Outlier-Lite-7B-GGUF
52
+ ollama run hf.co/Outlier-Ai/Outlier-Lite-7B-GGUF "Hello"
 
 
 
53
  ```
54
 
55
+ ## Desktop app
56
 
57
+ Outlier ships a native Mac desktop app with streaming chat, markdown rendering, session history, model switching, and a full settings panel. See [outlier.host](https://outlier.host) (coming soon) or download from the Releases page.
 
 
58
 
59
+ ## Research
60
 
61
+ Three provisional patents filed April 2026 (61 claims total) on the **Path F+ BTX merge architecture** that produces the tier family from base models. Non-provisional consolidation due April 2027.
62
 
63
+ ## License
64
 
65
+ Apache 2.0 on all models. Base weights inherit Qwen 2.5's license. Post-training artifacts released under the same terms.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
66
 
67
+ ## Contact
68
 
69
+ [mattkerr09@gmail.com](mailto:mattkerr09@gmail.com) β€” Matt Kerr, solo founder, Michigan.