Proven REAPs
Collection
Benchmarked REAP checkpoints with >=500 all-time downloads. GLM/Qwen/MiniMax/DeepSeek/Kimi/gemma. • 26 items • Updated • 2
Support this work: donate.sybilsolutions.ai
REAP surfaces: GLM | MiniMax | Qwen | Gemma | Paper | Code | PR17 | Cerebras Collection
50% expert-pruned version of PrimeIntellect/INTELLECT-3 using Cerebras REAP (Router-weighted Expert Activation Pruning).
| Property | Value |
|---|---|
| Base Model | PrimeIntellect/INTELLECT-3 (248B MoE) |
| Architecture | GLM-4 MoE (glm4_moe) |
| Compression | 50% (64 experts pruned) |
| Remaining Experts | 64 per layer |
| Parameters | ~124B |
| Format | BF16 SafeTensors |
| Size | 107 GB |
dataset: 0xSero/glm47-reap-calibration-v2
samples: 1360
- evol-codealpaca-v1: 700 (code generation)
- xlam-function-calling-60k: 330 (function calling)
- SWE-smith-trajectories: 330 (agentic multi-turn)
distance_measure: angular
seed: 42
model_max_length: 2048
compression_ratio: 0.50
prune_method: reap
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained(
"0xSero/INTELLECT-3-REAP-50",
torch_dtype="auto",
device_map="auto",
trust_remote_code=True
)
tokenizer = AutoTokenizer.from_pretrained("0xSero/INTELLECT-3-REAP-50", trust_remote_code=True)
messages = [{"role": "user", "content": "Write a Python function to calculate fibonacci numbers"}]
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt", add_generation_prompt=True).to(model.device)
outputs = model.generate(inputs, max_new_tokens=512)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
| Model | Compression | Format | Size |
|---|---|---|---|
| INTELLECT-3-REAP-50 | 50% | BF16 | 107GB |
| INTELLECT-3-REAP-50-W4A16 | 50% | W4A16 GPTQ | ~30GB (coming soon) |
@article{cerebras2025reap,
title={REAP: Router-weighted Expert Activation Pruning for MoE Models},
author={Cerebras Systems},
year={2025}
}
This model was created as part of efficiency research for large MoE models.
If this work is useful, support Sybil Solutions here: https://donate.sybilsolutions.ai
Thank you for the kind sponsors, wouldn't be possible without them: