Nemotron-Post-Training-v3 Collection Collection of datasets used in the post-training phase of Nemotron Nano and Super v3. • 28 items • Updated 6 days ago • 119
Open Pangram Collection Open models and datasets based on Pangram's ICLR 2026 EditLens paper licensed for noncommercial use ONLY under CC BY-NC-SA 4.0 • 4 items • Updated 19 days ago • 8
CodeScout Collection RL-trained code search agents (1.7B, 4B, 14B) that outperform 2–18× larger models using only a Unix terminal. 📄 arxiv.org/abs/2603.17829 • 12 items • Updated 25 days ago • 7
view article Article Keep the Tokens Flowing: Lessons from 16 Open-Source RL Libraries +7 Mar 10 • 124
Distil Efficiency Benchmarks Collection Collection of models used in the blog post www.distillabs.ai/blog/the-10x-inference-tax-you-dont-have-to-pay • 9 items • Updated Mar 2 • 3
Quantized Qwen3.5 Collection Verified models. Compatible with Transformers v5.3 and vLLM v0.16.1rc1 (nightly). Under evaluation. • 9 items • Updated Mar 12 • 9
REAM Collection Compressed MoE models with a reduced number of experts. See additional models at https://huggingface.co/bknyaz. • 11 items • Updated 3 days ago • 5
Cerebras REAP Collection Sparse MoE models compressed using REAP (Router-weighted Expert Activation Pruning) method • 30 items • Updated Feb 25 • 135
gliner2 family Collection GLiNER2 extends the original GLiNER architecture to support multi-task information extraction with a schema-driven interface. This base model provid • 4 items • Updated Feb 10 • 42
view article Article From Golden Gate Bridge to Broken JSON: Why Anthropic's SAE Steering Fails for Structured Output Feb 7 • 22
view article Article Tensor Parallelism (TP) in Transformers: 5 Minutes to Understand Dec 4, 2025 • 68
The Bestiary Collection Decensored language models made using Heretic (https://github.com/p-e-w/heretic) • 6 items • Updated Nov 16, 2025 • 110