TIPSv2 Collection TIPSv2 foundational vision-language models. Webpage: https://gdm-tipsv2.github.io/ • 9 items • Updated 3 days ago • 15
MobileLLM-R1 Collection MobileLLM-R1, a series of sub-billion parameter reasoning models • 10 items • Updated Nov 21, 2025 • 30
MobileLLM Collection Optimizing Sub-billion Parameter Language Models for On-Device Use Cases (ICML 2024) https://arxiv.org/abs/2402.14905 • 49 items • Updated Mar 2 • 141
Continual Quantization-Aware Pre-Training: When to transition from 16-bit to 1.58-bit pre-training for BitNet language models? Paper • 2502.11895 • Published Feb 17, 2025 • 3
An Extra RMSNorm is All You Need for Fine Tuning to 1.58 Bits Paper • 2505.08823 • Published May 12, 2025 • 2
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits Paper • 2402.17764 • Published Feb 27, 2024 • 628
Byte Latent Transformer: Patches Scale Better Than Tokens Paper • 2412.09871 • Published Dec 13, 2024 • 108
llama.vim Collection Recommended models for the llama.vim and llama.vscode plugins • 10 items • Updated Dec 16, 2025 • 67
story writing favourites Collection Models I personally liked for generating stories in the past. Not a recommendation, most of these are outdated. • 17 items • Updated Mar 2 • 98
Spectra: A Comprehensive Study of Ternary, Quantized, and FP16 Language Models Paper • 2407.12327 • Published Jul 17, 2024 • 79