LLM-Drop Collection Model weights of paper "What Matters in Transformers? Not All Attention is Needed" (https://arxiv.org/abs/2406.15786) • 15 items • Updated 3 days ago
LLM-Drop Collection Model weights of paper "Uncovering the Redundancy in Transformers via a Unified Study of Layer Dropping (TMLR)". • 18 items • Updated 2 days ago • 6
Demystifying When Pruning Works via Representation Hierarchies Paper • 2603.24652 • Published 9 days ago • 20
Jackrong/Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled Image-Text-to-Text • 28B • Updated 9 days ago • 589k • 2.65k
facebook/dinov3-vitl16-pretrain-lvd1689m Image Feature Extraction • 0.3B • Updated Aug 19, 2025 • 710k • 216