SFT/Alignment - Phase 007-06-MLP8: ethicalabs/Kurtis-EON1-SFT Mix (1 epoch)

#9
by mrs83 - opened
ethicalabs.ai org

Screenshot 2026-03-16 at 14.47.06
Screenshot 2026-03-16 at 14.46.38

ethicalabs.ai org

Errata corrige: that's v0.7.6 not v0.7.5. v0.7.5 was still on Finetome

mrs83 changed discussion title from SFT/Alignment - Phase 007-05-MLP8: ethicalabs/Kurtis-EON1-SFT Mix (1 epoch) to SFT/Alignment - Phase 007-06-MLP8: ethicalabs/Kurtis-EON1-SFT Mix (1 epoch)
ethicalabs.ai org

Look, linear GPU MEM! ๐Ÿš€ Just running an lm_eval with batch_size=16 on the new architecture.

Notice the orange line on the left. That is a perfectly flat, constant allocation of ~15GB out of 96GB, while compute (blue line) is pinned at 100%.

By shifting to a custom 3-Pass Triton kernel the memory footprint now scales linearly, allowing us to hold massive batch sizes and long contexts entirely in memory without triggering PyTorch's garbage collection or OOMs.
Schermata del 2026-03-16 18-44-50

Sign up or log in to comment