byteshape/Devstral-Small-2-24B-Instruct-2512-GGUF Text Generation โข 24B โข Updated Feb 18 โข 3.65k โข 27
view post Post 5213 We collaborated with Hugging Face to enable you to train MoE models 12ร faster with 35% less VRAM via our new Triton kernels (no accuracy loss). ๐คTrain gpt-oss locally on 12.8GB VRAM with our free notebooks: https://unsloth.ai/docs/new/faster-moe See translation 1 reply ยท ๐ฅ 29 29 ๐ค 5 5 + Reply