These are AMD GFX906-focused GGUF quantizations of Kimi-Linear-48B-A3B-Instruct.
For GFX906 users: Kimi-Linear support has been merged into the llama.cpp-gfx906 fork.
You can git clone it and compile locally with the following commands:
git clone https://github.com/iacopPBK/llama.cpp-gfx906.git
cd llama.cpp-gfx906
./SCRIPT_compile_MI50.sh # edit ROCM_PATH if not using /opt/rocm
Full credits for the Kimi-Linear implementation goes to ymcki! See their github repo here.
- Downloads last month
- 12
Hardware compatibility
Log In to add your hardware
4-bit
16-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support
Model tree for Kamali-Lab/Kimi-Linear-48B-A3B-Instruct-GGUF
Base model
moonshotai/Kimi-Linear-48B-A3B-Instruct