This model is a pruned variant of ananayarora/Kimi-K2.5-BF16 that retains the first 2 layer(s) of the original 61 layer(s) architecture. It is intended for pipeline testing and performance research rather than production use.

Made with ❤️ by Model Pruner

Kimi K2.5 — BF16 Safetensors

BF16 (bfloat16) conversion of Moonshot AI's Kimi K2.5.

Converted from the official native INT4 weights so that LlamaFactory + KTransformers can run LoRA SFT directly without a per-run conversion step.

Spec Value
Source moonshotai/Kimi-K2.5
Format safetensors (BF16)
Shards 64
Total size ~1913 GB
Architecture MoE — 1T total, 32B active
Context 256K tokens

Usage with LlamaFactory + KTransformers

model_name_or_path: ananayarora/Kimi-K2.5-BF16
stage: sft
finetuning_type: lora
bf16: true
use_kt: true

License

Same as the base model: Modified MIT.

Downloads last month
13
Safetensors
Model size
20B params
Tensor type
BF16
·
F8_E4M3
·
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for ubicloud/Kimi-K2.5-BF16-Pruned-20B-FP8

Quantized
(1)
this model