14B? Or 17B?
#2
by pymo - opened
Hi, can you tell me why your GGUF quantized version is 17B? Why isn't it 14B or 1.7B?
https://huggingface.co/bytedance-research/HuMo
The original model is 17B, not sure why others renamed it 14B
VeryAladeen changed discussion status to closed