Great model, please release ~16B model with FP8

#9
by Duonglv - opened

Hello,
27B dense model is so great. Could you please train a 16B dense model with FP8?
16B model has ~16G vRam for weights, so it fit well in 32G vRam ( 2 gpus with 16G each).
Many thanks.

Duonglv changed discussion title from Great model, please release 16B model with FP8 to Great model, please release ~16B model with FP8

I think you will fit if you use gguf. Not sure if the rule is still correct, but using a model with more knowledge in lower precision is always better than less parameters in bf16/bf8, but avoid 1-bits.

If so, a 27B model with 4 bit quantization may be better than 16B model FP8.
There are many people/group that provides quantized models, it’s worth to try or use. However, an offical quantized model from Qwen team with sufficicent evaluations is best.
Hope that Qwen team will officially release 27B model with 4 bit, such as AWQ INT4,…
I use vLLM, so gguf isn’t suitable.
I think that we can train models to fit well in 4G, 8G, 16G, 32G or 64G,…

Sign up or log in to comment