IQ2 quant

#1
by piloponth - opened

Hi and thanks for the service :)

I'm kindly asking for IQ2 quants.
Best, Pilo

Me, too, unfortunately, llama.cpp currently fails to generate them :(

As mentioned in https://huggingface.co/mradermacher/model_requests/discussions/1239#6891b806a57cc4b88f6d0437 there unfortunately won't be any low bits-per-wight quants for any GLM 4.5 based models from us until MTP support is implemented by llama.cpp after which we will have to redo all the GLM 4.5 based models anyways.

@mradermacher When we are at GLM 4.5 please mark GLM-4.5 and GLM-4.5-Base to use the imatrix RPC setup which I will fix as soon I feel better again.

Thank you Señors.
I will be looking forward for the update in the future.

Best, Pilo.

piloponth changed discussion status to closed

Sign up or log in to comment